From a10e35c0a83c437aabb6fa030bc415a95832aa78 Mon Sep 17 00:00:00 2001 From: Claudio Mignanti Date: Fri, 3 Jul 2009 10:05:18 +0000 Subject: [PATCH] Port the usb host driver and install some missing headers SVN-Revision: 16660 --- target/linux/etrax/Makefile | 8 + .../drivers/usb/host/hc-cris-dbg.h | 282 + .../drivers/usb/host/hc-crisv10.c | 9918 +++++++++++++++++ .../drivers/usb/host/hc-crisv10.h | 666 ++ .../patches-2.6.30/300-usb_support.patch | 24 + 5 files changed, 10898 insertions(+) create mode 100644 target/linux/etrax/files-2.6.30/drivers/usb/host/hc-cris-dbg.h create mode 100644 target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.c create mode 100644 target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.h create mode 100644 target/linux/etrax/patches-2.6.30/300-usb_support.patch diff --git a/target/linux/etrax/Makefile b/target/linux/etrax/Makefile index b7970fad27..ec81d9f754 100644 --- a/target/linux/etrax/Makefile +++ b/target/linux/etrax/Makefile @@ -21,4 +21,12 @@ define Target/Description Build firmware images for the FOXBOARD made by acmesystems.it endef +define Kernel/Prepare + $(call Kernel/Prepare/Default) + mkdir -p $(LINUX_DIR)/include/asm-cris + $(CP) $(LINUX_DIR)/arch/cris/include/asm/*.h $(LINUX_DIR)/include/asm-cris + mkdir -p $(LINUX_DIR)/include/asm-cris/arch + $(CP) $(LINUX_DIR)/arch/cris/include/arch-v10/arch/*.h $(LINUX_DIR)/include/asm-cris/arch +endef + $(eval $(call BuildTarget)) diff --git a/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-cris-dbg.h b/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-cris-dbg.h new file mode 100644 index 0000000000..c3f34c296a --- /dev/null +++ b/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-cris-dbg.h @@ -0,0 +1,282 @@ + +/* macros for debug output */ + +#define hcd_dbg(hcd, fmt, args...) \ + dev_info(hcd->self.controller, fmt, ## args) +#define hcd_err(hcd, fmt, args...) \ + dev_err(hcd->self.controller, fmt, ## args) +#define hcd_info(hcd, fmt, args...) \ + dev_info(hcd->self.controller, fmt, ## args) +#define hcd_warn(hcd, fmt, args...) \ + dev_warn(hcd->self.controller, fmt, ## args) + +/* +#define devdrv_dbg(fmt, args...) \ + printk(KERN_INFO "usb_devdrv dbg: ");printk(fmt, ## args) +*/ +#define devdrv_dbg(fmt, args...) {} + +#define devdrv_err(fmt, args...) \ + printk(KERN_ERR "usb_devdrv error: ");printk(fmt, ## args) +#define devdrv_info(fmt, args...) \ + printk(KERN_INFO "usb_devdrv: ");printk(fmt, ## args) + +#define irq_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_irq dbg: ");printk(fmt, ## args) +#define irq_err(fmt, args...) \ + printk(KERN_ERR "crisv10_irq error: ");printk(fmt, ## args) +#define irq_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_irq warn: ");printk(fmt, ## args) +#define irq_info(fmt, args...) \ + printk(KERN_INFO "crisv10_hcd: ");printk(fmt, ## args) + +/* +#define rh_dbg(fmt, args...) \ + printk(KERN_DEBUG "crisv10_rh dbg: ");printk(fmt, ## args) +*/ +#define rh_dbg(fmt, args...) {} + +#define rh_err(fmt, args...) \ + printk(KERN_ERR "crisv10_rh error: ");printk(fmt, ## args) +#define rh_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_rh warning: ");printk(fmt, ## args) +#define rh_info(fmt, args...) \ + printk(KERN_INFO "crisv10_rh: ");printk(fmt, ## args) + +/* +#define tc_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_tc dbg: ");printk(fmt, ## args) +*/ +#define tc_dbg(fmt, args...) {while(0){}} + +#define tc_err(fmt, args...) \ + printk(KERN_ERR "crisv10_tc error: ");printk(fmt, ## args) +/* +#define tc_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_tc warning: ");printk(fmt, ## args) +*/ +#define tc_warn(fmt, args...) {while(0){}} + +#define tc_info(fmt, args...) \ + printk(KERN_INFO "crisv10_tc: ");printk(fmt, ## args) + + +/* Debug print-outs for various traffic types */ + +#define intr_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_intr warning: ");printk(fmt, ## args) +/* +#define intr_dbg(fmt, args...) \ + printk(KERN_DEBUG "crisv10_intr dbg: ");printk(fmt, ## args) +*/ +#define intr_dbg(fmt, args...) {while(0){}} + + +#define isoc_err(fmt, args...) \ + printk(KERN_ERR "crisv10_isoc error: ");printk(fmt, ## args) +/* +#define isoc_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_isoc warning: ");printk(fmt, ## args) +*/ +#define isoc_warn(fmt, args...) {while(0){}} + +/* +#define isoc_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_isoc dbg: ");printk(fmt, ## args) +*/ +#define isoc_dbg(fmt, args...) {while(0){}} + +/* +#define timer_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_timer warning: ");printk(fmt, ## args) +*/ +#define timer_warn(fmt, args...) {while(0){}} + +/* +#define timer_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_timer dbg: ");printk(fmt, ## args) +*/ +#define timer_dbg(fmt, args...) {while(0){}} + + +/* Debug printouts for events related to late finishing of URBs */ +/* +#define late_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_late dbg: ");printk(fmt, ## args) +*/ +#define late_dbg(fmt, args...) {while(0){}} + +#define late_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_late warning: ");printk(fmt, ## args) +/* +#define errno_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_errno dbg: ");printk(fmt, ## args) +*/ +#define errno_dbg(fmt, args...) {while(0){}} + + +#define dma_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_dma dbg: ");printk(fmt, ## args) +#define dma_err(fmt, args...) \ + printk(KERN_ERR "crisv10_dma error: ");printk(fmt, ## args) +#define dma_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_dma warning: ");printk(fmt, ## args) +#define dma_info(fmt, args...) \ + printk(KERN_INFO "crisv10_dma: ");printk(fmt, ## args) + + + +#define str_dir(pipe) \ + (usb_pipeout(pipe) ? "out" : "in") +#define str_type(pipe) \ + ({ \ + char *s = "?"; \ + switch (usb_pipetype(pipe)) { \ + case PIPE_ISOCHRONOUS: s = "iso"; break; \ + case PIPE_INTERRUPT: s = "intr"; break; \ + case PIPE_CONTROL: s = "ctrl"; break; \ + case PIPE_BULK: s = "bulk"; break; \ + }; \ + s; \ + }) + +/* macros for debug output */ + +#define hcd_dbg(hcd, fmt, args...) \ + dev_info(hcd->self.controller, fmt, ## args) +#define hcd_err(hcd, fmt, args...) \ + dev_err(hcd->self.controller, fmt, ## args) +#define hcd_info(hcd, fmt, args...) \ + dev_info(hcd->self.controller, fmt, ## args) +#define hcd_warn(hcd, fmt, args...) \ + dev_warn(hcd->self.controller, fmt, ## args) + +/* +#define devdrv_dbg(fmt, args...) \ + printk(KERN_INFO "usb_devdrv dbg: ");printk(fmt, ## args) +*/ +#define devdrv_dbg(fmt, args...) {} + +#define devdrv_err(fmt, args...) \ + printk(KERN_ERR "usb_devdrv error: ");printk(fmt, ## args) +#define devdrv_info(fmt, args...) \ + printk(KERN_INFO "usb_devdrv: ");printk(fmt, ## args) + +#define irq_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_irq dbg: ");printk(fmt, ## args) +#define irq_err(fmt, args...) \ + printk(KERN_ERR "crisv10_irq error: ");printk(fmt, ## args) +#define irq_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_irq warn: ");printk(fmt, ## args) +#define irq_info(fmt, args...) \ + printk(KERN_INFO "crisv10_hcd: ");printk(fmt, ## args) + +/* +#define rh_dbg(fmt, args...) \ + printk(KERN_DEBUG "crisv10_rh dbg: ");printk(fmt, ## args) +*/ +#define rh_dbg(fmt, args...) {} + +#define rh_err(fmt, args...) \ + printk(KERN_ERR "crisv10_rh error: ");printk(fmt, ## args) +#define rh_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_rh warning: ");printk(fmt, ## args) +#define rh_info(fmt, args...) \ + printk(KERN_INFO "crisv10_rh: ");printk(fmt, ## args) + +/* +#define tc_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_tc dbg: ");printk(fmt, ## args) +*/ +#define tc_dbg(fmt, args...) {while(0){}} + +#define tc_err(fmt, args...) \ + printk(KERN_ERR "crisv10_tc error: ");printk(fmt, ## args) +/* +#define tc_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_tc warning: ");printk(fmt, ## args) +*/ +#define tc_warn(fmt, args...) {while(0){}} + +#define tc_info(fmt, args...) \ + printk(KERN_INFO "crisv10_tc: ");printk(fmt, ## args) + + +/* Debug print-outs for various traffic types */ + +#define intr_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_intr warning: ");printk(fmt, ## args) +/* +#define intr_dbg(fmt, args...) \ + printk(KERN_DEBUG "crisv10_intr dbg: ");printk(fmt, ## args) +*/ +#define intr_dbg(fmt, args...) {while(0){}} + + +#define isoc_err(fmt, args...) \ + printk(KERN_ERR "crisv10_isoc error: ");printk(fmt, ## args) +/* +#define isoc_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_isoc warning: ");printk(fmt, ## args) +*/ +#define isoc_warn(fmt, args...) {while(0){}} + +/* +#define isoc_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_isoc dbg: ");printk(fmt, ## args) +*/ +#define isoc_dbg(fmt, args...) {while(0){}} + +/* +#define timer_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_timer warning: ");printk(fmt, ## args) +*/ +#define timer_warn(fmt, args...) {while(0){}} + +/* +#define timer_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_timer dbg: ");printk(fmt, ## args) +*/ +#define timer_dbg(fmt, args...) {while(0){}} + + +/* Debug printouts for events related to late finishing of URBs */ +/* +#define late_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_late dbg: ");printk(fmt, ## args) +*/ +#define late_dbg(fmt, args...) {while(0){}} + +#define late_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_late warning: ");printk(fmt, ## args) +/* +#define errno_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_errno dbg: ");printk(fmt, ## args) +*/ +#define errno_dbg(fmt, args...) {while(0){}} + + +#define dma_dbg(fmt, args...) \ + printk(KERN_INFO "crisv10_dma dbg: ");printk(fmt, ## args) +#define dma_err(fmt, args...) \ + printk(KERN_ERR "crisv10_dma error: ");printk(fmt, ## args) +#define dma_warn(fmt, args...) \ + printk(KERN_INFO "crisv10_dma warning: ");printk(fmt, ## args) +#define dma_info(fmt, args...) \ + printk(KERN_INFO "crisv10_dma: ");printk(fmt, ## args) + + + +#define str_dir(pipe) \ + (usb_pipeout(pipe) ? "out" : "in") +#define str_type(pipe) \ + ({ \ + char *s = "?"; \ + switch (usb_pipetype(pipe)) { \ + case PIPE_ISOCHRONOUS: s = "iso"; break; \ + case PIPE_INTERRUPT: s = "intr"; break; \ + case PIPE_CONTROL: s = "ctrl"; break; \ + case PIPE_BULK: s = "bulk"; break; \ + }; \ + s; \ + }) diff --git a/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.c b/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.c new file mode 100644 index 0000000000..de06e89c2f --- /dev/null +++ b/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.c @@ -0,0 +1,9918 @@ +/* + * + * ETRAX 100LX USB Host Controller Driver + * + * Copyright (C) 2005 - 2008 Axis Communications AB + * + * Author: Konrad Eriksson + * + */ + +#include +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include + +#include "../core/hcd.h" +#include "../core/hub.h" +#include "hc-crisv10.h" +#include "hc-cris-dbg.h" + + +/***************************************************************************/ +/***************************************************************************/ +/* Host Controller settings */ +/***************************************************************************/ +/***************************************************************************/ + +#define VERSION "1.00-openwrt_diff" +#define COPYRIGHT "(c) 2005, 2006 Axis Communications AB" +#define DESCRIPTION "ETRAX 100LX USB Host Controller" + +#define ETRAX_USB_HC_IRQ USB_HC_IRQ_NBR +#define ETRAX_USB_RX_IRQ USB_DMA_RX_IRQ_NBR +#define ETRAX_USB_TX_IRQ USB_DMA_TX_IRQ_NBR + +/* Number of physical ports in Etrax 100LX */ +#define USB_ROOT_HUB_PORTS 2 + +const char hc_name[] = "hc-crisv10"; +const char product_desc[] = DESCRIPTION; + +/* The number of epids is, among other things, used for pre-allocating + ctrl, bulk and isoc EP descriptors (one for each epid). + Assumed to be > 1 when initiating the DMA lists. */ +#define NBR_OF_EPIDS 32 + +/* Support interrupt traffic intervals up to 128 ms. */ +#define MAX_INTR_INTERVAL 128 + +/* If periodic traffic (intr or isoc) is to be used, then one entry in the EP + table must be "invalid". By this we mean that we shouldn't care about epid + attentions for this epid, or at least handle them differently from epid + attentions for "valid" epids. This define determines which one to use + (don't change it). */ +#define INVALID_EPID 31 +/* A special epid for the bulk dummys. */ +#define DUMMY_EPID 30 + +/* Module settings */ + +MODULE_DESCRIPTION(DESCRIPTION); +MODULE_LICENSE("GPL"); +MODULE_AUTHOR("Konrad Eriksson "); + + +/* Module parameters */ + +/* 0 = No ports enabled + 1 = Only port 1 enabled (on board ethernet on devboard) + 2 = Only port 2 enabled (external connector on devboard) + 3 = Both ports enabled +*/ +static unsigned int ports = 3; +module_param(ports, uint, S_IRUGO); +MODULE_PARM_DESC(ports, "Bitmask indicating USB ports to use"); + + +/***************************************************************************/ +/***************************************************************************/ +/* Shared global variables for this module */ +/***************************************************************************/ +/***************************************************************************/ + +/* EP descriptor lists for non period transfers. Must be 32-bit aligned. */ +static volatile struct USB_EP_Desc TxBulkEPList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); + +static volatile struct USB_EP_Desc TxCtrlEPList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); + +/* EP descriptor lists for period transfers. Must be 32-bit aligned. */ +static volatile struct USB_EP_Desc TxIntrEPList[MAX_INTR_INTERVAL] __attribute__ ((aligned (4))); +static volatile struct USB_SB_Desc TxIntrSB_zout __attribute__ ((aligned (4))); + +static volatile struct USB_EP_Desc TxIsocEPList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); +static volatile struct USB_SB_Desc TxIsocSB_zout __attribute__ ((aligned (4))); + +static volatile struct USB_SB_Desc TxIsocSBList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); + +/* After each enabled bulk EP IN we put two disabled EP descriptors with the eol flag set, + causing the DMA to stop the DMA channel. The first of these two has the intr flag set, which + gives us a dma8_sub0_descr interrupt. When we receive this, we advance the DMA one step in the + EP list and then restart the bulk channel, thus forcing a switch between bulk EP descriptors + in each frame. */ +static volatile struct USB_EP_Desc TxBulkDummyEPList[NBR_OF_EPIDS][2] __attribute__ ((aligned (4))); + +/* List of URB pointers, where each points to the active URB for a epid. + For Bulk, Ctrl and Intr this means which URB that currently is added to + DMA lists (Isoc URBs are all directly added to DMA lists). As soon as + URB has completed is the queue examined and the first URB in queue is + removed and moved to the activeUrbList while its state change to STARTED and + its transfer(s) gets added to DMA list (exception Isoc where URBs enter + state STARTED directly and added transfers added to DMA lists). */ +static struct urb *activeUrbList[NBR_OF_EPIDS]; + +/* Additional software state info for each epid */ +static struct etrax_epid epid_state[NBR_OF_EPIDS]; + +/* Timer handles for bulk traffic timer used to avoid DMA bug where DMA stops + even if there is new data waiting to be processed */ +static struct timer_list bulk_start_timer = TIMER_INITIALIZER(NULL, 0, 0); +static struct timer_list bulk_eot_timer = TIMER_INITIALIZER(NULL, 0, 0); + +/* We want the start timer to expire before the eot timer, because the former + might start traffic, thus making it unnecessary for the latter to time + out. */ +#define BULK_START_TIMER_INTERVAL (HZ/50) /* 20 ms */ +#define BULK_EOT_TIMER_INTERVAL (HZ/16) /* 60 ms */ + +/* Delay before a URB completion happen when it's scheduled to be delayed */ +#define LATER_TIMER_DELAY (HZ/50) /* 20 ms */ + +/* Simplifying macros for checking software state info of a epid */ +/* ----------------------------------------------------------------------- */ +#define epid_inuse(epid) epid_state[epid].inuse +#define epid_out_traffic(epid) epid_state[epid].out_traffic +#define epid_isoc(epid) (epid_state[epid].type == PIPE_ISOCHRONOUS ? 1 : 0) +#define epid_intr(epid) (epid_state[epid].type == PIPE_INTERRUPT ? 1 : 0) + + +/***************************************************************************/ +/***************************************************************************/ +/* DEBUG FUNCTIONS */ +/***************************************************************************/ +/***************************************************************************/ +/* Note that these functions are always available in their "__" variants, + for use in error situations. The "__" missing variants are controlled by + the USB_DEBUG_DESC/USB_DEBUG_URB macros. */ +static void __dump_urb(struct urb* purb) +{ + struct crisv10_urb_priv *urb_priv = purb->hcpriv; + int urb_num = -1; + if(urb_priv) { + urb_num = urb_priv->urb_num; + } + printk("\nURB:0x%x[%d]\n", (unsigned int)purb, urb_num); + printk("dev :0x%08lx\n", (unsigned long)purb->dev); + printk("pipe :0x%08x\n", purb->pipe); + printk("status :%d\n", purb->status); + printk("transfer_flags :0x%08x\n", purb->transfer_flags); + printk("transfer_buffer :0x%08lx\n", (unsigned long)purb->transfer_buffer); + printk("transfer_buffer_length:%d\n", purb->transfer_buffer_length); + printk("actual_length :%d\n", purb->actual_length); + printk("setup_packet :0x%08lx\n", (unsigned long)purb->setup_packet); + printk("start_frame :%d\n", purb->start_frame); + printk("number_of_packets :%d\n", purb->number_of_packets); + printk("interval :%d\n", purb->interval); + printk("error_count :%d\n", purb->error_count); + printk("context :0x%08lx\n", (unsigned long)purb->context); + printk("complete :0x%08lx\n\n", (unsigned long)purb->complete); +} + +static void __dump_in_desc(volatile struct USB_IN_Desc *in) +{ + printk("\nUSB_IN_Desc at 0x%08lx\n", (unsigned long)in); + printk(" sw_len : 0x%04x (%d)\n", in->sw_len, in->sw_len); + printk(" command : 0x%04x\n", in->command); + printk(" next : 0x%08lx\n", in->next); + printk(" buf : 0x%08lx\n", in->buf); + printk(" hw_len : 0x%04x (%d)\n", in->hw_len, in->hw_len); + printk(" status : 0x%04x\n\n", in->status); +} + +static void __dump_sb_desc(volatile struct USB_SB_Desc *sb) +{ + char tt = (sb->command & 0x30) >> 4; + char *tt_string; + + switch (tt) { + case 0: + tt_string = "zout"; + break; + case 1: + tt_string = "in"; + break; + case 2: + tt_string = "out"; + break; + case 3: + tt_string = "setup"; + break; + default: + tt_string = "unknown (weird)"; + } + + printk(" USB_SB_Desc at 0x%08lx ", (unsigned long)sb); + printk(" command:0x%04x (", sb->command); + printk("rem:%d ", (sb->command & 0x3f00) >> 8); + printk("full:%d ", (sb->command & 0x40) >> 6); + printk("tt:%d(%s) ", tt, tt_string); + printk("intr:%d ", (sb->command & 0x8) >> 3); + printk("eot:%d ", (sb->command & 0x2) >> 1); + printk("eol:%d)", sb->command & 0x1); + printk(" sw_len:0x%04x(%d)", sb->sw_len, sb->sw_len); + printk(" next:0x%08lx", sb->next); + printk(" buf:0x%08lx\n", sb->buf); +} + + +static void __dump_ep_desc(volatile struct USB_EP_Desc *ep) +{ + printk("USB_EP_Desc at 0x%08lx ", (unsigned long)ep); + printk(" command:0x%04x (", ep->command); + printk("ep_id:%d ", (ep->command & 0x1f00) >> 8); + printk("enable:%d ", (ep->command & 0x10) >> 4); + printk("intr:%d ", (ep->command & 0x8) >> 3); + printk("eof:%d ", (ep->command & 0x2) >> 1); + printk("eol:%d)", ep->command & 0x1); + printk(" hw_len:0x%04x(%d)", ep->hw_len, ep->hw_len); + printk(" next:0x%08lx", ep->next); + printk(" sub:0x%08lx\n", ep->sub); +} + +static inline void __dump_ep_list(int pipe_type) +{ + volatile struct USB_EP_Desc *ep; + volatile struct USB_EP_Desc *first_ep; + volatile struct USB_SB_Desc *sb; + + switch (pipe_type) + { + case PIPE_BULK: + first_ep = &TxBulkEPList[0]; + break; + case PIPE_CONTROL: + first_ep = &TxCtrlEPList[0]; + break; + case PIPE_INTERRUPT: + first_ep = &TxIntrEPList[0]; + break; + case PIPE_ISOCHRONOUS: + first_ep = &TxIsocEPList[0]; + break; + default: + return; + } + ep = first_ep; + + printk("\n\nDumping EP list...\n\n"); + + do { + __dump_ep_desc(ep); + /* Cannot phys_to_virt on 0 as it turns into 80000000, which is != 0. */ + sb = ep->sub ? phys_to_virt(ep->sub) : 0; + while (sb) { + __dump_sb_desc(sb); + sb = sb->next ? phys_to_virt(sb->next) : 0; + } + ep = (volatile struct USB_EP_Desc *)(phys_to_virt(ep->next)); + + } while (ep != first_ep); +} + +static inline void __dump_ept_data(int epid) +{ + unsigned long flags; + __u32 r_usb_ept_data; + + if (epid < 0 || epid > 31) { + printk("Cannot dump ept data for invalid epid %d\n", epid); + return; + } + + local_irq_save(flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, epid); + nop(); + r_usb_ept_data = *R_USB_EPT_DATA; + local_irq_restore(flags); + + printk(" R_USB_EPT_DATA = 0x%x for epid %d :\n", r_usb_ept_data, epid); + if (r_usb_ept_data == 0) { + /* No need for more detailed printing. */ + return; + } + printk(" valid : %d\n", (r_usb_ept_data & 0x80000000) >> 31); + printk(" hold : %d\n", (r_usb_ept_data & 0x40000000) >> 30); + printk(" error_count_in : %d\n", (r_usb_ept_data & 0x30000000) >> 28); + printk(" t_in : %d\n", (r_usb_ept_data & 0x08000000) >> 27); + printk(" low_speed : %d\n", (r_usb_ept_data & 0x04000000) >> 26); + printk(" port : %d\n", (r_usb_ept_data & 0x03000000) >> 24); + printk(" error_code : %d\n", (r_usb_ept_data & 0x00c00000) >> 22); + printk(" t_out : %d\n", (r_usb_ept_data & 0x00200000) >> 21); + printk(" error_count_out : %d\n", (r_usb_ept_data & 0x00180000) >> 19); + printk(" max_len : %d\n", (r_usb_ept_data & 0x0003f800) >> 11); + printk(" ep : %d\n", (r_usb_ept_data & 0x00000780) >> 7); + printk(" dev : %d\n", (r_usb_ept_data & 0x0000003f)); +} + +static inline void __dump_ept_data_iso(int epid) +{ + unsigned long flags; + __u32 ept_data; + + if (epid < 0 || epid > 31) { + printk("Cannot dump ept data for invalid epid %d\n", epid); + return; + } + + local_irq_save(flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, epid); + nop(); + ept_data = *R_USB_EPT_DATA_ISO; + local_irq_restore(flags); + + printk(" R_USB_EPT_DATA = 0x%x for epid %d :\n", ept_data, epid); + if (ept_data == 0) { + /* No need for more detailed printing. */ + return; + } + printk(" valid : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, valid, + ept_data)); + printk(" port : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, port, + ept_data)); + printk(" error_code : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, error_code, + ept_data)); + printk(" max_len : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, max_len, + ept_data)); + printk(" ep : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, ep, + ept_data)); + printk(" dev : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, dev, + ept_data)); +} + +static inline void __dump_ept_data_list(void) +{ + int i; + + printk("Dumping the whole R_USB_EPT_DATA list\n"); + + for (i = 0; i < 32; i++) { + __dump_ept_data(i); + } +} + +static void debug_epid(int epid) { + int i; + + if(epid_isoc(epid)) { + __dump_ept_data_iso(epid); + } else { + __dump_ept_data(epid); + } + + printk("Bulk:\n"); + for(i = 0; i < 32; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxBulkEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxBulkEPList[i])); + } + } + + printk("Ctrl:\n"); + for(i = 0; i < 32; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxCtrlEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxCtrlEPList[i])); + } + } + + printk("Intr:\n"); + for(i = 0; i < MAX_INTR_INTERVAL; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxIntrEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxIntrEPList[i])); + } + } + + printk("Isoc:\n"); + for(i = 0; i < 32; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxIsocEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxIsocEPList[i])); + } + } + + __dump_ept_data_list(); + __dump_ep_list(PIPE_INTERRUPT); + printk("\n\n"); +} + + + +char* hcd_status_to_str(__u8 bUsbStatus) { + static char hcd_status_str[128]; + hcd_status_str[0] = '\0'; + if(bUsbStatus & IO_STATE(R_USB_STATUS, ourun, yes)) { + strcat(hcd_status_str, "ourun "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, perror, yes)) { + strcat(hcd_status_str, "perror "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, device_mode, yes)) { + strcat(hcd_status_str, "device_mode "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, host_mode, yes)) { + strcat(hcd_status_str, "host_mode "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, started, yes)) { + strcat(hcd_status_str, "started "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, running, yes)) { + strcat(hcd_status_str, "running "); + } + return hcd_status_str; +} + + +char* sblist_to_str(struct USB_SB_Desc* sb_desc) { + static char sblist_to_str_buff[128]; + char tmp[32], tmp2[32]; + sblist_to_str_buff[0] = '\0'; + while(sb_desc != NULL) { + switch(IO_EXTRACT(USB_SB_command, tt, sb_desc->command)) { + case 0: sprintf(tmp, "zout"); break; + case 1: sprintf(tmp, "in"); break; + case 2: sprintf(tmp, "out"); break; + case 3: sprintf(tmp, "setup"); break; + } + sprintf(tmp2, "(%s %d)", tmp, sb_desc->sw_len); + strcat(sblist_to_str_buff, tmp2); + if(sb_desc->next != 0) { + sb_desc = phys_to_virt(sb_desc->next); + } else { + sb_desc = NULL; + } + } + return sblist_to_str_buff; +} + +char* port_status_to_str(__u16 wPortStatus) { + static char port_status_str[128]; + port_status_str[0] = '\0'; + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, connected, yes)) { + strcat(port_status_str, "connected "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) { + strcat(port_status_str, "enabled "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, suspended, yes)) { + strcat(port_status_str, "suspended "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, reset, yes)) { + strcat(port_status_str, "reset "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, speed, full)) { + strcat(port_status_str, "full-speed "); + } else { + strcat(port_status_str, "low-speed "); + } + return port_status_str; +} + + +char* endpoint_to_str(struct usb_endpoint_descriptor *ed) { + static char endpoint_to_str_buff[128]; + char tmp[32]; + int epnum = ed->bEndpointAddress & 0x0F; + int dir = ed->bEndpointAddress & 0x80; + int type = ed->bmAttributes & 0x03; + endpoint_to_str_buff[0] = '\0'; + sprintf(endpoint_to_str_buff, "ep:%d ", epnum); + switch(type) { + case 0: + sprintf(tmp, " ctrl"); + break; + case 1: + sprintf(tmp, " isoc"); + break; + case 2: + sprintf(tmp, " bulk"); + break; + case 3: + sprintf(tmp, " intr"); + break; + } + strcat(endpoint_to_str_buff, tmp); + if(dir) { + sprintf(tmp, " in"); + } else { + sprintf(tmp, " out"); + } + strcat(endpoint_to_str_buff, tmp); + + return endpoint_to_str_buff; +} + +/* Debug helper functions for Transfer Controller */ +char* pipe_to_str(unsigned int pipe) { + static char pipe_to_str_buff[128]; + char tmp[64]; + sprintf(pipe_to_str_buff, "dir:%s", str_dir(pipe)); + sprintf(tmp, " type:%s", str_type(pipe)); + strcat(pipe_to_str_buff, tmp); + + sprintf(tmp, " dev:%d", usb_pipedevice(pipe)); + strcat(pipe_to_str_buff, tmp); + sprintf(tmp, " ep:%d", usb_pipeendpoint(pipe)); + strcat(pipe_to_str_buff, tmp); + return pipe_to_str_buff; +} + + +#define USB_DEBUG_DESC 1 + +#ifdef USB_DEBUG_DESC +#define dump_in_desc(x) __dump_in_desc(x) +#define dump_sb_desc(...) __dump_sb_desc(...) +#define dump_ep_desc(x) __dump_ep_desc(x) +#define dump_ept_data(x) __dump_ept_data(x) +#else +#define dump_in_desc(...) do {} while (0) +#define dump_sb_desc(...) do {} while (0) +#define dump_ep_desc(...) do {} while (0) +#endif + + +/* Uncomment this to enable massive function call trace + #define USB_DEBUG_TRACE */ + +#ifdef USB_DEBUG_TRACE +#define DBFENTER (printk(": Entering: %s\n", __FUNCTION__)) +#define DBFEXIT (printk(": Exiting: %s\n", __FUNCTION__)) +#else +#define DBFENTER do {} while (0) +#define DBFEXIT do {} while (0) +#endif + +#define CHECK_ALIGN(x) if (((__u32)(x)) & 0x00000003) \ +{panic("Alignment check (DWORD) failed at %s:%s:%d\n", __FILE__, __FUNCTION__, __LINE__);} + +/* Most helpful debugging aid */ +#define ASSERT(expr) ((void) ((expr) ? 0 : (err("assert failed at: %s %d",__FUNCTION__, __LINE__)))) + + +/***************************************************************************/ +/***************************************************************************/ +/* Forward declarations */ +/***************************************************************************/ +/***************************************************************************/ +void crisv10_hcd_epid_attn_irq(struct crisv10_irq_reg *reg); +void crisv10_hcd_port_status_irq(struct crisv10_irq_reg *reg); +void crisv10_hcd_ctl_status_irq(struct crisv10_irq_reg *reg); +void crisv10_hcd_isoc_eof_irq(struct crisv10_irq_reg *reg); + +void rh_port_status_change(__u16[]); +int rh_clear_port_feature(__u8, __u16); +int rh_set_port_feature(__u8, __u16); +static void rh_disable_port(unsigned int port); + +static void check_finished_bulk_tx_epids(struct usb_hcd *hcd, + int timer); + +static int tc_setup_epid(struct usb_host_endpoint *ep, struct urb *urb, + int mem_flags); +static void tc_free_epid(struct usb_host_endpoint *ep); +static int tc_allocate_epid(void); +static void tc_finish_urb(struct usb_hcd *hcd, struct urb *urb, int status); +static void tc_finish_urb_later(struct usb_hcd *hcd, struct urb *urb, + int status); + +static int urb_priv_create(struct usb_hcd *hcd, struct urb *urb, int epid, + int mem_flags); +static void urb_priv_free(struct usb_hcd *hcd, struct urb *urb); + +static int crisv10_usb_check_bandwidth(struct usb_device *dev,struct urb *urb); +static void crisv10_usb_claim_bandwidth( + struct usb_device *dev, struct urb *urb, int bustime, int isoc); +static void crisv10_usb_release_bandwidth( + struct usb_hcd *hcd, int isoc, int bandwidth); + +static inline struct urb *urb_list_first(int epid); +static inline void urb_list_add(struct urb *urb, int epid, + int mem_flags); +static inline urb_entry_t *urb_list_entry(struct urb *urb, int epid); +static inline void urb_list_del(struct urb *urb, int epid); +static inline void urb_list_move_last(struct urb *urb, int epid); +static inline struct urb *urb_list_next(struct urb *urb, int epid); + +int create_sb_for_urb(struct urb *urb, int mem_flags); +int init_intr_urb(struct urb *urb, int mem_flags); + +static inline void etrax_epid_set(__u8 index, __u32 data); +static inline void etrax_epid_clear_error(__u8 index); +static inline void etrax_epid_set_toggle(__u8 index, __u8 dirout, + __u8 toggle); +static inline __u8 etrax_epid_get_toggle(__u8 index, __u8 dirout); +static inline __u32 etrax_epid_get(__u8 index); + +/* We're accessing the same register position in Etrax so + when we do full access the internal difference doesn't matter */ +#define etrax_epid_iso_set(index, data) etrax_epid_set(index, data) +#define etrax_epid_iso_get(index) etrax_epid_get(index) + + +static void tc_dma_process_isoc_urb(struct urb *urb); +static void tc_dma_process_queue(int epid); +static void tc_dma_unlink_intr_urb(struct urb *urb); +static irqreturn_t tc_dma_tx_interrupt(int irq, void *vhc); +static irqreturn_t tc_dma_rx_interrupt(int irq, void *vhc); + +static void tc_bulk_start_timer_func(unsigned long dummy); +static void tc_bulk_eot_timer_func(unsigned long dummy); + + +/*************************************************************/ +/*************************************************************/ +/* Host Controler Driver block */ +/*************************************************************/ +/*************************************************************/ + +/* HCD operations */ +static irqreturn_t crisv10_hcd_top_irq(int irq, void*); +static int crisv10_hcd_reset(struct usb_hcd *); +static int crisv10_hcd_start(struct usb_hcd *); +static void crisv10_hcd_stop(struct usb_hcd *); +#ifdef CONFIG_PM +static int crisv10_hcd_suspend(struct device *, u32, u32); +static int crisv10_hcd_resume(struct device *, u32); +#endif /* CONFIG_PM */ +static int crisv10_hcd_get_frame(struct usb_hcd *); + +static int tc_urb_enqueue(struct usb_hcd *, struct urb *, gfp_t mem_flags); +static int tc_urb_dequeue(struct usb_hcd *, struct urb *, int); +static void tc_endpoint_disable(struct usb_hcd *, struct usb_host_endpoint *ep); + +static int rh_status_data_request(struct usb_hcd *, char *); +static int rh_control_request(struct usb_hcd *, u16, u16, u16, char*, u16); + +#ifdef CONFIG_PM +static int crisv10_hcd_hub_suspend(struct usb_hcd *); +static int crisv10_hcd_hub_resume(struct usb_hcd *); +#endif /* CONFIG_PM */ +#ifdef CONFIG_USB_OTG +static int crisv10_hcd_start_port_reset(struct usb_hcd *, unsigned); +#endif /* CONFIG_USB_OTG */ + +/* host controller driver interface */ +static const struct hc_driver crisv10_hc_driver = + { + .description = hc_name, + .product_desc = product_desc, + .hcd_priv_size = sizeof(struct crisv10_hcd), + + /* Attaching IRQ handler manualy in probe() */ + /* .irq = crisv10_hcd_irq, */ + + .flags = HCD_USB11, + + /* called to init HCD and root hub */ + .reset = crisv10_hcd_reset, + .start = crisv10_hcd_start, + + /* cleanly make HCD stop writing memory and doing I/O */ + .stop = crisv10_hcd_stop, + + /* return current frame number */ + .get_frame_number = crisv10_hcd_get_frame, + + + /* Manage i/o requests via the Transfer Controller */ + .urb_enqueue = tc_urb_enqueue, + .urb_dequeue = tc_urb_dequeue, + + /* hw synch, freeing endpoint resources that urb_dequeue can't */ + .endpoint_disable = tc_endpoint_disable, + + + /* Root Hub support */ + .hub_status_data = rh_status_data_request, + .hub_control = rh_control_request, +#ifdef CONFIG_PM + .hub_suspend = rh_suspend_request, + .hub_resume = rh_resume_request, +#endif /* CONFIG_PM */ +#ifdef CONFIG_USB_OTG + .start_port_reset = crisv10_hcd_start_port_reset, +#endif /* CONFIG_USB_OTG */ + }; + + +/* + * conversion between pointers to a hcd and the corresponding + * crisv10_hcd + */ + +static inline struct crisv10_hcd *hcd_to_crisv10_hcd(struct usb_hcd *hcd) +{ + return (struct crisv10_hcd *) hcd->hcd_priv; +} + +static inline struct usb_hcd *crisv10_hcd_to_hcd(struct crisv10_hcd *hcd) +{ + return container_of((void *) hcd, struct usb_hcd, hcd_priv); +} + +/* check if specified port is in use */ +static inline int port_in_use(unsigned int port) +{ + return ports & (1 << port); +} + +/* number of ports in use */ +static inline unsigned int num_ports(void) +{ + unsigned int i, num = 0; + for (i = 0; i < USB_ROOT_HUB_PORTS; i++) + if (port_in_use(i)) + num++; + return num; +} + +/* map hub port number to the port number used internally by the HC */ +static inline unsigned int map_port(unsigned int port) +{ + unsigned int i, num = 0; + for (i = 0; i < USB_ROOT_HUB_PORTS; i++) + if (port_in_use(i)) + if (++num == port) + return i; + return -1; +} + +/* size of descriptors in slab cache */ +#ifndef MAX +#define MAX(x, y) ((x) > (y) ? (x) : (y)) +#endif + + +/******************************************************************/ +/* Hardware Interrupt functions */ +/******************************************************************/ + +/* Fast interrupt handler for HC */ +static irqreturn_t crisv10_hcd_top_irq(int irq, void *vcd) +{ + struct usb_hcd *hcd = vcd; + struct crisv10_irq_reg reg; + __u32 irq_mask; + unsigned long flags; + + DBFENTER; + + ASSERT(hcd != NULL); + reg.hcd = hcd; + + /* Turn of other interrupts while handling these sensitive cases */ + local_irq_save(flags); + + /* Read out which interrupts that are flaged */ + irq_mask = *R_USB_IRQ_MASK_READ; + reg.r_usb_irq_mask_read = irq_mask; + + /* Reading R_USB_STATUS clears the ctl_status interrupt. Note that + R_USB_STATUS must be read before R_USB_EPID_ATTN since reading the latter + clears the ourun and perror fields of R_USB_STATUS. */ + reg.r_usb_status = *R_USB_STATUS; + + /* Reading R_USB_EPID_ATTN clears the iso_eof, bulk_eot and epid_attn + interrupts. */ + reg.r_usb_epid_attn = *R_USB_EPID_ATTN; + + /* Reading R_USB_RH_PORT_STATUS_1 and R_USB_RH_PORT_STATUS_2 clears the + port_status interrupt. */ + reg.r_usb_rh_port_status_1 = *R_USB_RH_PORT_STATUS_1; + reg.r_usb_rh_port_status_2 = *R_USB_RH_PORT_STATUS_2; + + /* Reading R_USB_FM_NUMBER clears the sof interrupt. */ + /* Note: the lower 11 bits contain the actual frame number, sent with each + sof. */ + reg.r_usb_fm_number = *R_USB_FM_NUMBER; + + /* Interrupts are handled in order of priority. */ + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, port_status)) { + crisv10_hcd_port_status_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, epid_attn)) { + crisv10_hcd_epid_attn_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, ctl_status)) { + crisv10_hcd_ctl_status_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, iso_eof)) { + crisv10_hcd_isoc_eof_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, bulk_eot)) { + /* Update/restart the bulk start timer since obviously the channel is + running. */ + mod_timer(&bulk_start_timer, jiffies + BULK_START_TIMER_INTERVAL); + /* Update/restart the bulk eot timer since we just received an bulk eot + interrupt. */ + mod_timer(&bulk_eot_timer, jiffies + BULK_EOT_TIMER_INTERVAL); + + /* Check for finished bulk transfers on epids */ + check_finished_bulk_tx_epids(hcd, 0); + } + local_irq_restore(flags); + + DBFEXIT; + return IRQ_HANDLED; +} + + +void crisv10_hcd_epid_attn_irq(struct crisv10_irq_reg *reg) { + struct usb_hcd *hcd = reg->hcd; + struct crisv10_urb_priv *urb_priv; + int epid; + DBFENTER; + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if (test_bit(epid, (void *)®->r_usb_epid_attn)) { + struct urb *urb; + __u32 ept_data; + int error_code; + + if (epid == DUMMY_EPID || epid == INVALID_EPID) { + /* We definitely don't care about these ones. Besides, they are + always disabled, so any possible disabling caused by the + epid attention interrupt is irrelevant. */ + continue; + } + + if(!epid_inuse(epid)) { + irq_err("Epid attention on epid:%d that isn't in use\n", epid); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + debug_epid(epid); + continue; + } + + /* Note that although there are separate R_USB_EPT_DATA and + R_USB_EPT_DATA_ISO registers, they are located at the same address and + are of the same size. In other words, this read should be ok for isoc + also. */ + ept_data = etrax_epid_get(epid); + error_code = IO_EXTRACT(R_USB_EPT_DATA, error_code, ept_data); + + /* Get the active URB for this epid. We blatantly assume + that only this URB could have caused the epid attention. */ + urb = activeUrbList[epid]; + if (urb == NULL) { + irq_err("Attention on epid:%d error:%d with no active URB.\n", + epid, error_code); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + debug_epid(epid); + continue; + } + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + /* Using IO_STATE_VALUE on R_USB_EPT_DATA should be ok for isoc also. */ + if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + + /* Isoc traffic doesn't have error_count_in/error_count_out. */ + if ((usb_pipetype(urb->pipe) != PIPE_ISOCHRONOUS) && + (IO_EXTRACT(R_USB_EPT_DATA, error_count_in, ept_data) == 3 || + IO_EXTRACT(R_USB_EPT_DATA, error_count_out, ept_data) == 3)) { + /* Check if URB allready is marked for late-finish, we can get + several 3rd error for Intr traffic when a device is unplugged */ + if(urb_priv->later_data == NULL) { + /* 3rd error. */ + irq_warn("3rd error for epid:%d (%s %s) URB:0x%x[%d]\n", epid, + str_dir(urb->pipe), str_type(urb->pipe), + (unsigned int)urb, urb_priv->urb_num); + + tc_finish_urb_later(hcd, urb, -EPROTO); + } + + } else if (reg->r_usb_status & IO_MASK(R_USB_STATUS, perror)) { + irq_warn("Perror for epid:%d\n", epid); + printk("FM_NUMBER: %d\n", reg->r_usb_fm_number & 0x7ff); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + __dump_urb(urb); + debug_epid(epid); + + if (!(ept_data & IO_MASK(R_USB_EPT_DATA, valid))) { + /* invalid ep_id */ + panic("Perror because of invalid epid." + " Deconfigured too early?"); + } else { + /* past eof1, near eof, zout transfer, setup transfer */ + /* Dump the urb and the relevant EP descriptor. */ + panic("Something wrong with DMA descriptor contents." + " Too much traffic inserted?"); + } + } else if (reg->r_usb_status & IO_MASK(R_USB_STATUS, ourun)) { + /* buffer ourun */ + printk("FM_NUMBER: %d\n", reg->r_usb_fm_number & 0x7ff); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + __dump_urb(urb); + debug_epid(epid); + + panic("Buffer overrun/underrun for epid:%d. DMA too busy?", epid); + } else { + irq_warn("Attention on epid:%d (%s %s) with no error code\n", epid, + str_dir(urb->pipe), str_type(urb->pipe)); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + __dump_urb(urb); + debug_epid(epid); + } + + } else if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, + stall)) { + /* Not really a protocol error, just says that the endpoint gave + a stall response. Note that error_code cannot be stall for isoc. */ + if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) { + panic("Isoc traffic cannot stall"); + } + + tc_dbg("Stall for epid:%d (%s %s) URB:0x%x\n", epid, + str_dir(urb->pipe), str_type(urb->pipe), (unsigned int)urb); + tc_finish_urb(hcd, urb, -EPIPE); + + } else if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, + bus_error)) { + /* Two devices responded to a transaction request. Must be resolved + by software. FIXME: Reset ports? */ + panic("Bus error for epid %d." + " Two devices responded to transaction request\n", + epid); + + } else if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, + buffer_error)) { + /* DMA overrun or underrun. */ + irq_warn("Buffer overrun/underrun for epid:%d (%s %s)\n", epid, + str_dir(urb->pipe), str_type(urb->pipe)); + + /* It seems that error_code = buffer_error in + R_USB_EPT_DATA/R_USB_EPT_DATA_ISO and ourun = yes in R_USB_STATUS + are the same error. */ + tc_finish_urb(hcd, urb, -EPROTO); + } else { + irq_warn("Unknown attention on epid:%d (%s %s)\n", epid, + str_dir(urb->pipe), str_type(urb->pipe)); + dump_ept_data(epid); + } + } + } + DBFEXIT; +} + +void crisv10_hcd_port_status_irq(struct crisv10_irq_reg *reg) +{ + __u16 port_reg[USB_ROOT_HUB_PORTS]; + DBFENTER; + port_reg[0] = reg->r_usb_rh_port_status_1; + port_reg[1] = reg->r_usb_rh_port_status_2; + rh_port_status_change(port_reg); + DBFEXIT; +} + +void crisv10_hcd_isoc_eof_irq(struct crisv10_irq_reg *reg) +{ + int epid; + struct urb *urb; + struct crisv10_urb_priv *urb_priv; + + DBFENTER; + + for (epid = 0; epid < NBR_OF_EPIDS - 1; epid++) { + + /* Only check epids that are in use, is valid and has SB list */ + if (!epid_inuse(epid) || epid == INVALID_EPID || + TxIsocEPList[epid].sub == 0 || epid == DUMMY_EPID) { + /* Nothing here to see. */ + continue; + } + ASSERT(epid_isoc(epid)); + + /* Get the active URB for this epid (if any). */ + urb = activeUrbList[epid]; + if (urb == 0) { + isoc_warn("Ignoring NULL urb for epid:%d\n", epid); + continue; + } + if(!epid_out_traffic(epid)) { + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + if (urb_priv->urb_state == NOT_STARTED) { + /* If ASAP is not set and urb->start_frame is the current frame, + start the transfer. */ + if (!(urb->transfer_flags & URB_ISO_ASAP) && + (urb->start_frame == (*R_USB_FM_NUMBER & 0x7ff))) { + /* EP should not be enabled if we're waiting for start_frame */ + ASSERT((TxIsocEPList[epid].command & + IO_STATE(USB_EP_command, enable, yes)) == 0); + + isoc_warn("Enabling isoc IN EP descr for epid %d\n", epid); + TxIsocEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + + /* This urb is now active. */ + urb_priv->urb_state = STARTED; + continue; + } + } + } + } + + DBFEXIT; +} + +void crisv10_hcd_ctl_status_irq(struct crisv10_irq_reg *reg) +{ + struct crisv10_hcd* crisv10_hcd = hcd_to_crisv10_hcd(reg->hcd); + + DBFENTER; + ASSERT(crisv10_hcd); + +/* irq_dbg("ctr_status_irq, controller status: %s\n", + hcd_status_to_str(reg->r_usb_status));*/ + + /* FIXME: What should we do if we get ourun or perror? Dump the EP and SB + list for the corresponding epid? */ + if (reg->r_usb_status & IO_MASK(R_USB_STATUS, ourun)) { + panic("USB controller got ourun."); + } + if (reg->r_usb_status & IO_MASK(R_USB_STATUS, perror)) { + + /* Before, etrax_usb_do_intr_recover was called on this epid if it was + an interrupt pipe. I don't see how re-enabling all EP descriptors + will help if there was a programming error. */ + panic("USB controller got perror."); + } + + /* Keep track of USB Controller, if it's running or not */ + if(reg->r_usb_status & IO_STATE(R_USB_STATUS, running, yes)) { + crisv10_hcd->running = 1; + } else { + crisv10_hcd->running = 0; + } + + if (reg->r_usb_status & IO_MASK(R_USB_STATUS, device_mode)) { + /* We should never operate in device mode. */ + panic("USB controller in device mode."); + } + + /* Set the flag to avoid getting "Unlink after no-IRQ? Controller is probably + using the wrong IRQ" from hcd_unlink_urb() in drivers/usb/core/hcd.c */ + set_bit(HCD_FLAG_SAW_IRQ, ®->hcd->flags); + + DBFEXIT; +} + + +/******************************************************************/ +/* Host Controller interface functions */ +/******************************************************************/ + +static inline void crisv10_ready_wait(void) { + volatile int timeout = 10000; + /* Check the busy bit of USB controller in Etrax */ + while((*R_USB_COMMAND & IO_MASK(R_USB_COMMAND, busy)) && + (timeout-- > 0)); +} + +/* reset host controller */ +static int crisv10_hcd_reset(struct usb_hcd *hcd) +{ + DBFENTER; + hcd_dbg(hcd, "reset\n"); + + + /* Reset the USB interface. */ + /* + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, reset); + nop(); + */ + DBFEXIT; + return 0; +} + +/* start host controller */ +static int crisv10_hcd_start(struct usb_hcd *hcd) +{ + DBFENTER; + hcd_dbg(hcd, "start\n"); + + crisv10_ready_wait(); + + /* Start processing of USB traffic. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + + nop(); + + hcd->state = HC_STATE_RUNNING; + + DBFEXIT; + return 0; +} + +/* stop host controller */ +static void crisv10_hcd_stop(struct usb_hcd *hcd) +{ + DBFENTER; + hcd_dbg(hcd, "stop\n"); + crisv10_hcd_reset(hcd); + DBFEXIT; +} + +/* return the current frame number */ +static int crisv10_hcd_get_frame(struct usb_hcd *hcd) +{ + DBFENTER; + DBFEXIT; + return (*R_USB_FM_NUMBER & 0x7ff); +} + +#ifdef CONFIG_USB_OTG + +static int crisv10_hcd_start_port_reset(struct usb_hcd *hcd, unsigned port) +{ + return 0; /* no-op for now */ +} + +#endif /* CONFIG_USB_OTG */ + + +/******************************************************************/ +/* Root Hub functions */ +/******************************************************************/ + +/* root hub status */ +static const struct usb_hub_status rh_hub_status = + { + .wHubStatus = 0, + .wHubChange = 0, + }; + +/* root hub descriptor */ +static const u8 rh_hub_descr[] = + { + 0x09, /* bDescLength */ + 0x29, /* bDescriptorType */ + USB_ROOT_HUB_PORTS, /* bNbrPorts */ + 0x00, /* wHubCharacteristics */ + 0x00, + 0x01, /* bPwrOn2pwrGood */ + 0x00, /* bHubContrCurrent */ + 0x00, /* DeviceRemovable */ + 0xff /* PortPwrCtrlMask */ + }; + +/* Actual holder of root hub status*/ +struct crisv10_rh rh; + +/* Initialize root hub data structures (called from dvdrv_hcd_probe()) */ +int rh_init(void) { + int i; + /* Reset port status flags */ + for (i = 0; i < USB_ROOT_HUB_PORTS; i++) { + rh.wPortChange[i] = 0; + rh.wPortStatusPrev[i] = 0; + } + return 0; +} + +#define RH_FEAT_MASK ((1<lock); + for (i = 1; i <= crisv10_hcd->num_ports; i++) { + if (rh.wPortChange[map_port(i)]) { + *buf |= (1 << i); + rh_dbg("rh_status_data_request, change on port %d: %s Current Status: %s\n", i, + port_status_to_str(rh.wPortChange[map_port(i)]), + port_status_to_str(rh.wPortStatusPrev[map_port(i)])); + } + } + spin_unlock(&crisv10_hcd->lock); + DBFEXIT; + return *buf == 0 ? 0 : 1; +} + +/* Handle a control request for the root hub (called from hcd_driver) */ +static int rh_control_request(struct usb_hcd *hcd, + u16 typeReq, + u16 wValue, + u16 wIndex, + char *buf, + u16 wLength) { + + struct crisv10_hcd *crisv10_hcd = hcd_to_crisv10_hcd(hcd); + int retval = 0; + int len; + DBFENTER; + + switch (typeReq) { + case GetHubDescriptor: + rh_dbg("GetHubDescriptor\n"); + len = min_t(unsigned int, sizeof rh_hub_descr, wLength); + memcpy(buf, rh_hub_descr, len); + buf[2] = crisv10_hcd->num_ports; + break; + case GetHubStatus: + rh_dbg("GetHubStatus\n"); + len = min_t(unsigned int, sizeof rh_hub_status, wLength); + memcpy(buf, &rh_hub_status, len); + break; + case GetPortStatus: + if (!wIndex || wIndex > crisv10_hcd->num_ports) + goto error; + rh_dbg("GetportStatus, port:%d change:%s status:%s\n", wIndex, + port_status_to_str(rh.wPortChange[map_port(wIndex)]), + port_status_to_str(rh.wPortStatusPrev[map_port(wIndex)])); + *(u16 *) buf = cpu_to_le16(rh.wPortStatusPrev[map_port(wIndex)]); + *(u16 *) (buf + 2) = cpu_to_le16(rh.wPortChange[map_port(wIndex)]); + break; + case SetHubFeature: + rh_dbg("SetHubFeature\n"); + case ClearHubFeature: + rh_dbg("ClearHubFeature\n"); + switch (wValue) { + case C_HUB_OVER_CURRENT: + case C_HUB_LOCAL_POWER: + rh_warn("Not implemented hub request:%d \n", typeReq); + /* not implemented */ + break; + default: + goto error; + } + break; + case SetPortFeature: + if (!wIndex || wIndex > crisv10_hcd->num_ports) + goto error; + if(rh_set_port_feature(map_port(wIndex), wValue)) + goto error; + break; + case ClearPortFeature: + if (!wIndex || wIndex > crisv10_hcd->num_ports) + goto error; + if(rh_clear_port_feature(map_port(wIndex), wValue)) + goto error; + break; + default: + rh_warn("Unknown hub request: %d\n", typeReq); + error: + retval = -EPIPE; + } + DBFEXIT; + return retval; +} + +int rh_set_port_feature(__u8 bPort, __u16 wFeature) { + __u8 bUsbCommand = 0; + __u8 reset_cnt; + switch(wFeature) { + case USB_PORT_FEAT_RESET: + rh_dbg("SetPortFeature: reset\n"); + + if (rh.wPortStatusPrev[bPort] & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) + { + __u8 restart_controller = 0; + + if ( (rh.wPortStatusPrev[0] & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) && + (rh.wPortStatusPrev[1] & + IO_STATE(R_USB_RH_PORT_STATUS_2, enabled, yes)) ) + { + /* Both ports is enabled. The USB controller will not change state. */ + restart_controller = 0; + } + else + { + /* Only ports is enabled. The USB controller will change state and + must be restarted. */ + restart_controller = 1; + } + /* + In ETRAX 100LX it's not possible to reset an enabled root hub port. + The workaround is to disable and enable the port before resetting it. + Disabling the port can, if both ports are disabled at once, cause the + USB controller to change state to HOST_MODE state. + The USB controller state transition causes a lot of unwanted + interrupts that must be avoided. + Disabling the USB controller status and port status interrupts before + disabling/resetting the port stops these interrupts. + + These actions are performed: + 1. Disable USB controller status and port status interrupts. + 2. Disable the port + 3. Wait for the port to be disabled. + 4. Enable the port. + 5. Wait for the port to be enabled. + 6. Reset the port. + 7. Wait for for the reset to end. + 8. Wait for the USB controller entering started state. + 9. Order the USB controller to running state. + 10. Wait for the USB controller reaching running state. + 11. Clear all interrupts generated during the disable/enable/reset + procedure. + 12. Enable the USB controller status and port status interrupts. + */ + + /* 1. Disable USB controller status and USB port status interrupts. */ + *R_USB_IRQ_MASK_CLR = IO_STATE(R_USB_IRQ_MASK_CLR, ctl_status, clr); + __asm__ __volatile__ (" nop"); + *R_USB_IRQ_MASK_CLR = IO_STATE(R_USB_IRQ_MASK_CLR, port_status, clr); + __asm__ __volatile__ (" nop"); + + { + + /* Since an root hub port reset shall be 50 ms and the ETRAX 100LX + root hub port reset is 10 ms we must perform 5 port resets to + achieve a proper root hub port reset. */ + for (reset_cnt = 0; reset_cnt < 5; reset_cnt ++) + { + rh_dbg("Disable Port %d\n", bPort + 1); + + /* 2. Disable the port*/ + if (bPort == 0) + { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, yes); + } + else + { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, yes); + } + + /* 3. Wait for the port to be disabled. */ + while ( (bPort == 0) ? + *R_USB_RH_PORT_STATUS_1 & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes) : + *R_USB_RH_PORT_STATUS_2 & + IO_STATE(R_USB_RH_PORT_STATUS_2, enabled, yes) ) {} + + rh_dbg("Port %d is disabled. Enable it!\n", bPort + 1); + + /* 4. Enable the port. */ + if (bPort == 0) + { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, no); + } + else + { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, no); + } + + /* 5. Wait for the port to be enabled again. */ + while (!( (bPort == 0) ? + *R_USB_RH_PORT_STATUS_1 & + IO_STATE(R_USB_RH_PORT_STATUS_1, connected, yes) : + *R_USB_RH_PORT_STATUS_2 & + IO_STATE(R_USB_RH_PORT_STATUS_2, connected, yes) ) ) {} + + rh_dbg("Port %d is enabled.\n", bPort + 1); + + /* 6. Reset the port */ + crisv10_ready_wait(); + *R_USB_COMMAND = + ( (bPort == 0) ? + IO_STATE(R_USB_COMMAND, port_sel, port1): + IO_STATE(R_USB_COMMAND, port_sel, port2) ) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, busy, no) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, nop); + rh_dbg("Port %d is resetting.\n", bPort + 1); + + /* 7. The USB specification says that we should wait for at least + 10ms for device recover */ + udelay(10500); /* 10,5ms blocking wait */ + + crisv10_ready_wait(); + } + } + + + /* Check if the USB controller needs to be restarted. */ + if (restart_controller) + { + /* 8. Wait for the USB controller entering started state. */ + while (!(*R_USB_STATUS & IO_STATE(R_USB_STATUS, started, yes))) {} + + /* 9. Order the USB controller to running state. */ + crisv10_ready_wait(); + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, busy, no) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + + /* 10. Wait for the USB controller reaching running state. */ + while (!(*R_USB_STATUS & IO_STATE(R_USB_STATUS, running, yes))) {} + } + + /* 11. Clear any controller or port satus interrupts before enabling + the interrupts. */ + { + u16 dummy; + + /* Clear the port status interrupt of the reset port. */ + if (bPort == 0) + { + rh_dbg("Clearing port 1 interrupts\n"); + dummy = *R_USB_RH_PORT_STATUS_1; + } + else + { + rh_dbg("Clearing port 2 interrupts\n"); + dummy = *R_USB_RH_PORT_STATUS_2; + } + + if (restart_controller) + { + /* The USB controller is restarted. Clear all interupts. */ + rh_dbg("Clearing all interrupts\n"); + dummy = *R_USB_STATUS; + dummy = *R_USB_RH_PORT_STATUS_1; + dummy = *R_USB_RH_PORT_STATUS_2; + } + } + + /* 12. Enable USB controller status and USB port status interrupts. */ + *R_USB_IRQ_MASK_SET = IO_STATE(R_USB_IRQ_MASK_SET, ctl_status, set); + __asm__ __volatile__ (" nop"); + *R_USB_IRQ_MASK_SET = IO_STATE(R_USB_IRQ_MASK_SET, port_status, set); + __asm__ __volatile__ (" nop"); + + } + else + { + + bUsbCommand |= IO_STATE(R_USB_COMMAND, port_cmd, reset); + /* Select which port via the port_sel field */ + bUsbCommand |= IO_FIELD(R_USB_COMMAND, port_sel, bPort+1); + + /* Make sure the controller isn't busy. */ + crisv10_ready_wait(); + /* Send out the actual command to the USB controller */ + *R_USB_COMMAND = bUsbCommand; + + /* Wait a while for controller to first become started after port reset */ + udelay(12000); /* 12ms blocking wait */ + + /* Make sure the controller isn't busy. */ + crisv10_ready_wait(); + + /* If all enabled ports were disabled the host controller goes down into + started mode, so we need to bring it back into the running state. + (This is safe even if it's already in the running state.) */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + } + + break; + case USB_PORT_FEAT_SUSPEND: + rh_dbg("SetPortFeature: suspend\n"); + bUsbCommand |= IO_STATE(R_USB_COMMAND, port_cmd, suspend); + goto set; + break; + case USB_PORT_FEAT_POWER: + rh_dbg("SetPortFeature: power\n"); + break; + case USB_PORT_FEAT_C_CONNECTION: + rh_dbg("SetPortFeature: c_connection\n"); + break; + case USB_PORT_FEAT_C_RESET: + rh_dbg("SetPortFeature: c_reset\n"); + break; + case USB_PORT_FEAT_C_OVER_CURRENT: + rh_dbg("SetPortFeature: c_over_current\n"); + break; + + set: + /* Select which port via the port_sel field */ + bUsbCommand |= IO_FIELD(R_USB_COMMAND, port_sel, bPort+1); + + /* Make sure the controller isn't busy. */ + crisv10_ready_wait(); + /* Send out the actual command to the USB controller */ + *R_USB_COMMAND = bUsbCommand; + break; + default: + rh_dbg("SetPortFeature: unknown feature\n"); + return -1; + } + return 0; +} + +int rh_clear_port_feature(__u8 bPort, __u16 wFeature) { + switch(wFeature) { + case USB_PORT_FEAT_ENABLE: + rh_dbg("ClearPortFeature: enable\n"); + rh_disable_port(bPort); + break; + case USB_PORT_FEAT_SUSPEND: + rh_dbg("ClearPortFeature: suspend\n"); + break; + case USB_PORT_FEAT_POWER: + rh_dbg("ClearPortFeature: power\n"); + break; + + case USB_PORT_FEAT_C_ENABLE: + rh_dbg("ClearPortFeature: c_enable\n"); + goto clear; + case USB_PORT_FEAT_C_SUSPEND: + rh_dbg("ClearPortFeature: c_suspend\n"); + goto clear; + case USB_PORT_FEAT_C_CONNECTION: + rh_dbg("ClearPortFeature: c_connection\n"); + goto clear; + case USB_PORT_FEAT_C_OVER_CURRENT: + rh_dbg("ClearPortFeature: c_over_current\n"); + goto clear; + case USB_PORT_FEAT_C_RESET: + rh_dbg("ClearPortFeature: c_reset\n"); + goto clear; + clear: + rh.wPortChange[bPort] &= ~(1 << (wFeature - 16)); + break; + default: + rh_dbg("ClearPortFeature: unknown feature\n"); + return -1; + } + return 0; +} + + +#ifdef CONFIG_PM +/* Handle a suspend request for the root hub (called from hcd_driver) */ +static int rh_suspend_request(struct usb_hcd *hcd) +{ + return 0; /* no-op for now */ +} + +/* Handle a resume request for the root hub (called from hcd_driver) */ +static int rh_resume_request(struct usb_hcd *hcd) +{ + return 0; /* no-op for now */ +} +#endif /* CONFIG_PM */ + + + +/* Wrapper function for workaround port disable registers in USB controller */ +static void rh_disable_port(unsigned int port) { + volatile int timeout = 10000; + volatile char* usb_portx_disable; + switch(port) { + case 0: + usb_portx_disable = R_USB_PORT1_DISABLE; + break; + case 1: + usb_portx_disable = R_USB_PORT2_DISABLE; + break; + default: + /* Invalid port index */ + return; + } + /* Set disable flag in special register */ + *usb_portx_disable = IO_STATE(R_USB_PORT1_DISABLE, disable, yes); + /* Wait until not enabled anymore */ + while((rh.wPortStatusPrev[port] & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) && + (timeout-- > 0)); + + /* clear disable flag in special register */ + *usb_portx_disable = IO_STATE(R_USB_PORT1_DISABLE, disable, no); + rh_info("Physical port %d disabled\n", port+1); +} + + +/******************************************************************/ +/* Transfer Controller (TC) functions */ +/******************************************************************/ + +/* FIXME: Should RX_BUF_SIZE be a config option, or maybe we should adjust it + dynamically? + To adjust it dynamically we would have to get an interrupt when we reach + the end of the rx descriptor list, or when we get close to the end, and + then allocate more descriptors. */ +#define NBR_OF_RX_DESC 512 +#define RX_DESC_BUF_SIZE 1024 +#define RX_BUF_SIZE (NBR_OF_RX_DESC * RX_DESC_BUF_SIZE) + + +/* Local variables for Transfer Controller */ +/* --------------------------------------- */ + +/* This is a circular (double-linked) list of the active urbs for each epid. + The head is never removed, and new urbs are linked onto the list as + urb_entry_t elements. Don't reference urb_list directly; use the wrapper + functions instead (which includes spin_locks) */ +static struct list_head urb_list[NBR_OF_EPIDS]; + +/* Read about the need and usage of this lock in submit_ctrl_urb. */ +/* Lock for URB lists for each EPID */ +static spinlock_t urb_list_lock; + +/* Lock for EPID array register (R_USB_EPT_x) in Etrax */ +static spinlock_t etrax_epid_lock; + +/* Lock for dma8 sub0 handling */ +static spinlock_t etrax_dma8_sub0_lock; + +/* DMA IN cache bug. Align the DMA IN buffers to 32 bytes, i.e. a cache line. + Since RX_DESC_BUF_SIZE is 1024 is a multiple of 32, all rx buffers will be + cache aligned. */ +static volatile unsigned char RxBuf[RX_BUF_SIZE] __attribute__ ((aligned (32))); +static volatile struct USB_IN_Desc RxDescList[NBR_OF_RX_DESC] __attribute__ ((aligned (4))); + +/* Pointers into RxDescList. */ +static volatile struct USB_IN_Desc *myNextRxDesc; +static volatile struct USB_IN_Desc *myLastRxDesc; + +/* A zout transfer makes a memory access at the address of its buf pointer, + which means that setting this buf pointer to 0 will cause an access to the + flash. In addition to this, setting sw_len to 0 results in a 16/32 bytes + (depending on DMA burst size) transfer. + Instead, we set it to 1, and point it to this buffer. */ +static int zout_buffer[4] __attribute__ ((aligned (4))); + +/* Cache for allocating new EP and SB descriptors. */ +static struct kmem_cache *usb_desc_cache; + +/* Cache for the data allocated in the isoc descr top half. */ +static struct kmem_cache *isoc_compl_cache; + +/* Cache for the data allocated when delayed finishing of URBs */ +static struct kmem_cache *later_data_cache; + + +/* Counter to keep track of how many Isoc EP we have sat up. Used to enable + and disable iso_eof interrupt. We only need these interrupts when we have + Isoc data endpoints (consumes CPU cycles). + FIXME: This could be more fine granular, so this interrupt is only enabled + when we have a In Isoc URB not URB_ISO_ASAP flaged queued. */ +static int isoc_epid_counter; + +/* Protecting wrapper functions for R_USB_EPT_x */ +/* -------------------------------------------- */ +static inline void etrax_epid_set(__u8 index, __u32 data) { + unsigned long flags; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + *R_USB_EPT_DATA = data; + spin_unlock_irqrestore(&etrax_epid_lock, flags); +} + +static inline void etrax_epid_clear_error(__u8 index) { + unsigned long flags; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + *R_USB_EPT_DATA &= + ~(IO_MASK(R_USB_EPT_DATA, error_count_in) | + IO_MASK(R_USB_EPT_DATA, error_count_out) | + IO_MASK(R_USB_EPT_DATA, error_code)); + spin_unlock_irqrestore(&etrax_epid_lock, flags); +} + +static inline void etrax_epid_set_toggle(__u8 index, __u8 dirout, + __u8 toggle) { + unsigned long flags; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + if(dirout) { + *R_USB_EPT_DATA &= ~IO_MASK(R_USB_EPT_DATA, t_out); + *R_USB_EPT_DATA |= IO_FIELD(R_USB_EPT_DATA, t_out, toggle); + } else { + *R_USB_EPT_DATA &= ~IO_MASK(R_USB_EPT_DATA, t_in); + *R_USB_EPT_DATA |= IO_FIELD(R_USB_EPT_DATA, t_in, toggle); + } + spin_unlock_irqrestore(&etrax_epid_lock, flags); +} + +static inline __u8 etrax_epid_get_toggle(__u8 index, __u8 dirout) { + unsigned long flags; + __u8 toggle; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + if (dirout) { + toggle = IO_EXTRACT(R_USB_EPT_DATA, t_out, *R_USB_EPT_DATA); + } else { + toggle = IO_EXTRACT(R_USB_EPT_DATA, t_in, *R_USB_EPT_DATA); + } + spin_unlock_irqrestore(&etrax_epid_lock, flags); + return toggle; +} + + +static inline __u32 etrax_epid_get(__u8 index) { + unsigned long flags; + __u32 data; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + data = *R_USB_EPT_DATA; + spin_unlock_irqrestore(&etrax_epid_lock, flags); + return data; +} + + + + +/* Main functions for Transfer Controller */ +/* -------------------------------------- */ + +/* Init structs, memories and lists used by Transfer Controller */ +int tc_init(struct usb_hcd *hcd) { + int i; + /* Clear software state info for all epids */ + memset(epid_state, 0, sizeof(struct etrax_epid) * NBR_OF_EPIDS); + + /* Set Invalid and Dummy as being in use and disabled */ + epid_state[INVALID_EPID].inuse = 1; + epid_state[DUMMY_EPID].inuse = 1; + epid_state[INVALID_EPID].disabled = 1; + epid_state[DUMMY_EPID].disabled = 1; + + /* Clear counter for how many Isoc epids we have sat up */ + isoc_epid_counter = 0; + + /* Initialize the urb list by initiating a head for each list. + Also reset list hodling active URB for each epid */ + for (i = 0; i < NBR_OF_EPIDS; i++) { + INIT_LIST_HEAD(&urb_list[i]); + activeUrbList[i] = NULL; + } + + /* Init lock for URB lists */ + spin_lock_init(&urb_list_lock); + /* Init lock for Etrax R_USB_EPT register */ + spin_lock_init(&etrax_epid_lock); + /* Init lock for Etrax dma8 sub0 handling */ + spin_lock_init(&etrax_dma8_sub0_lock); + + /* We use kmem_cache_* to make sure that all DMA desc. are dword aligned */ + + /* Note that we specify sizeof(struct USB_EP_Desc) as the size, but also + allocate SB descriptors from this cache. This is ok since + sizeof(struct USB_EP_Desc) == sizeof(struct USB_SB_Desc). */ + usb_desc_cache = kmem_cache_create("usb_desc_cache", + sizeof(struct USB_EP_Desc), 0, + SLAB_HWCACHE_ALIGN, 0); + if(usb_desc_cache == NULL) { + return -ENOMEM; + } + + /* Create slab cache for speedy allocation of memory for isoc bottom-half + interrupt handling */ + isoc_compl_cache = + kmem_cache_create("isoc_compl_cache", + sizeof(struct crisv10_isoc_complete_data), + 0, SLAB_HWCACHE_ALIGN, 0); + if(isoc_compl_cache == NULL) { + return -ENOMEM; + } + + /* Create slab cache for speedy allocation of memory for later URB finish + struct */ + later_data_cache = + kmem_cache_create("later_data_cache", + sizeof(struct urb_later_data), + 0, SLAB_HWCACHE_ALIGN, 0); + if(later_data_cache == NULL) { + return -ENOMEM; + } + + + /* Initiate the bulk start timer. */ + init_timer(&bulk_start_timer); + bulk_start_timer.expires = jiffies + BULK_START_TIMER_INTERVAL; + bulk_start_timer.function = tc_bulk_start_timer_func; + add_timer(&bulk_start_timer); + + + /* Initiate the bulk eot timer. */ + init_timer(&bulk_eot_timer); + bulk_eot_timer.expires = jiffies + BULK_EOT_TIMER_INTERVAL; + bulk_eot_timer.function = tc_bulk_eot_timer_func; + bulk_eot_timer.data = (unsigned long)hcd; + add_timer(&bulk_eot_timer); + + return 0; +} + +/* Uninitialize all resources used by Transfer Controller */ +void tc_destroy(void) { + + /* Destroy all slab cache */ + kmem_cache_destroy(usb_desc_cache); + kmem_cache_destroy(isoc_compl_cache); + kmem_cache_destroy(later_data_cache); + + /* Remove timers */ + del_timer(&bulk_start_timer); + del_timer(&bulk_eot_timer); +} + +static void restart_dma8_sub0(void) { + unsigned long flags; + spin_lock_irqsave(&etrax_dma8_sub0_lock, flags); + /* Verify that the dma is not running */ + if ((*R_DMA_CH8_SUB0_CMD & IO_MASK(R_DMA_CH8_SUB0_CMD, cmd)) == 0) { + struct USB_EP_Desc *ep = (struct USB_EP_Desc *)phys_to_virt(*R_DMA_CH8_SUB0_EP); + while (DUMMY_EPID == IO_EXTRACT(USB_EP_command, epid, ep->command)) { + ep = (struct USB_EP_Desc *)phys_to_virt(ep->next); + } + /* Advance the DMA to the next EP descriptor that is not a DUMMY_EPID. */ + *R_DMA_CH8_SUB0_EP = virt_to_phys(ep); + /* Restart the DMA */ + *R_DMA_CH8_SUB0_CMD = IO_STATE(R_DMA_CH8_SUB0_CMD, cmd, start); + } + spin_unlock_irqrestore(&etrax_dma8_sub0_lock, flags); +} + +/* queue an URB with the transfer controller (called from hcd_driver) */ +static int tc_urb_enqueue(struct usb_hcd *hcd, + struct urb *urb, + gfp_t mem_flags) { + int epid; + int retval; + int bustime = 0; + int maxpacket; + unsigned long flags; + struct crisv10_urb_priv *urb_priv; + struct crisv10_hcd* crisv10_hcd = hcd_to_crisv10_hcd(hcd); + DBFENTER; + + if(!(crisv10_hcd->running)) { + /* The USB Controller is not running, probably because no device is + attached. No idea to enqueue URBs then */ + tc_warn("Rejected enqueueing of URB:0x%x because no dev attached\n", + (unsigned int)urb); + return -ENOENT; + } + + maxpacket = usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe)); + /* Special case check for In Isoc transfers. Specification states that each + In Isoc transfer consists of one packet and therefore it should fit into + the transfer-buffer of an URB. + We do the check here to be sure (an invalid scenario can be produced with + parameters to the usbtest suite) */ + if(usb_pipeisoc(urb->pipe) && usb_pipein(urb->pipe) && + (urb->transfer_buffer_length < maxpacket)) { + tc_err("Submit In Isoc URB with buffer length:%d to pipe with maxpacketlen: %d\n", urb->transfer_buffer_length, maxpacket); + return -EMSGSIZE; + } + + /* Check if there is a epid for URBs destination, if not this function + set up one. */ + epid = tc_setup_epid(urb->ep, urb, mem_flags); + if (epid < 0) { + tc_err("Failed setup epid:%d for URB:0x%x\n", epid, (unsigned int)urb); + DBFEXIT; + return -ENOMEM; + } + + if(urb == activeUrbList[epid]) { + tc_err("Resubmition of allready active URB:0x%x\n", (unsigned int)urb); + return -ENXIO; + } + + if(urb_list_entry(urb, epid)) { + tc_err("Resubmition of allready queued URB:0x%x\n", (unsigned int)urb); + return -ENXIO; + } + + /* If we actively have flaged endpoint as disabled then refuse submition */ + if(epid_state[epid].disabled) { + return -ENOENT; + } + + /* Allocate and init HC-private data for URB */ + if(urb_priv_create(hcd, urb, epid, mem_flags) != 0) { + DBFEXIT; + return -ENOMEM; + } + urb_priv = urb->hcpriv; + + /* Check if there is enough bandwidth for periodic transfer */ + if(usb_pipeint(urb->pipe) || usb_pipeisoc(urb->pipe)) { + /* only check (and later claim) if not already claimed */ + if (urb_priv->bandwidth == 0) { + bustime = crisv10_usb_check_bandwidth(urb->dev, urb); + if (bustime < 0) { + tc_err("Not enough periodic bandwidth\n"); + urb_priv_free(hcd, urb); + DBFEXIT; + return -ENOSPC; + } + } + } + + tc_dbg("Enqueue URB:0x%x[%d] epid:%d (%s) bufflen:%d\n", + (unsigned int)urb, urb_priv->urb_num, epid, + pipe_to_str(urb->pipe), urb->transfer_buffer_length); + + /* Create and link SBs required for this URB */ + retval = create_sb_for_urb(urb, mem_flags); + if(retval != 0) { + tc_err("Failed to create SBs for URB:0x%x[%d]\n", (unsigned int)urb, + urb_priv->urb_num); + urb_priv_free(hcd, urb); + DBFEXIT; + return retval; + } + + /* Init intr EP pool if this URB is a INTR transfer. This pool is later + used when inserting EPs in the TxIntrEPList. We do the alloc here + so we can't run out of memory later */ + if(usb_pipeint(urb->pipe)) { + retval = init_intr_urb(urb, mem_flags); + if(retval != 0) { + tc_warn("Failed to init Intr URB\n"); + urb_priv_free(hcd, urb); + DBFEXIT; + return retval; + } + } + + /* Disable other access when inserting USB */ + local_irq_save(flags); + + /* Claim bandwidth, if needed */ + if(bustime) { + crisv10_usb_claim_bandwidth(urb->dev, + urb, + bustime, + (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS)); + } + + /* Add URB to EP queue */ + urb_list_add(urb, epid, mem_flags); + + if(usb_pipeisoc(urb->pipe)) { + /* Special processing of Isoc URBs. */ + tc_dma_process_isoc_urb(urb); + } else { + /* Process EP queue for rest of the URB types (Bulk, Ctrl, Intr) */ + tc_dma_process_queue(epid); + } + + local_irq_restore(flags); + + DBFEXIT; + return 0; +} + +/* remove an URB from the transfer controller queues (called from hcd_driver)*/ +static int tc_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status) { + struct crisv10_urb_priv *urb_priv; + unsigned long flags; + int epid; + + DBFENTER; + /* Disable interrupts here since a descriptor interrupt for the isoc epid + will modify the sb list. This could possibly be done more granular, but + urb_dequeue should not be used frequently anyway. + */ + local_irq_save(flags); + + urb->status = status; + urb_priv = urb->hcpriv; + + if (!urb_priv) { + /* This happens if a device driver calls unlink on an urb that + was never submitted (lazy driver) or if the urb was completed + while dequeue was being called. */ + tc_warn("Dequeing of not enqueued URB:0x%x\n", (unsigned int)urb); + local_irq_restore(flags); + return 0; + } + epid = urb_priv->epid; + + tc_warn("Dequeing %s URB:0x%x[%d] (%s %s epid:%d) status:%d %s\n", + (urb == activeUrbList[epid]) ? "active" : "queued", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), epid, urb->status, + (urb_priv->later_data) ? "later-sched" : ""); + + /* For Bulk, Ctrl and Intr are only one URB active at a time. So any URB + that isn't active can be dequeued by just removing it from the queue */ + if(usb_pipebulk(urb->pipe) || usb_pipecontrol(urb->pipe) || + usb_pipeint(urb->pipe)) { + + /* Check if URB haven't gone further than the queue */ + if(urb != activeUrbList[epid]) { + ASSERT(urb_priv->later_data == NULL); + tc_warn("Dequeing URB:0x%x[%d] (%s %s epid:%d) from queue" + " (not active)\n", (unsigned int)urb, urb_priv->urb_num, + str_dir(urb->pipe), str_type(urb->pipe), epid); + + /* Finish the URB with error status from USB core */ + tc_finish_urb(hcd, urb, urb->status); + local_irq_restore(flags); + return 0; + } + } + + /* Set URB status to Unlink for handling when interrupt comes. */ + urb_priv->urb_state = UNLINK; + + /* Differentiate dequeing of Bulk and Ctrl from Isoc and Intr */ + switch(usb_pipetype(urb->pipe)) { + case PIPE_BULK: + /* Check if EP still is enabled */ + if (TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxBulkEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + } + /* Kicking dummy list out of the party. */ + TxBulkEPList[epid].next = virt_to_phys(&TxBulkEPList[(epid + 1) % NBR_OF_EPIDS]); + break; + case PIPE_CONTROL: + /* Check if EP still is enabled */ + if (TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxCtrlEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + } + break; + case PIPE_ISOCHRONOUS: + /* Disabling, busy-wait and unlinking of Isoc SBs will be done in + finish_isoc_urb(). Because there might the case when URB is dequeued + but there are other valid URBs waiting */ + + /* Check if In Isoc EP still is enabled */ + if (TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxIsocEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + } + break; + case PIPE_INTERRUPT: + /* Special care is taken for interrupt URBs. EPs are unlinked in + tc_finish_urb */ + break; + default: + break; + } + + /* Asynchronous unlink, finish the URB later from scheduled or other + event (data finished, error) */ + tc_finish_urb_later(hcd, urb, urb->status); + + local_irq_restore(flags); + DBFEXIT; + return 0; +} + + +static void tc_sync_finish_epid(struct usb_hcd *hcd, int epid) { + volatile int timeout = 10000; + struct urb* urb; + struct crisv10_urb_priv* urb_priv; + unsigned long flags; + + volatile struct USB_EP_Desc *first_ep; /* First EP in the list. */ + volatile struct USB_EP_Desc *curr_ep; /* Current EP, the iterator. */ + volatile struct USB_EP_Desc *next_ep; /* The EP after current. */ + + int type = epid_state[epid].type; + + /* Setting this flag will cause enqueue() to return -ENOENT for new + submitions on this endpoint and finish_urb() wont process queue further */ + epid_state[epid].disabled = 1; + + switch(type) { + case PIPE_BULK: + /* Check if EP still is enabled */ + if (TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxBulkEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + tc_warn("sync_finish: Disabling EP for epid:%d\n", epid); + + /* Do busy-wait until DMA not using this EP descriptor anymore */ + while((*R_DMA_CH8_SUB0_EP == + virt_to_phys(&TxBulkEPList[epid])) && + (timeout-- > 0)); + + } + break; + + case PIPE_CONTROL: + /* Check if EP still is enabled */ + if (TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxCtrlEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + tc_warn("sync_finish: Disabling EP for epid:%d\n", epid); + + /* Do busy-wait until DMA not using this EP descriptor anymore */ + while((*R_DMA_CH8_SUB1_EP == + virt_to_phys(&TxCtrlEPList[epid])) && + (timeout-- > 0)); + } + break; + + case PIPE_INTERRUPT: + local_irq_save(flags); + /* Disable all Intr EPs belonging to epid */ + first_ep = &TxIntrEPList[0]; + curr_ep = first_ep; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if (IO_EXTRACT(USB_EP_command, epid, next_ep->command) == epid) { + /* Disable EP */ + next_ep->command &= ~IO_MASK(USB_EP_command, enable); + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != first_ep); + + local_irq_restore(flags); + break; + + case PIPE_ISOCHRONOUS: + /* Check if EP still is enabled */ + if (TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + tc_warn("sync_finish: Disabling Isoc EP for epid:%d\n", epid); + /* The EP was enabled, disable it. */ + TxIsocEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + + while((*R_DMA_CH8_SUB3_EP == virt_to_phys(&TxIsocEPList[epid])) && + (timeout-- > 0)); + } + break; + } + + local_irq_save(flags); + + /* Finish if there is active URB for this endpoint */ + if(activeUrbList[epid] != NULL) { + urb = activeUrbList[epid]; + urb_priv = urb->hcpriv; + ASSERT(urb_priv); + tc_warn("Sync finish %s URB:0x%x[%d] (%s %s epid:%d) status:%d %s\n", + (urb == activeUrbList[epid]) ? "active" : "queued", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), epid, urb->status, + (urb_priv->later_data) ? "later-sched" : ""); + + tc_finish_urb(hcd, activeUrbList[epid], -ENOENT); + ASSERT(activeUrbList[epid] == NULL); + } + + /* Finish any queued URBs for this endpoint. There won't be any resubmitions + because epid_disabled causes enqueue() to fail for this endpoint */ + while((urb = urb_list_first(epid)) != NULL) { + urb_priv = urb->hcpriv; + ASSERT(urb_priv); + + tc_warn("Sync finish %s URB:0x%x[%d] (%s %s epid:%d) status:%d %s\n", + (urb == activeUrbList[epid]) ? "active" : "queued", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), epid, urb->status, + (urb_priv->later_data) ? "later-sched" : ""); + + tc_finish_urb(hcd, urb, -ENOENT); + } + epid_state[epid].disabled = 0; + local_irq_restore(flags); +} + +/* free resources associated with an endpoint (called from hcd_driver) */ +static void tc_endpoint_disable(struct usb_hcd *hcd, + struct usb_host_endpoint *ep) { + DBFENTER; + /* Only free epid if it has been allocated. We get two endpoint_disable + requests for ctrl endpoints so ignore the second one */ + if(ep->hcpriv != NULL) { + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + int epid = ep_priv->epid; + tc_warn("endpoint_disable ep:0x%x ep-priv:0x%x (%s) (epid:%d freed)\n", + (unsigned int)ep, (unsigned int)ep->hcpriv, + endpoint_to_str(&(ep->desc)), epid); + + tc_sync_finish_epid(hcd, epid); + + ASSERT(activeUrbList[epid] == NULL); + ASSERT(list_empty(&urb_list[epid])); + + tc_free_epid(ep); + } else { + tc_dbg("endpoint_disable ep:0x%x ep-priv:0x%x (%s)\n", (unsigned int)ep, + (unsigned int)ep->hcpriv, endpoint_to_str(&(ep->desc))); + } + DBFEXIT; +} + +static void tc_finish_urb_later_proc(struct work_struct* work) { + unsigned long flags; + struct urb_later_data* uld; + + local_irq_save(flags); + uld = container_of(work, struct urb_later_data, dws.work); + if(uld->urb == NULL) { + late_dbg("Later finish of URB = NULL (allready finished)\n"); + } else { + struct crisv10_urb_priv* urb_priv = uld->urb->hcpriv; + ASSERT(urb_priv); + if(urb_priv->urb_num == uld->urb_num) { + late_dbg("Later finish of URB:0x%x[%d]\n", (unsigned int)(uld->urb), + urb_priv->urb_num); + if(uld->status != uld->urb->status) { + errno_dbg("Later-finish URB with status:%d, later-status:%d\n", + uld->urb->status, uld->status); + } + if(uld != urb_priv->later_data) { + panic("Scheduled uld not same as URBs uld\n"); + } + tc_finish_urb(uld->hcd, uld->urb, uld->status); + } else { + late_warn("Ignoring later finish of URB:0x%x[%d]" + ", urb_num doesn't match current URB:0x%x[%d]", + (unsigned int)(uld->urb), uld->urb_num, + (unsigned int)(uld->urb), urb_priv->urb_num); + } + } + local_irq_restore(flags); + kmem_cache_free(later_data_cache, uld); +} + +static void tc_finish_urb_later(struct usb_hcd *hcd, struct urb *urb, + int status) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + struct urb_later_data* uld; + + ASSERT(urb_priv); + + if(urb_priv->later_data != NULL) { + /* Later-finish allready scheduled for this URB, just update status to + return when finishing later */ + errno_dbg("Later-finish schedule change URB status:%d with new" + " status:%d\n", urb_priv->later_data->status, status); + + urb_priv->later_data->status = status; + return; + } + + uld = kmem_cache_alloc(later_data_cache, GFP_ATOMIC); + ASSERT(uld); + + uld->hcd = hcd; + uld->urb = urb; + uld->urb_num = urb_priv->urb_num; + uld->status = status; + + INIT_DELAYED_WORK(&uld->dws, tc_finish_urb_later_proc); + urb_priv->later_data = uld; + + /* Schedule the finishing of the URB to happen later */ + schedule_delayed_work(&uld->dws, LATER_TIMER_DELAY); +} + +static void tc_finish_isoc_urb(struct usb_hcd *hcd, struct urb *urb, + int status); + +static void tc_finish_urb(struct usb_hcd *hcd, struct urb *urb, int status) { + struct crisv10_hcd* crisv10_hcd = hcd_to_crisv10_hcd(hcd); + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + int epid; + char toggle; + int urb_num; + + DBFENTER; + ASSERT(urb_priv != NULL); + epid = urb_priv->epid; + urb_num = urb_priv->urb_num; + + if(urb != activeUrbList[epid]) { + if(urb_list_entry(urb, epid)) { + /* Remove this URB from the list. Only happens when URB are finished + before having been processed (dequeing) */ + urb_list_del(urb, epid); + } else { + tc_warn("Finishing of URB:0x%x[%d] neither active or in queue for" + " epid:%d\n", (unsigned int)urb, urb_num, epid); + } + } + + /* Cancel any pending later-finish of this URB */ + if(urb_priv->later_data) { + urb_priv->later_data->urb = NULL; + } + + /* For an IN pipe, we always set the actual length, regardless of whether + there was an error or not (which means the device driver can use the data + if it wants to). */ + if(usb_pipein(urb->pipe)) { + urb->actual_length = urb_priv->rx_offset; + } else { + /* Set actual_length for OUT urbs also; the USB mass storage driver seems + to want that. */ + if (status == 0 && urb->status == -EINPROGRESS) { + urb->actual_length = urb->transfer_buffer_length; + } else { + /* We wouldn't know of any partial writes if there was an error. */ + urb->actual_length = 0; + } + } + + + /* URB status mangling */ + if(urb->status == -EINPROGRESS) { + /* The USB core hasn't changed the status, let's set our finish status */ + urb->status = status; + + if ((status == 0) && (urb->transfer_flags & URB_SHORT_NOT_OK) && + usb_pipein(urb->pipe) && + (urb->actual_length != urb->transfer_buffer_length)) { + /* URB_SHORT_NOT_OK means that short reads (shorter than the endpoint's + max length) is to be treated as an error. */ + errno_dbg("Finishing URB:0x%x[%d] with SHORT_NOT_OK flag and short" + " data:%d\n", (unsigned int)urb, urb_num, + urb->actual_length); + urb->status = -EREMOTEIO; + } + + if(urb_priv->urb_state == UNLINK) { + /* URB has been requested to be unlinked asynchronously */ + urb->status = -ECONNRESET; + errno_dbg("Fixing unlink status of URB:0x%x[%d] to:%d\n", + (unsigned int)urb, urb_num, urb->status); + } + } else { + /* The USB Core wants to signal some error via the URB, pass it through */ + } + + /* use completely different finish function for Isoc URBs */ + if(usb_pipeisoc(urb->pipe)) { + tc_finish_isoc_urb(hcd, urb, status); + return; + } + + /* Do special unlinking of EPs for Intr traffic */ + if(usb_pipeint(urb->pipe)) { + tc_dma_unlink_intr_urb(urb); + } + + /* Release allocated bandwidth for periodic transfers */ + if(usb_pipeint(urb->pipe) || usb_pipeisoc(urb->pipe)) + crisv10_usb_release_bandwidth(hcd, + usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS, + urb_priv->bandwidth); + + /* This URB is active on EP */ + if(urb == activeUrbList[epid]) { + /* We need to fiddle with the toggle bits because the hardware doesn't do + it for us. */ + toggle = etrax_epid_get_toggle(epid, usb_pipeout(urb->pipe)); + usb_settoggle(urb->dev, usb_pipeendpoint(urb->pipe), + usb_pipeout(urb->pipe), toggle); + + /* Checks for Ctrl and Bulk EPs */ + switch(usb_pipetype(urb->pipe)) { + case PIPE_BULK: + /* Check so Bulk EP realy is disabled before finishing active URB */ + ASSERT((TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) == + IO_STATE(USB_EP_command, enable, no)); + /* Disable sub-pointer for EP to avoid next tx_interrupt() to + process Bulk EP. */ + TxBulkEPList[epid].sub = 0; + /* No need to wait for the DMA before changing the next pointer. + The modulo NBR_OF_EPIDS isn't actually necessary, since we will never use + the last one (INVALID_EPID) for actual traffic. */ + TxBulkEPList[epid].next = + virt_to_phys(&TxBulkEPList[(epid + 1) % NBR_OF_EPIDS]); + break; + case PIPE_CONTROL: + /* Check so Ctrl EP realy is disabled before finishing active URB */ + ASSERT((TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) == + IO_STATE(USB_EP_command, enable, no)); + /* Disable sub-pointer for EP to avoid next tx_interrupt() to + process Ctrl EP. */ + TxCtrlEPList[epid].sub = 0; + break; + } + } + + /* Free HC-private URB data*/ + urb_priv_free(hcd, urb); + + if(urb->status) { + errno_dbg("finish_urb (URB:0x%x[%d] %s %s) (data:%d) status:%d\n", + (unsigned int)urb, urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb->actual_length, urb->status); + } else { + tc_dbg("finish_urb (URB:0x%x[%d] %s %s) (data:%d) status:%d\n", + (unsigned int)urb, urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb->actual_length, urb->status); + } + + /* If we just finished an active URB, clear active pointer. */ + if (urb == activeUrbList[epid]) { + /* Make URB not active on EP anymore */ + activeUrbList[epid] = NULL; + + if(urb->status == 0) { + /* URB finished sucessfully, process queue to see if there are any more + URBs waiting before we call completion function.*/ + if(crisv10_hcd->running) { + /* Only process queue if USB controller is running */ + tc_dma_process_queue(epid); + } else { + tc_warn("No processing of queue for epid:%d, USB Controller not" + " running\n", epid); + } + } + } + + /* Hand the URB from HCD to its USB device driver, using its completion + functions */ + usb_hcd_giveback_urb (hcd, urb, status); + + /* Check the queue once more if the URB returned with error, because we + didn't do it before the completion function because the specification + states that the queue should not restart until all it's unlinked + URBs have been fully retired, with the completion functions run */ + if(crisv10_hcd->running) { + /* Only process queue if USB controller is running */ + tc_dma_process_queue(epid); + } else { + tc_warn("No processing of queue for epid:%d, USB Controller not running\n", + epid); + } + + DBFEXIT; +} + +static void tc_finish_isoc_urb(struct usb_hcd *hcd, struct urb *urb, + int status) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + int epid, i; + volatile int timeout = 10000; + int bandwidth = 0; + + ASSERT(urb_priv); + epid = urb_priv->epid; + + ASSERT(usb_pipeisoc(urb->pipe)); + + /* Set that all isoc packets have status and length set before + completing the urb. */ + for (i = urb_priv->isoc_packet_counter; i < urb->number_of_packets; i++){ + urb->iso_frame_desc[i].actual_length = 0; + urb->iso_frame_desc[i].status = -EPROTO; + } + + /* Check if the URB is currently active (done or error) */ + if(urb == activeUrbList[epid]) { + /* Check if there are another In Isoc URB queued for this epid */ + if (!list_empty(&urb_list[epid])&& !epid_state[epid].disabled) { + /* Move it from queue to active and mark it started so Isoc transfers + won't be interrupted. + All Isoc URBs data transfers are already added to DMA lists so we + don't have to insert anything in DMA lists here. */ + activeUrbList[epid] = urb_list_first(epid); + ((struct crisv10_urb_priv *)(activeUrbList[epid]->hcpriv))->urb_state = + STARTED; + urb_list_del(activeUrbList[epid], epid); + + if(urb->status) { + errno_dbg("finish_isoc_urb (URB:0x%x[%d] %s %s) (%d of %d packets)" + " status:%d, new waiting URB:0x%x[%d]\n", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb_priv->isoc_packet_counter, + urb->number_of_packets, urb->status, + (unsigned int)activeUrbList[epid], + ((struct crisv10_urb_priv *)(activeUrbList[epid]->hcpriv))->urb_num); + } + + } else { /* No other URB queued for this epid */ + if(urb->status) { + errno_dbg("finish_isoc_urb (URB:0x%x[%d] %s %s) (%d of %d packets)" + " status:%d, no new URB waiting\n", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb_priv->isoc_packet_counter, + urb->number_of_packets, urb->status); + } + + /* Check if EP is still enabled, then shut it down. */ + if (TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + isoc_dbg("Isoc EP enabled for epid:%d, disabling it\n", epid); + + /* Should only occur for In Isoc EPs where SB isn't consumed. */ + ASSERT(usb_pipein(urb->pipe)); + + /* Disable it and wait for it to stop */ + TxIsocEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + + /* Ah, the luxury of busy-wait. */ + while((*R_DMA_CH8_SUB3_EP == virt_to_phys(&TxIsocEPList[epid])) && + (timeout-- > 0)); + } + + /* Unlink SB to say that epid is finished. */ + TxIsocEPList[epid].sub = 0; + TxIsocEPList[epid].hw_len = 0; + + /* No URB active for EP anymore */ + activeUrbList[epid] = NULL; + } + } else { /* Finishing of not active URB (queued up with SBs thought) */ + isoc_warn("finish_isoc_urb (URB:0x%x %s) (%d of %d packets) status:%d," + " SB queued but not active\n", + (unsigned int)urb, str_dir(urb->pipe), + urb_priv->isoc_packet_counter, urb->number_of_packets, + urb->status); + if(usb_pipeout(urb->pipe)) { + /* Finishing of not yet active Out Isoc URB needs unlinking of SBs. */ + struct USB_SB_Desc *iter_sb, *prev_sb, *next_sb; + + iter_sb = TxIsocEPList[epid].sub ? + phys_to_virt(TxIsocEPList[epid].sub) : 0; + prev_sb = 0; + + /* SB that is linked before this URBs first SB */ + while (iter_sb && (iter_sb != urb_priv->first_sb)) { + prev_sb = iter_sb; + iter_sb = iter_sb->next ? phys_to_virt(iter_sb->next) : 0; + } + + if (iter_sb == 0) { + /* Unlink of the URB currently being transmitted. */ + prev_sb = 0; + iter_sb = TxIsocEPList[epid].sub ? phys_to_virt(TxIsocEPList[epid].sub) : 0; + } + + while (iter_sb && (iter_sb != urb_priv->last_sb)) { + iter_sb = iter_sb->next ? phys_to_virt(iter_sb->next) : 0; + } + + if (iter_sb) { + next_sb = iter_sb->next ? phys_to_virt(iter_sb->next) : 0; + } else { + /* This should only happen if the DMA has completed + processing the SB list for this EP while interrupts + are disabled. */ + isoc_dbg("Isoc urb not found, already sent?\n"); + next_sb = 0; + } + if (prev_sb) { + prev_sb->next = next_sb ? virt_to_phys(next_sb) : 0; + } else { + TxIsocEPList[epid].sub = next_sb ? virt_to_phys(next_sb) : 0; + } + } + } + + /* Free HC-private URB data*/ + bandwidth = urb_priv->bandwidth; + urb_priv_free(hcd, urb); + + crisv10_usb_release_bandwidth(hcd, usb_pipeisoc(urb->pipe), bandwidth); + + /* Hand the URB from HCD to its USB device driver, using its completion + functions */ + usb_hcd_giveback_urb (hcd, urb, status); +} + +static __u32 urb_num = 0; + +/* allocate and initialize URB private data */ +static int urb_priv_create(struct usb_hcd *hcd, struct urb *urb, int epid, + int mem_flags) { + struct crisv10_urb_priv *urb_priv; + + urb_priv = kmalloc(sizeof *urb_priv, mem_flags); + if (!urb_priv) + return -ENOMEM; + memset(urb_priv, 0, sizeof *urb_priv); + + urb_priv->epid = epid; + urb_priv->urb_state = NOT_STARTED; + + urb->hcpriv = urb_priv; + /* Assign URB a sequence number, and increment counter */ + urb_priv->urb_num = urb_num; + urb_num++; + urb_priv->bandwidth = 0; + return 0; +} + +/* free URB private data */ +static void urb_priv_free(struct usb_hcd *hcd, struct urb *urb) { + int i; + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + ASSERT(urb_priv != 0); + + /* Check it has any SBs linked that needs to be freed*/ + if(urb_priv->first_sb != NULL) { + struct USB_SB_Desc *next_sb, *first_sb, *last_sb; + int i = 0; + first_sb = urb_priv->first_sb; + last_sb = urb_priv->last_sb; + ASSERT(last_sb); + while(first_sb != last_sb) { + next_sb = (struct USB_SB_Desc *)phys_to_virt(first_sb->next); + kmem_cache_free(usb_desc_cache, first_sb); + first_sb = next_sb; + i++; + } + kmem_cache_free(usb_desc_cache, last_sb); + i++; + } + + /* Check if it has any EPs in its Intr pool that also needs to be freed */ + if(urb_priv->intr_ep_pool_length > 0) { + for(i = 0; i < urb_priv->intr_ep_pool_length; i++) { + kfree(urb_priv->intr_ep_pool[i]); + } + /* + tc_dbg("Freed %d EPs from URB:0x%x EP pool\n", + urb_priv->intr_ep_pool_length, (unsigned int)urb); + */ + } + + kfree(urb_priv); + urb->hcpriv = NULL; +} + +static int ep_priv_create(struct usb_host_endpoint *ep, int mem_flags) { + struct crisv10_ep_priv *ep_priv; + + ep_priv = kmalloc(sizeof *ep_priv, mem_flags); + if (!ep_priv) + return -ENOMEM; + memset(ep_priv, 0, sizeof *ep_priv); + + ep->hcpriv = ep_priv; + return 0; +} + +static void ep_priv_free(struct usb_host_endpoint *ep) { + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + ASSERT(ep_priv); + kfree(ep_priv); + ep->hcpriv = NULL; +} + +/* + * usb_check_bandwidth(): + * + * old_alloc is from host_controller->bandwidth_allocated in microseconds; + * bustime is from calc_bus_time(), but converted to microseconds. + * + * returns if successful, + * or -ENOSPC if bandwidth request fails. + * + * FIXME: + * This initial implementation does not use Endpoint.bInterval + * in managing bandwidth allocation. + * It probably needs to be expanded to use Endpoint.bInterval. + * This can be done as a later enhancement (correction). + * + * This will also probably require some kind of + * frame allocation tracking...meaning, for example, + * that if multiple drivers request interrupts every 10 USB frames, + * they don't all have to be allocated at + * frame numbers N, N+10, N+20, etc. Some of them could be at + * N+11, N+21, N+31, etc., and others at + * N+12, N+22, N+32, etc. + * + * Similarly for isochronous transfers... + * + * Individual HCDs can schedule more directly ... this logic + * is not correct for high speed transfers. + */ +static int crisv10_usb_check_bandwidth( + struct usb_device *dev, + struct urb *urb) +{ + unsigned int pipe = urb->pipe; + long bustime; + int is_in = usb_pipein (pipe); + int is_iso = usb_pipeisoc (pipe); + int old_alloc = dev->bus->bandwidth_allocated; + int new_alloc; + + bustime = NS_TO_US (usb_calc_bus_time (dev->speed, is_in, is_iso, + usb_maxpacket (dev, pipe, !is_in))); + if (is_iso) + bustime /= urb->number_of_packets; + + new_alloc = old_alloc + (int) bustime; + if (new_alloc > FRAME_TIME_MAX_USECS_ALLOC) { + dev_dbg (&dev->dev, "usb_check_bandwidth FAILED: %d + %ld = %d usec\n", + old_alloc, bustime, new_alloc); + bustime = -ENOSPC; /* report error */ + } + + return bustime; +} + +/** + * usb_claim_bandwidth - records bandwidth for a periodic transfer + * @dev: source/target of request + * @urb: request (urb->dev == dev) + * @bustime: bandwidth consumed, in (average) microseconds per frame + * @isoc: true iff the request is isochronous + * + * HCDs are expected not to overcommit periodic bandwidth, and to record such + * reservations whenever endpoints are added to the periodic schedule. + * + * FIXME averaging per-frame is suboptimal. Better to sum over the HCD's + * entire periodic schedule ... 32 frames for OHCI, 1024 for UHCI, settable + * for EHCI (256/512/1024 frames, default 1024) and have the bus expose how + * large its periodic schedule is. + */ +static void crisv10_usb_claim_bandwidth( + struct usb_device *dev, + struct urb *urb, int bustime, int isoc) +{ + dev->bus->bandwidth_allocated += bustime; + if (isoc) + dev->bus->bandwidth_isoc_reqs++; + else + dev->bus->bandwidth_int_reqs++; + struct crisv10_urb_priv *urb_priv; + urb_priv = urb->hcpriv; + urb_priv->bandwidth = bustime; +} + +/** + * usb_release_bandwidth - reverses effect of usb_claim_bandwidth() + * @hcd: host controller + * @isoc: true iff the request is isochronous + * @bandwidth: bandwidth returned + * + * This records that previously allocated bandwidth has been released. + * Bandwidth is released when endpoints are removed from the host controller's + * periodic schedule. + */ +static void crisv10_usb_release_bandwidth( + struct usb_hcd *hcd, + int isoc, + int bandwidth) +{ + hcd_to_bus(hcd)->bandwidth_allocated -= bandwidth; + if (isoc) + hcd_to_bus(hcd)->bandwidth_isoc_reqs--; + else + hcd_to_bus(hcd)->bandwidth_int_reqs--; +} + + +/* EPID handling functions, managing EP-list in Etrax through wrappers */ +/* ------------------------------------------------------------------- */ + +/* Sets up a new EPID for an endpoint or returns existing if found */ +static int tc_setup_epid(struct usb_host_endpoint *ep, struct urb *urb, + int mem_flags) { + int epid; + char devnum, endpoint, out_traffic, slow; + int maxlen; + __u32 epid_data; + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + + DBFENTER; + + /* Check if a valid epid already is setup for this endpoint */ + if(ep_priv != NULL) { + return ep_priv->epid; + } + + /* We must find and initiate a new epid for this urb. */ + epid = tc_allocate_epid(); + + if (epid == -1) { + /* Failed to allocate a new epid. */ + DBFEXIT; + return epid; + } + + /* We now have a new epid to use. Claim it. */ + epid_state[epid].inuse = 1; + + /* Init private data for new endpoint */ + if(ep_priv_create(ep, mem_flags) != 0) { + return -ENOMEM; + } + ep_priv = ep->hcpriv; + ep_priv->epid = epid; + + devnum = usb_pipedevice(urb->pipe); + endpoint = usb_pipeendpoint(urb->pipe); + slow = (urb->dev->speed == USB_SPEED_LOW); + maxlen = usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe)); + + if (usb_pipetype(urb->pipe) == PIPE_CONTROL) { + /* We want both IN and OUT control traffic to be put on the same + EP/SB list. */ + out_traffic = 1; + } else { + out_traffic = usb_pipeout(urb->pipe); + } + + if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) { + epid_data = IO_STATE(R_USB_EPT_DATA_ISO, valid, yes) | + /* FIXME: Change any to the actual port? */ + IO_STATE(R_USB_EPT_DATA_ISO, port, any) | + IO_FIELD(R_USB_EPT_DATA_ISO, max_len, maxlen) | + IO_FIELD(R_USB_EPT_DATA_ISO, ep, endpoint) | + IO_FIELD(R_USB_EPT_DATA_ISO, dev, devnum); + etrax_epid_iso_set(epid, epid_data); + } else { + epid_data = IO_STATE(R_USB_EPT_DATA, valid, yes) | + IO_FIELD(R_USB_EPT_DATA, low_speed, slow) | + /* FIXME: Change any to the actual port? */ + IO_STATE(R_USB_EPT_DATA, port, any) | + IO_FIELD(R_USB_EPT_DATA, max_len, maxlen) | + IO_FIELD(R_USB_EPT_DATA, ep, endpoint) | + IO_FIELD(R_USB_EPT_DATA, dev, devnum); + etrax_epid_set(epid, epid_data); + } + + epid_state[epid].out_traffic = out_traffic; + epid_state[epid].type = usb_pipetype(urb->pipe); + + tc_warn("Setting up ep:0x%x epid:%d (addr:%d endp:%d max_len:%d %s %s %s)\n", + (unsigned int)ep, epid, devnum, endpoint, maxlen, + str_type(urb->pipe), out_traffic ? "out" : "in", + slow ? "low" : "full"); + + /* Enable Isoc eof interrupt if we set up the first Isoc epid */ + if(usb_pipeisoc(urb->pipe)) { + isoc_epid_counter++; + if(isoc_epid_counter == 1) { + isoc_warn("Enabled Isoc eof interrupt\n"); + *R_USB_IRQ_MASK_SET = IO_STATE(R_USB_IRQ_MASK_SET, iso_eof, set); + } + } + + DBFEXIT; + return epid; +} + +static void tc_free_epid(struct usb_host_endpoint *ep) { + unsigned long flags; + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + int epid; + volatile int timeout = 10000; + + DBFENTER; + + if (ep_priv == NULL) { + tc_warn("Trying to free unused epid on ep:0x%x\n", (unsigned int)ep); + DBFEXIT; + return; + } + + epid = ep_priv->epid; + + /* Disable Isoc eof interrupt if we free the last Isoc epid */ + if(epid_isoc(epid)) { + ASSERT(isoc_epid_counter > 0); + isoc_epid_counter--; + if(isoc_epid_counter == 0) { + *R_USB_IRQ_MASK_CLR = IO_STATE(R_USB_IRQ_MASK_CLR, iso_eof, clr); + isoc_warn("Disabled Isoc eof interrupt\n"); + } + } + + /* Take lock manualy instead of in epid_x_x wrappers, + because we need to be polling here */ + spin_lock_irqsave(&etrax_epid_lock, flags); + + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, epid); + nop(); + while((*R_USB_EPT_DATA & IO_MASK(R_USB_EPT_DATA, hold)) && + (timeout-- > 0)); + /* This will, among other things, set the valid field to 0. */ + *R_USB_EPT_DATA = 0; + spin_unlock_irqrestore(&etrax_epid_lock, flags); + + /* Free resource in software state info list */ + epid_state[epid].inuse = 0; + + /* Free private endpoint data */ + ep_priv_free(ep); + + DBFEXIT; +} + +static int tc_allocate_epid(void) { + int i; + DBFENTER; + for (i = 0; i < NBR_OF_EPIDS; i++) { + if (!epid_inuse(i)) { + DBFEXIT; + return i; + } + } + + tc_warn("Found no free epids\n"); + DBFEXIT; + return -1; +} + + +/* Wrappers around the list functions (include/linux/list.h). */ +/* ---------------------------------------------------------- */ +static inline int __urb_list_empty(int epid) { + int retval; + retval = list_empty(&urb_list[epid]); + return retval; +} + +/* Returns first urb for this epid, or NULL if list is empty. */ +static inline struct urb *urb_list_first(int epid) { + unsigned long flags; + struct urb *first_urb = 0; + spin_lock_irqsave(&urb_list_lock, flags); + if (!__urb_list_empty(epid)) { + /* Get the first urb (i.e. head->next). */ + urb_entry_t *urb_entry = list_entry((&urb_list[epid])->next, urb_entry_t, list); + first_urb = urb_entry->urb; + } + spin_unlock_irqrestore(&urb_list_lock, flags); + return first_urb; +} + +/* Adds an urb_entry last in the list for this epid. */ +static inline void urb_list_add(struct urb *urb, int epid, int mem_flags) { + unsigned long flags; + urb_entry_t *urb_entry = (urb_entry_t *)kmalloc(sizeof(urb_entry_t), mem_flags); + ASSERT(urb_entry); + + urb_entry->urb = urb; + spin_lock_irqsave(&urb_list_lock, flags); + list_add_tail(&urb_entry->list, &urb_list[epid]); + spin_unlock_irqrestore(&urb_list_lock, flags); +} + +/* Search through the list for an element that contains this urb. (The list + is expected to be short and the one we are about to delete will often be + the first in the list.) + Should be protected by spin_locks in calling function */ +static inline urb_entry_t *__urb_list_entry(struct urb *urb, int epid) { + struct list_head *entry; + struct list_head *tmp; + urb_entry_t *urb_entry; + + list_for_each_safe(entry, tmp, &urb_list[epid]) { + urb_entry = list_entry(entry, urb_entry_t, list); + ASSERT(urb_entry); + ASSERT(urb_entry->urb); + + if (urb_entry->urb == urb) { + return urb_entry; + } + } + return 0; +} + +/* Same function as above but for global use. Protects list by spinlock */ +static inline urb_entry_t *urb_list_entry(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + spin_unlock_irqrestore(&urb_list_lock, flags); + return (urb_entry); +} + +/* Delete an urb from the list. */ +static inline void urb_list_del(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + + /* Delete entry and free. */ + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + ASSERT(urb_entry); + + list_del(&urb_entry->list); + spin_unlock_irqrestore(&urb_list_lock, flags); + kfree(urb_entry); +} + +/* Move an urb to the end of the list. */ +static inline void urb_list_move_last(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + ASSERT(urb_entry); + + list_del(&urb_entry->list); + list_add_tail(&urb_entry->list, &urb_list[epid]); + spin_unlock_irqrestore(&urb_list_lock, flags); +} + +/* Get the next urb in the list. */ +static inline struct urb *urb_list_next(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + ASSERT(urb_entry); + + if (urb_entry->list.next != &urb_list[epid]) { + struct list_head *elem = urb_entry->list.next; + urb_entry = list_entry(elem, urb_entry_t, list); + spin_unlock_irqrestore(&urb_list_lock, flags); + return urb_entry->urb; + } else { + spin_unlock_irqrestore(&urb_list_lock, flags); + return NULL; + } +} + +struct USB_EP_Desc* create_ep(int epid, struct USB_SB_Desc* sb_desc, + int mem_flags) { + struct USB_EP_Desc *ep_desc; + ep_desc = (struct USB_EP_Desc *) kmem_cache_alloc(usb_desc_cache, mem_flags); + if(ep_desc == NULL) + return NULL; + memset(ep_desc, 0, sizeof(struct USB_EP_Desc)); + + ep_desc->hw_len = 0; + ep_desc->command = (IO_FIELD(USB_EP_command, epid, epid) | + IO_STATE(USB_EP_command, enable, yes)); + if(sb_desc == NULL) { + ep_desc->sub = 0; + } else { + ep_desc->sub = virt_to_phys(sb_desc); + } + return ep_desc; +} + +#define TT_ZOUT 0 +#define TT_IN 1 +#define TT_OUT 2 +#define TT_SETUP 3 + +#define CMD_EOL IO_STATE(USB_SB_command, eol, yes) +#define CMD_INTR IO_STATE(USB_SB_command, intr, yes) +#define CMD_FULL IO_STATE(USB_SB_command, full, yes) + +/* Allocation and setup of a generic SB. Used to create SETUP, OUT and ZOUT + SBs. Also used by create_sb_in() to avoid same allocation procedure at two + places */ +struct USB_SB_Desc* create_sb(struct USB_SB_Desc* sb_prev, int tt, void* data, + int datalen, int mem_flags) { + struct USB_SB_Desc *sb_desc; + sb_desc = (struct USB_SB_Desc*)kmem_cache_alloc(usb_desc_cache, mem_flags); + if(sb_desc == NULL) + return NULL; + memset(sb_desc, 0, sizeof(struct USB_SB_Desc)); + + sb_desc->command = IO_FIELD(USB_SB_command, tt, tt) | + IO_STATE(USB_SB_command, eot, yes); + + sb_desc->sw_len = datalen; + if(data != NULL) { + sb_desc->buf = virt_to_phys(data); + } else { + sb_desc->buf = 0; + } + if(sb_prev != NULL) { + sb_prev->next = virt_to_phys(sb_desc); + } + return sb_desc; +} + +/* Creates a copy of an existing SB by allocation space for it and copy + settings */ +struct USB_SB_Desc* create_sb_copy(struct USB_SB_Desc* sb_orig, int mem_flags) { + struct USB_SB_Desc *sb_desc; + sb_desc = (struct USB_SB_Desc*)kmem_cache_alloc(usb_desc_cache, mem_flags); + if(sb_desc == NULL) + return NULL; + + memcpy(sb_desc, sb_orig, sizeof(struct USB_SB_Desc)); + return sb_desc; +} + +/* A specific create_sb function for creation of in SBs. This is due to + that datalen in In SBs shows how many packets we are expecting. It also + sets up the rem field to show if how many bytes we expect in last packet + if it's not a full one */ +struct USB_SB_Desc* create_sb_in(struct USB_SB_Desc* sb_prev, int datalen, + int maxlen, int mem_flags) { + struct USB_SB_Desc *sb_desc; + sb_desc = create_sb(sb_prev, TT_IN, NULL, + datalen ? (datalen - 1) / maxlen + 1 : 0, mem_flags); + if(sb_desc == NULL) + return NULL; + sb_desc->command |= IO_FIELD(USB_SB_command, rem, datalen % maxlen); + return sb_desc; +} + +void set_sb_cmds(struct USB_SB_Desc *sb_desc, __u16 flags) { + sb_desc->command |= flags; +} + +int create_sb_for_urb(struct urb *urb, int mem_flags) { + int is_out = !usb_pipein(urb->pipe); + int type = usb_pipetype(urb->pipe); + int maxlen = usb_maxpacket(urb->dev, urb->pipe, is_out); + int buf_len = urb->transfer_buffer_length; + void *buf = buf_len > 0 ? urb->transfer_buffer : NULL; + struct USB_SB_Desc *sb_desc = NULL; + + struct crisv10_urb_priv *urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv != NULL); + + switch(type) { + case PIPE_CONTROL: + /* Setup stage */ + sb_desc = create_sb(NULL, TT_SETUP, urb->setup_packet, 8, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + set_sb_cmds(sb_desc, CMD_FULL); + + /* Attach first SB to URB */ + urb_priv->first_sb = sb_desc; + + if (is_out) { /* Out Control URB */ + /* If this Control OUT transfer has an optional data stage we add + an OUT token before the mandatory IN (status) token */ + if ((buf_len > 0) && buf) { + sb_desc = create_sb(sb_desc, TT_OUT, buf, buf_len, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + set_sb_cmds(sb_desc, CMD_FULL); + } + + /* Status stage */ + /* The data length has to be exactly 1. This is due to a requirement + of the USB specification that a host must be prepared to receive + data in the status phase */ + sb_desc = create_sb(sb_desc, TT_IN, NULL, 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + } else { /* In control URB */ + /* Data stage */ + sb_desc = create_sb_in(sb_desc, buf_len, maxlen, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + + /* Status stage */ + /* Read comment at zout_buffer declaration for an explanation to this. */ + sb_desc = create_sb(sb_desc, TT_ZOUT, &zout_buffer[0], 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + /* Set descriptor interrupt flag for in URBs so we can finish URB after + zout-packet has been sent */ + set_sb_cmds(sb_desc, CMD_INTR | CMD_FULL); + } + /* Set end-of-list flag in last SB */ + set_sb_cmds(sb_desc, CMD_EOL); + /* Attach last SB to URB */ + urb_priv->last_sb = sb_desc; + break; + + case PIPE_BULK: + if (is_out) { /* Out Bulk URB */ + sb_desc = create_sb(NULL, TT_OUT, buf, buf_len, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + /* The full field is set to yes, even if we don't actually check that + this is a full-length transfer (i.e., that transfer_buffer_length % + maxlen = 0). + Setting full prevents the USB controller from sending an empty packet + in that case. However, if URB_ZERO_PACKET was set we want that. */ + if (!(urb->transfer_flags & URB_ZERO_PACKET)) { + set_sb_cmds(sb_desc, CMD_FULL); + } + } else { /* In Bulk URB */ + sb_desc = create_sb_in(NULL, buf_len, maxlen, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + } + /* Set end-of-list flag for last SB */ + set_sb_cmds(sb_desc, CMD_EOL); + + /* Attach SB to URB */ + urb_priv->first_sb = sb_desc; + urb_priv->last_sb = sb_desc; + break; + + case PIPE_INTERRUPT: + if(is_out) { /* Out Intr URB */ + sb_desc = create_sb(NULL, TT_OUT, buf, buf_len, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + + /* The full field is set to yes, even if we don't actually check that + this is a full-length transfer (i.e., that transfer_buffer_length % + maxlen = 0). + Setting full prevents the USB controller from sending an empty packet + in that case. However, if URB_ZERO_PACKET was set we want that. */ + if (!(urb->transfer_flags & URB_ZERO_PACKET)) { + set_sb_cmds(sb_desc, CMD_FULL); + } + /* Only generate TX interrupt if it's a Out URB*/ + set_sb_cmds(sb_desc, CMD_INTR); + + } else { /* In Intr URB */ + sb_desc = create_sb_in(NULL, buf_len, maxlen, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + } + /* Set end-of-list flag for last SB */ + set_sb_cmds(sb_desc, CMD_EOL); + + /* Attach SB to URB */ + urb_priv->first_sb = sb_desc; + urb_priv->last_sb = sb_desc; + + break; + case PIPE_ISOCHRONOUS: + if(is_out) { /* Out Isoc URB */ + int i; + if(urb->number_of_packets == 0) { + tc_err("Can't create SBs for Isoc URB with zero packets\n"); + return -EPIPE; + } + /* Create one SB descriptor for each packet and link them together. */ + for(i = 0; i < urb->number_of_packets; i++) { + if (urb->iso_frame_desc[i].length > 0) { + + sb_desc = create_sb(sb_desc, TT_OUT, urb->transfer_buffer + + urb->iso_frame_desc[i].offset, + urb->iso_frame_desc[i].length, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + + /* Check if it's a full length packet */ + if (urb->iso_frame_desc[i].length == + usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe))) { + set_sb_cmds(sb_desc, CMD_FULL); + } + + } else { /* zero length packet */ + sb_desc = create_sb(sb_desc, TT_ZOUT, &zout_buffer[0], 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + set_sb_cmds(sb_desc, CMD_FULL); + } + /* Attach first SB descriptor to URB */ + if (i == 0) { + urb_priv->first_sb = sb_desc; + } + } + /* Set interrupt and end-of-list flags in last SB */ + set_sb_cmds(sb_desc, CMD_INTR | CMD_EOL); + /* Attach last SB descriptor to URB */ + urb_priv->last_sb = sb_desc; + tc_dbg("Created %d out SBs for Isoc URB:0x%x\n", + urb->number_of_packets, (unsigned int)urb); + } else { /* In Isoc URB */ + /* Actual number of packets is not relevant for periodic in traffic as + long as it is more than zero. Set to 1 always. */ + sb_desc = create_sb(sb_desc, TT_IN, NULL, 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + /* Set end-of-list flags for SB */ + set_sb_cmds(sb_desc, CMD_EOL); + + /* Attach SB to URB */ + urb_priv->first_sb = sb_desc; + urb_priv->last_sb = sb_desc; + } + break; + default: + tc_err("Unknown pipe-type\n"); + return -EPIPE; + break; + } + return 0; +} + +int init_intr_urb(struct urb *urb, int mem_flags) { + struct crisv10_urb_priv *urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + struct USB_EP_Desc* ep_desc; + int interval; + int i; + int ep_count; + + ASSERT(urb_priv != NULL); + ASSERT(usb_pipeint(urb->pipe)); + /* We can't support interval longer than amount of eof descriptors in + TxIntrEPList */ + if(urb->interval > MAX_INTR_INTERVAL) { + tc_err("Interrupt interval %dms too big (max: %dms)\n", urb->interval, + MAX_INTR_INTERVAL); + return -EINVAL; + } + + /* We assume that the SB descriptors already have been setup */ + ASSERT(urb_priv->first_sb != NULL); + + /* Round of the interval to 2^n, it is obvious that this code favours + smaller numbers, but that is actually a good thing */ + /* FIXME: The "rounding error" for larger intervals will be quite + large. For in traffic this shouldn't be a problem since it will only + mean that we "poll" more often. */ + interval = urb->interval; + for (i = 0; interval; i++) { + interval = interval >> 1; + } + urb_priv->interval = 1 << (i - 1); + + /* We can only have max interval for Out Interrupt due to that we can only + handle one linked in EP for a certain epid in the Intr descr array at the + time. The USB Controller in the Etrax 100LX continues to process Intr EPs + so we have no way of knowing which one that caused the actual transfer if + we have several linked in. */ + if(usb_pipeout(urb->pipe)) { + urb_priv->interval = MAX_INTR_INTERVAL; + } + + /* Calculate amount of EPs needed */ + ep_count = MAX_INTR_INTERVAL / urb_priv->interval; + + for(i = 0; i < ep_count; i++) { + ep_desc = create_ep(urb_priv->epid, urb_priv->first_sb, mem_flags); + if(ep_desc == NULL) { + /* Free any descriptors that we may have allocated before failure */ + while(i > 0) { + i--; + kfree(urb_priv->intr_ep_pool[i]); + } + return -ENOMEM; + } + urb_priv->intr_ep_pool[i] = ep_desc; + } + urb_priv->intr_ep_pool_length = ep_count; + return 0; +} + +/* DMA RX/TX functions */ +/* ----------------------- */ + +static void tc_dma_init_rx_list(void) { + int i; + + /* Setup descriptor list except last one */ + for (i = 0; i < (NBR_OF_RX_DESC - 1); i++) { + RxDescList[i].sw_len = RX_DESC_BUF_SIZE; + RxDescList[i].command = 0; + RxDescList[i].next = virt_to_phys(&RxDescList[i + 1]); + RxDescList[i].buf = virt_to_phys(RxBuf + (i * RX_DESC_BUF_SIZE)); + RxDescList[i].hw_len = 0; + RxDescList[i].status = 0; + + /* DMA IN cache bug. (struct etrax_dma_descr has the same layout as + USB_IN_Desc for the relevant fields.) */ + prepare_rx_descriptor((struct etrax_dma_descr*)&RxDescList[i]); + + } + /* Special handling of last descriptor */ + RxDescList[i].sw_len = RX_DESC_BUF_SIZE; + RxDescList[i].command = IO_STATE(USB_IN_command, eol, yes); + RxDescList[i].next = virt_to_phys(&RxDescList[0]); + RxDescList[i].buf = virt_to_phys(RxBuf + (i * RX_DESC_BUF_SIZE)); + RxDescList[i].hw_len = 0; + RxDescList[i].status = 0; + + /* Setup list pointers that show progress in list */ + myNextRxDesc = &RxDescList[0]; + myLastRxDesc = &RxDescList[NBR_OF_RX_DESC - 1]; + + flush_etrax_cache(); + /* Point DMA to first descriptor in list and start it */ + *R_DMA_CH9_FIRST = virt_to_phys(myNextRxDesc); + *R_DMA_CH9_CMD = IO_STATE(R_DMA_CH9_CMD, cmd, start); +} + + +static void tc_dma_init_tx_bulk_list(void) { + int i; + volatile struct USB_EP_Desc *epDescr; + + for (i = 0; i < (NBR_OF_EPIDS - 1); i++) { + epDescr = &(TxBulkEPList[i]); + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxBulkEPList[i + 1]); + + /* Initiate two EPs, disabled and with the eol flag set. No need for any + preserved epid. */ + + /* The first one has the intr flag set so we get an interrupt when the DMA + channel is about to become disabled. */ + CHECK_ALIGN(&TxBulkDummyEPList[i][0]); + TxBulkDummyEPList[i][0].hw_len = 0; + TxBulkDummyEPList[i][0].command = (IO_FIELD(USB_EP_command, epid, DUMMY_EPID) | + IO_STATE(USB_EP_command, eol, yes) | + IO_STATE(USB_EP_command, intr, yes)); + TxBulkDummyEPList[i][0].sub = 0; + TxBulkDummyEPList[i][0].next = virt_to_phys(&TxBulkDummyEPList[i][1]); + + /* The second one. */ + CHECK_ALIGN(&TxBulkDummyEPList[i][1]); + TxBulkDummyEPList[i][1].hw_len = 0; + TxBulkDummyEPList[i][1].command = (IO_FIELD(USB_EP_command, epid, DUMMY_EPID) | + IO_STATE(USB_EP_command, eol, yes)); + TxBulkDummyEPList[i][1].sub = 0; + /* The last dummy's next pointer is the same as the current EP's next pointer. */ + TxBulkDummyEPList[i][1].next = virt_to_phys(&TxBulkEPList[i + 1]); + } + + /* Special handling of last descr in list, make list circular */ + epDescr = &TxBulkEPList[i]; + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_STATE(USB_EP_command, eol, yes) | + IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxBulkEPList[0]); + + /* Init DMA sub-channel pointers to last item in each list */ + *R_DMA_CH8_SUB0_EP = virt_to_phys(&TxBulkEPList[i]); + /* No point in starting the bulk channel yet. + *R_DMA_CH8_SUB0_CMD = IO_STATE(R_DMA_CH8_SUB0_CMD, cmd, start); */ +} + +static void tc_dma_init_tx_ctrl_list(void) { + int i; + volatile struct USB_EP_Desc *epDescr; + + for (i = 0; i < (NBR_OF_EPIDS - 1); i++) { + epDescr = &(TxCtrlEPList[i]); + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxCtrlEPList[i + 1]); + } + /* Special handling of last descr in list, make list circular */ + epDescr = &TxCtrlEPList[i]; + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_STATE(USB_EP_command, eol, yes) | + IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxCtrlEPList[0]); + + /* Init DMA sub-channel pointers to last item in each list */ + *R_DMA_CH8_SUB1_EP = virt_to_phys(&TxCtrlEPList[i]); + /* No point in starting the ctrl channel yet. + *R_DMA_CH8_SUB1_CMD = IO_STATE(R_DMA_CH8_SUB0_CMD, cmd, start); */ +} + + +static void tc_dma_init_tx_intr_list(void) { + int i; + + TxIntrSB_zout.sw_len = 1; + TxIntrSB_zout.next = 0; + TxIntrSB_zout.buf = virt_to_phys(&zout_buffer[0]); + TxIntrSB_zout.command = (IO_FIELD(USB_SB_command, rem, 0) | + IO_STATE(USB_SB_command, tt, zout) | + IO_STATE(USB_SB_command, full, yes) | + IO_STATE(USB_SB_command, eot, yes) | + IO_STATE(USB_SB_command, eol, yes)); + + for (i = 0; i < (MAX_INTR_INTERVAL - 1); i++) { + CHECK_ALIGN(&TxIntrEPList[i]); + TxIntrEPList[i].hw_len = 0; + TxIntrEPList[i].command = + (IO_STATE(USB_EP_command, eof, yes) | + IO_STATE(USB_EP_command, enable, yes) | + IO_FIELD(USB_EP_command, epid, INVALID_EPID)); + TxIntrEPList[i].sub = virt_to_phys(&TxIntrSB_zout); + TxIntrEPList[i].next = virt_to_phys(&TxIntrEPList[i + 1]); + } + + /* Special handling of last descr in list, make list circular */ + CHECK_ALIGN(&TxIntrEPList[i]); + TxIntrEPList[i].hw_len = 0; + TxIntrEPList[i].command = + (IO_STATE(USB_EP_command, eof, yes) | + IO_STATE(USB_EP_command, eol, yes) | + IO_STATE(USB_EP_command, enable, yes) | + IO_FIELD(USB_EP_command, epid, INVALID_EPID)); + TxIntrEPList[i].sub = virt_to_phys(&TxIntrSB_zout); + TxIntrEPList[i].next = virt_to_phys(&TxIntrEPList[0]); + + intr_dbg("Initiated Intr EP descriptor list\n"); + + + /* Connect DMA 8 sub-channel 2 to first in list */ + *R_DMA_CH8_SUB2_EP = virt_to_phys(&TxIntrEPList[0]); +} + +static void tc_dma_init_tx_isoc_list(void) { + int i; + + DBFENTER; + + /* Read comment at zout_buffer declaration for an explanation to this. */ + TxIsocSB_zout.sw_len = 1; + TxIsocSB_zout.next = 0; + TxIsocSB_zout.buf = virt_to_phys(&zout_buffer[0]); + TxIsocSB_zout.command = (IO_FIELD(USB_SB_command, rem, 0) | + IO_STATE(USB_SB_command, tt, zout) | + IO_STATE(USB_SB_command, full, yes) | + IO_STATE(USB_SB_command, eot, yes) | + IO_STATE(USB_SB_command, eol, yes)); + + /* The last isochronous EP descriptor is a dummy. */ + for (i = 0; i < (NBR_OF_EPIDS - 1); i++) { + CHECK_ALIGN(&TxIsocEPList[i]); + TxIsocEPList[i].hw_len = 0; + TxIsocEPList[i].command = IO_FIELD(USB_EP_command, epid, i); + TxIsocEPList[i].sub = 0; + TxIsocEPList[i].next = virt_to_phys(&TxIsocEPList[i + 1]); + } + + CHECK_ALIGN(&TxIsocEPList[i]); + TxIsocEPList[i].hw_len = 0; + + /* Must enable the last EP descr to get eof interrupt. */ + TxIsocEPList[i].command = (IO_STATE(USB_EP_command, enable, yes) | + IO_STATE(USB_EP_command, eof, yes) | + IO_STATE(USB_EP_command, eol, yes) | + IO_FIELD(USB_EP_command, epid, INVALID_EPID)); + TxIsocEPList[i].sub = virt_to_phys(&TxIsocSB_zout); + TxIsocEPList[i].next = virt_to_phys(&TxIsocEPList[0]); + + *R_DMA_CH8_SUB3_EP = virt_to_phys(&TxIsocEPList[0]); + *R_DMA_CH8_SUB3_CMD = IO_STATE(R_DMA_CH8_SUB3_CMD, cmd, start); +} + +static int tc_dma_init(struct usb_hcd *hcd) { + tc_dma_init_rx_list(); + tc_dma_init_tx_bulk_list(); + tc_dma_init_tx_ctrl_list(); + tc_dma_init_tx_intr_list(); + tc_dma_init_tx_isoc_list(); + + if (cris_request_dma(USB_TX_DMA_NBR, + "ETRAX 100LX built-in USB (Tx)", + DMA_VERBOSE_ON_ERROR, + dma_usb)) { + err("Could not allocate DMA ch 8 for USB"); + return -EBUSY; + } + + if (cris_request_dma(USB_RX_DMA_NBR, + "ETRAX 100LX built-in USB (Rx)", + DMA_VERBOSE_ON_ERROR, + dma_usb)) { + err("Could not allocate DMA ch 9 for USB"); + return -EBUSY; + } + + *R_IRQ_MASK2_SET = + /* Note that these interrupts are not used. */ + IO_STATE(R_IRQ_MASK2_SET, dma8_sub0_descr, set) | + /* Sub channel 1 (ctrl) descr. interrupts are used. */ + IO_STATE(R_IRQ_MASK2_SET, dma8_sub1_descr, set) | + IO_STATE(R_IRQ_MASK2_SET, dma8_sub2_descr, set) | + /* Sub channel 3 (isoc) descr. interrupts are used. */ + IO_STATE(R_IRQ_MASK2_SET, dma8_sub3_descr, set); + + /* Note that the dma9_descr interrupt is not used. */ + *R_IRQ_MASK2_SET = + IO_STATE(R_IRQ_MASK2_SET, dma9_eop, set) | + IO_STATE(R_IRQ_MASK2_SET, dma9_descr, set); + + if (request_irq(ETRAX_USB_RX_IRQ, tc_dma_rx_interrupt, 0, + "ETRAX 100LX built-in USB (Rx)", hcd)) { + err("Could not allocate IRQ %d for USB", ETRAX_USB_RX_IRQ); + return -EBUSY; + } + + if (request_irq(ETRAX_USB_TX_IRQ, tc_dma_tx_interrupt, 0, + "ETRAX 100LX built-in USB (Tx)", hcd)) { + err("Could not allocate IRQ %d for USB", ETRAX_USB_TX_IRQ); + return -EBUSY; + } + + return 0; +} + +static void tc_dma_destroy(void) { + free_irq(ETRAX_USB_RX_IRQ, NULL); + free_irq(ETRAX_USB_TX_IRQ, NULL); + + cris_free_dma(USB_TX_DMA_NBR, "ETRAX 100LX built-in USB (Tx)"); + cris_free_dma(USB_RX_DMA_NBR, "ETRAX 100LX built-in USB (Rx)"); + +} + +static void tc_dma_link_intr_urb(struct urb *urb); + +/* Handle processing of Bulk, Ctrl and Intr queues */ +static void tc_dma_process_queue(int epid) { + struct urb *urb; + struct crisv10_urb_priv *urb_priv; + unsigned long flags; + char toggle; + + if(epid_state[epid].disabled) { + /* Don't process any URBs on a disabled endpoint */ + return; + } + + /* Do not disturb us while fiddling with EPs and epids */ + local_irq_save(flags); + + /* For bulk, Ctrl and Intr can we only have one URB active at a time for + a specific EP. */ + if(activeUrbList[epid] != NULL) { + /* An URB is already active on EP, skip checking queue */ + local_irq_restore(flags); + return; + } + + urb = urb_list_first(epid); + if(urb == NULL) { + /* No URB waiting in EP queue. Nothing do to */ + local_irq_restore(flags); + return; + } + + urb_priv = urb->hcpriv; + ASSERT(urb_priv != NULL); + ASSERT(urb_priv->urb_state == NOT_STARTED); + ASSERT(!usb_pipeisoc(urb->pipe)); + + /* Remove this URB from the queue and move it to active */ + activeUrbList[epid] = urb; + urb_list_del(urb, epid); + + urb_priv->urb_state = STARTED; + + /* Reset error counters (regardless of which direction this traffic is). */ + etrax_epid_clear_error(epid); + + /* Special handling of Intr EP lists */ + if(usb_pipeint(urb->pipe)) { + tc_dma_link_intr_urb(urb); + local_irq_restore(flags); + return; + } + + /* Software must preset the toggle bits for Bulk and Ctrl */ + if(usb_pipecontrol(urb->pipe)) { + /* Toggle bits are initialized only during setup transaction in a + CTRL transfer */ + etrax_epid_set_toggle(epid, 0, 0); + etrax_epid_set_toggle(epid, 1, 0); + } else { + toggle = usb_gettoggle(urb->dev, usb_pipeendpoint(urb->pipe), + usb_pipeout(urb->pipe)); + etrax_epid_set_toggle(epid, usb_pipeout(urb->pipe), toggle); + } + + tc_dbg("Added SBs from (URB:0x%x %s %s) to epid %d: %s\n", + (unsigned int)urb, str_dir(urb->pipe), str_type(urb->pipe), epid, + sblist_to_str(urb_priv->first_sb)); + + /* We start the DMA sub channel without checking if it's running or not, + because: + 1) If it's already running, issuing the start command is a nop. + 2) We avoid a test-and-set race condition. */ + switch(usb_pipetype(urb->pipe)) { + case PIPE_BULK: + /* Assert that the EP descriptor is disabled. */ + ASSERT(!(TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable))); + + /* Set up and enable the EP descriptor. */ + TxBulkEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + TxBulkEPList[epid].hw_len = 0; + TxBulkEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + + /* Check if the dummy list is already with us (if several urbs were queued). */ + if (usb_pipein(urb->pipe) && (TxBulkEPList[epid].next != virt_to_phys(&TxBulkDummyEPList[epid][0]))) { + tc_dbg("Inviting dummy list to the party for urb 0x%lx, epid %d", + (unsigned long)urb, epid); + + /* We don't need to check if the DMA is at this EP or not before changing the + next pointer, since we will do it in one 32-bit write (EP descriptors are + 32-bit aligned). */ + TxBulkEPList[epid].next = virt_to_phys(&TxBulkDummyEPList[epid][0]); + } + + restart_dma8_sub0(); + + /* Update/restart the bulk start timer since we just started the channel.*/ + mod_timer(&bulk_start_timer, jiffies + BULK_START_TIMER_INTERVAL); + /* Update/restart the bulk eot timer since we just inserted traffic. */ + mod_timer(&bulk_eot_timer, jiffies + BULK_EOT_TIMER_INTERVAL); + break; + case PIPE_CONTROL: + /* Assert that the EP descriptor is disabled. */ + ASSERT(!(TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable))); + + /* Set up and enable the EP descriptor. */ + TxCtrlEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + TxCtrlEPList[epid].hw_len = 0; + TxCtrlEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + + *R_DMA_CH8_SUB1_CMD = IO_STATE(R_DMA_CH8_SUB1_CMD, cmd, start); + break; + } + local_irq_restore(flags); +} + +static void tc_dma_link_intr_urb(struct urb *urb) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + volatile struct USB_EP_Desc *tmp_ep; + struct USB_EP_Desc *ep_desc; + int i = 0, epid; + int pool_idx = 0; + + ASSERT(urb_priv != NULL); + epid = urb_priv->epid; + ASSERT(urb_priv->interval > 0); + ASSERT(urb_priv->intr_ep_pool_length > 0); + + tmp_ep = &TxIntrEPList[0]; + + /* Only insert one EP descriptor in list for Out Intr URBs. + We can only handle Out Intr with interval of 128ms because + it's not possible to insert several Out Intr EPs because they + are not consumed by the DMA. */ + if(usb_pipeout(urb->pipe)) { + ep_desc = urb_priv->intr_ep_pool[0]; + ASSERT(ep_desc); + ep_desc->next = tmp_ep->next; + tmp_ep->next = virt_to_phys(ep_desc); + i++; + } else { + /* Loop through Intr EP descriptor list and insert EP for URB at + specified interval */ + do { + /* Each EP descriptor with eof flag sat signals a new frame */ + if (tmp_ep->command & IO_MASK(USB_EP_command, eof)) { + /* Insert a EP from URBs EP pool at correct interval */ + if ((i % urb_priv->interval) == 0) { + ep_desc = urb_priv->intr_ep_pool[pool_idx]; + ASSERT(ep_desc); + ep_desc->next = tmp_ep->next; + tmp_ep->next = virt_to_phys(ep_desc); + pool_idx++; + ASSERT(pool_idx <= urb_priv->intr_ep_pool_length); + } + i++; + } + tmp_ep = (struct USB_EP_Desc *)phys_to_virt(tmp_ep->next); + } while(tmp_ep != &TxIntrEPList[0]); + } + + intr_dbg("Added SBs to intr epid %d: %s interval:%d (%d EP)\n", epid, + sblist_to_str(urb_priv->first_sb), urb_priv->interval, pool_idx); + + /* We start the DMA sub channel without checking if it's running or not, + because: + 1) If it's already running, issuing the start command is a nop. + 2) We avoid a test-and-set race condition. */ + *R_DMA_CH8_SUB2_CMD = IO_STATE(R_DMA_CH8_SUB2_CMD, cmd, start); +} + +static void tc_dma_process_isoc_urb(struct urb *urb) { + unsigned long flags; + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + int epid; + + /* Do not disturb us while fiddling with EPs and epids */ + local_irq_save(flags); + + ASSERT(urb_priv); + ASSERT(urb_priv->first_sb); + epid = urb_priv->epid; + + if(activeUrbList[epid] == NULL) { + /* EP is idle, so make this URB active */ + activeUrbList[epid] = urb; + urb_list_del(urb, epid); + ASSERT(TxIsocEPList[epid].sub == 0); + ASSERT(!(TxIsocEPList[epid].command & + IO_STATE(USB_EP_command, enable, yes))); + + /* Differentiate between In and Out Isoc. Because In SBs are not consumed*/ + if(usb_pipein(urb->pipe)) { + /* Each EP for In Isoc will have only one SB descriptor, setup when + submitting the first active urb. We do it here by copying from URBs + pre-allocated SB. */ + memcpy((void *)&(TxIsocSBList[epid]), urb_priv->first_sb, + sizeof(TxIsocSBList[epid])); + TxIsocEPList[epid].hw_len = 0; + TxIsocEPList[epid].sub = virt_to_phys(&(TxIsocSBList[epid])); + } else { + /* For Out Isoc we attach the pre-allocated list of SBs for the URB */ + TxIsocEPList[epid].hw_len = 0; + TxIsocEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + + isoc_dbg("Attached first URB:0x%x[%d] to epid:%d first_sb:0x%x" + " last_sb::0x%x\n", + (unsigned int)urb, urb_priv->urb_num, epid, + (unsigned int)(urb_priv->first_sb), + (unsigned int)(urb_priv->last_sb)); + } + + if (urb->transfer_flags & URB_ISO_ASAP) { + /* The isoc transfer should be started as soon as possible. The + start_frame field is a return value if URB_ISO_ASAP was set. Comparing + R_USB_FM_NUMBER with a USB Chief trace shows that the first isoc IN + token is sent 2 frames later. I'm not sure how this affects usage of + the start_frame field by the device driver, or how it affects things + when USB_ISO_ASAP is not set, so therefore there's no compensation for + the 2 frame "lag" here. */ + urb->start_frame = (*R_USB_FM_NUMBER & 0x7ff); + TxIsocEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + urb_priv->urb_state = STARTED; + isoc_dbg("URB_ISO_ASAP set, urb->start_frame set to %d\n", + urb->start_frame); + } else { + /* Not started yet. */ + urb_priv->urb_state = NOT_STARTED; + isoc_warn("urb_priv->urb_state set to NOT_STARTED for URB:0x%x\n", + (unsigned int)urb); + } + + } else { + /* An URB is already active on the EP. Leave URB in queue and let + finish_isoc_urb process it after current active URB */ + ASSERT(TxIsocEPList[epid].sub != 0); + + if(usb_pipein(urb->pipe)) { + /* Because there already is a active In URB on this epid we do nothing + and the finish_isoc_urb() function will handle switching to next URB*/ + + } else { /* For Out Isoc, insert new URBs traffic last in SB-list. */ + struct USB_SB_Desc *temp_sb_desc; + + /* Set state STARTED to all Out Isoc URBs added to SB list because we + don't know how many of them that are finished before descr interrupt*/ + urb_priv->urb_state = STARTED; + + /* Find end of current SB list by looking for SB with eol flag sat */ + temp_sb_desc = phys_to_virt(TxIsocEPList[epid].sub); + while ((temp_sb_desc->command & IO_MASK(USB_SB_command, eol)) != + IO_STATE(USB_SB_command, eol, yes)) { + ASSERT(temp_sb_desc->next); + temp_sb_desc = phys_to_virt(temp_sb_desc->next); + } + + isoc_dbg("Appended URB:0x%x[%d] (first:0x%x last:0x%x) to epid:%d" + " sub:0x%x eol:0x%x\n", + (unsigned int)urb, urb_priv->urb_num, + (unsigned int)(urb_priv->first_sb), + (unsigned int)(urb_priv->last_sb), epid, + (unsigned int)phys_to_virt(TxIsocEPList[epid].sub), + (unsigned int)temp_sb_desc); + + /* Next pointer must be set before eol is removed. */ + temp_sb_desc->next = virt_to_phys(urb_priv->first_sb); + /* Clear the previous end of list flag since there is a new in the + added SB descriptor list. */ + temp_sb_desc->command &= ~IO_MASK(USB_SB_command, eol); + + if (!(TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable))) { + __u32 epid_data; + /* 8.8.5 in Designer's Reference says we should check for and correct + any errors in the EP here. That should not be necessary if + epid_attn is handled correctly, so we assume all is ok. */ + epid_data = etrax_epid_iso_get(epid); + if (IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data) != + IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + isoc_err("Disabled Isoc EP with error:%d on epid:%d when appending" + " URB:0x%x[%d]\n", + IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data), epid, + (unsigned int)urb, urb_priv->urb_num); + } + + /* The SB list was exhausted. */ + if (virt_to_phys(urb_priv->last_sb) != TxIsocEPList[epid].sub) { + /* The new sublist did not get processed before the EP was + disabled. Setup the EP again. */ + + if(virt_to_phys(temp_sb_desc) == TxIsocEPList[epid].sub) { + isoc_dbg("EP for epid:%d stoped at SB:0x%x before newly inserted" + ", restarting from this URBs SB:0x%x\n", + epid, (unsigned int)temp_sb_desc, + (unsigned int)(urb_priv->first_sb)); + TxIsocEPList[epid].hw_len = 0; + TxIsocEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + urb->start_frame = (*R_USB_FM_NUMBER & 0x7ff); + /* Enable the EP again so data gets processed this time */ + TxIsocEPList[epid].command |= + IO_STATE(USB_EP_command, enable, yes); + + } else { + /* The EP has been disabled but not at end this URB (god knows + where). This should generate an epid_attn so we should not be + here */ + isoc_warn("EP was disabled on sb:0x%x before SB list for" + " URB:0x%x[%d] got processed\n", + (unsigned int)phys_to_virt(TxIsocEPList[epid].sub), + (unsigned int)urb, urb_priv->urb_num); + } + } else { + /* This might happend if we are slow on this function and isn't + an error. */ + isoc_dbg("EP was disabled and finished with SBs from appended" + " URB:0x%x[%d]\n", (unsigned int)urb, urb_priv->urb_num); + } + } + } + } + + /* Start the DMA sub channel */ + *R_DMA_CH8_SUB3_CMD = IO_STATE(R_DMA_CH8_SUB3_CMD, cmd, start); + + local_irq_restore(flags); +} + +static void tc_dma_unlink_intr_urb(struct urb *urb) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + volatile struct USB_EP_Desc *first_ep; /* First EP in the list. */ + volatile struct USB_EP_Desc *curr_ep; /* Current EP, the iterator. */ + volatile struct USB_EP_Desc *next_ep; /* The EP after current. */ + volatile struct USB_EP_Desc *unlink_ep; /* The one we should remove from + the list. */ + int count = 0; + volatile int timeout = 10000; + int epid; + + /* Read 8.8.4 in Designer's Reference, "Removing an EP Descriptor from the + List". */ + ASSERT(urb_priv); + ASSERT(urb_priv->intr_ep_pool_length > 0); + epid = urb_priv->epid; + + /* First disable all Intr EPs belonging to epid for this URB */ + first_ep = &TxIntrEPList[0]; + curr_ep = first_ep; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if (IO_EXTRACT(USB_EP_command, epid, next_ep->command) == epid) { + /* Disable EP */ + next_ep->command &= ~IO_MASK(USB_EP_command, enable); + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != first_ep); + + + /* Now unlink all EPs belonging to this epid from Descr list */ + first_ep = &TxIntrEPList[0]; + curr_ep = first_ep; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if (IO_EXTRACT(USB_EP_command, epid, next_ep->command) == epid) { + /* This is the one we should unlink. */ + unlink_ep = next_ep; + + /* Actually unlink the EP from the DMA list. */ + curr_ep->next = unlink_ep->next; + + /* Wait until the DMA is no longer at this descriptor. */ + while((*R_DMA_CH8_SUB2_EP == virt_to_phys(unlink_ep)) && + (timeout-- > 0)); + + count++; + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != first_ep); + + if(count != urb_priv->intr_ep_pool_length) { + intr_warn("Unlinked %d of %d Intr EPs for URB:0x%x[%d]\n", count, + urb_priv->intr_ep_pool_length, (unsigned int)urb, + urb_priv->urb_num); + } else { + intr_dbg("Unlinked %d of %d interrupt EPs for URB:0x%x\n", count, + urb_priv->intr_ep_pool_length, (unsigned int)urb); + } +} + +static void check_finished_bulk_tx_epids(struct usb_hcd *hcd, + int timer) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + __u32 epid_data; + + /* Protect TxEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + /* A finished EP descriptor is disabled and has a valid sub pointer */ + if (!(TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) && + (TxBulkEPList[epid].sub != 0)) { + + /* Get the active URB for this epid */ + urb = activeUrbList[epid]; + /* Sanity checks */ + ASSERT(urb); + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + /* Only handle finished out Bulk EPs here, + and let RX interrupt take care of the rest */ + if(!epid_out_traffic(epid)) { + continue; + } + + if(timer) { + tc_warn("Found finished %s Bulk epid:%d URB:0x%x[%d] from timeout\n", + epid_out_traffic(epid) ? "Out" : "In", epid, (unsigned int)urb, + urb_priv->urb_num); + } else { + tc_dbg("Found finished %s Bulk epid:%d URB:0x%x[%d] from interrupt\n", + epid_out_traffic(epid) ? "Out" : "In", epid, (unsigned int)urb, + urb_priv->urb_num); + } + + if(urb_priv->urb_state == UNLINK) { + /* This Bulk URB is requested to be unlinked, that means that the EP + has been disabled and we might not have sent all data */ + tc_finish_urb(hcd, urb, urb->status); + continue; + } + + ASSERT(urb_priv->urb_state == STARTED); + if (phys_to_virt(TxBulkEPList[epid].sub) != urb_priv->last_sb) { + tc_err("Endpoint got disabled before reaching last sb\n"); + } + + epid_data = etrax_epid_get(epid); + if (IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data) == + IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + /* This means that the endpoint has no error, is disabled + and had inserted traffic, i.e. transfer successfully completed. */ + tc_finish_urb(hcd, urb, 0); + } else { + /* Shouldn't happen. We expect errors to be caught by epid + attention. */ + tc_err("Found disabled bulk EP desc (epid:%d error:%d)\n", + epid, IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data)); + } + } else { + tc_dbg("Ignoring In Bulk epid:%d, let RX interrupt handle it\n", epid); + } + } + + local_irq_restore(flags); +} + +static void check_finished_ctrl_tx_epids(struct usb_hcd *hcd) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + __u32 epid_data; + + /* Protect TxEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if(epid == DUMMY_EPID) + continue; + + /* A finished EP descriptor is disabled and has a valid sub pointer */ + if (!(TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) && + (TxCtrlEPList[epid].sub != 0)) { + + /* Get the active URB for this epid */ + urb = activeUrbList[epid]; + + if(urb == NULL) { + tc_warn("Found finished Ctrl epid:%d with no active URB\n", epid); + continue; + } + + /* Sanity checks */ + ASSERT(usb_pipein(urb->pipe)); + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + if (phys_to_virt(TxCtrlEPList[epid].sub) != urb_priv->last_sb) { + tc_err("Endpoint got disabled before reaching last sb\n"); + } + + epid_data = etrax_epid_get(epid); + if (IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data) == + IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + /* This means that the endpoint has no error, is disabled + and had inserted traffic, i.e. transfer successfully completed. */ + + /* Check if RX-interrupt for In Ctrl has been processed before + finishing the URB */ + if(urb_priv->ctrl_rx_done) { + tc_dbg("Finishing In Ctrl URB:0x%x[%d] in tx_interrupt\n", + (unsigned int)urb, urb_priv->urb_num); + tc_finish_urb(hcd, urb, 0); + } else { + /* If we get zout descriptor interrupt before RX was done for a + In Ctrl transfer, then we flag that and it will be finished + in the RX-Interrupt */ + urb_priv->ctrl_zout_done = 1; + tc_dbg("Got zout descr interrupt before RX interrupt\n"); + } + } else { + /* Shouldn't happen. We expect errors to be caught by epid + attention. */ + tc_err("Found disabled Ctrl EP desc (epid:%d URB:0x%x[%d]) error_code:%d\n", epid, (unsigned int)urb, urb_priv->urb_num, IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data)); + __dump_ep_desc(&(TxCtrlEPList[epid])); + __dump_ept_data(epid); + } + } + } + local_irq_restore(flags); +} + +/* This function goes through all epids that are setup for Out Isoc transfers + and marks (isoc_out_done) all queued URBs that the DMA has finished + transfer for. + No URB completetion is done here to make interrupt routine return quickly. + URBs are completed later with help of complete_isoc_bottom_half() that + becomes schedules when this functions is finished. */ +static void check_finished_isoc_tx_epids(void) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + struct USB_SB_Desc* sb_desc; + int epid_done; + + /* Protect TxIsocEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if (TxIsocEPList[epid].sub == 0 || epid == INVALID_EPID || + !epid_out_traffic(epid)) { + /* Nothing here to see. */ + continue; + } + ASSERT(epid_inuse(epid)); + ASSERT(epid_isoc(epid)); + + sb_desc = phys_to_virt(TxIsocEPList[epid].sub); + /* Find the last descriptor of the currently active URB for this ep. + This is the first descriptor in the sub list marked for a descriptor + interrupt. */ + while (sb_desc && !IO_EXTRACT(USB_SB_command, intr, sb_desc->command)) { + sb_desc = sb_desc->next ? phys_to_virt(sb_desc->next) : 0; + } + ASSERT(sb_desc); + + isoc_dbg("Descr IRQ checking epid:%d sub:0x%x intr:0x%x\n", + epid, (unsigned int)phys_to_virt(TxIsocEPList[epid].sub), + (unsigned int)sb_desc); + + urb = activeUrbList[epid]; + if(urb == NULL) { + isoc_err("Isoc Descr irq on epid:%d with no active URB\n", epid); + continue; + } + + epid_done = 0; + while(urb && !epid_done) { + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS); + ASSERT(usb_pipeout(urb->pipe)); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + ASSERT(urb_priv->urb_state == STARTED || + urb_priv->urb_state == UNLINK); + + if (sb_desc != urb_priv->last_sb) { + /* This urb has been sent. */ + urb_priv->isoc_out_done = 1; + + } else { /* Found URB that has last_sb as the interrupt reason */ + + /* Check if EP has been disabled, meaning that all transfers are done*/ + if(!(TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable))) { + ASSERT((sb_desc->command & IO_MASK(USB_SB_command, eol)) == + IO_STATE(USB_SB_command, eol, yes)); + ASSERT(sb_desc->next == 0); + urb_priv->isoc_out_done = 1; + } else { + isoc_dbg("Skipping URB:0x%x[%d] because EP not disabled yet\n", + (unsigned int)urb, urb_priv->urb_num); + } + /* Stop looking any further in queue */ + epid_done = 1; + } + + if (!epid_done) { + if(urb == activeUrbList[epid]) { + urb = urb_list_first(epid); + } else { + urb = urb_list_next(urb, epid); + } + } + } /* END: while(urb && !epid_done) */ + } + + local_irq_restore(flags); +} + + +/* This is where the Out Isoc URBs are realy completed. This function is + scheduled from tc_dma_tx_interrupt() when one or more Out Isoc transfers + are done. This functions completes all URBs earlier marked with + isoc_out_done by fast interrupt routine check_finished_isoc_tx_epids() */ + +static void complete_isoc_bottom_half(struct work_struct* work) { + struct crisv10_isoc_complete_data *comp_data; + struct usb_iso_packet_descriptor *packet; + struct crisv10_urb_priv * urb_priv; + unsigned long flags; + struct urb* urb; + int epid_done; + int epid; + int i; + + comp_data = container_of(work, struct crisv10_isoc_complete_data, usb_bh); + + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS - 1; epid++) { + if(!epid_inuse(epid) || !epid_isoc(epid) || !epid_out_traffic(epid) || epid == DUMMY_EPID) { + /* Only check valid Out Isoc epids */ + continue; + } + + isoc_dbg("Isoc bottom-half checking epid:%d, sub:0x%x\n", epid, + (unsigned int)phys_to_virt(TxIsocEPList[epid].sub)); + + /* The descriptor interrupt handler has marked all transmitted Out Isoc + URBs with isoc_out_done. Now we traverse all epids and for all that + have out Isoc traffic we traverse its URB list and complete the + transmitted URBs. */ + epid_done = 0; + while (!epid_done) { + + /* Get the active urb (if any) */ + urb = activeUrbList[epid]; + if (urb == 0) { + isoc_dbg("No active URB on epid:%d anymore\n", epid); + epid_done = 1; + continue; + } + + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS); + ASSERT(usb_pipeout(urb->pipe)); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + if (!(urb_priv->isoc_out_done)) { + /* We have reached URB that isn't flaged done yet, stop traversing. */ + isoc_dbg("Stoped traversing Out Isoc URBs on epid:%d" + " before not yet flaged URB:0x%x[%d]\n", + epid, (unsigned int)urb, urb_priv->urb_num); + epid_done = 1; + continue; + } + + /* This urb has been sent. */ + isoc_dbg("Found URB:0x%x[%d] that is flaged isoc_out_done\n", + (unsigned int)urb, urb_priv->urb_num); + + /* Set ok on transfered packets for this URB and finish it */ + for (i = 0; i < urb->number_of_packets; i++) { + packet = &urb->iso_frame_desc[i]; + packet->status = 0; + packet->actual_length = packet->length; + } + urb_priv->isoc_packet_counter = urb->number_of_packets; + tc_finish_urb(comp_data->hcd, urb, 0); + + } /* END: while(!epid_done) */ + } /* END: for(epid...) */ + + local_irq_restore(flags); + kmem_cache_free(isoc_compl_cache, comp_data); +} + + +static void check_finished_intr_tx_epids(struct usb_hcd *hcd) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + volatile struct USB_EP_Desc *curr_ep; /* Current EP, the iterator. */ + volatile struct USB_EP_Desc *next_ep; /* The EP after current. */ + + /* Protect TxintrEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if(!epid_inuse(epid) || !epid_intr(epid) || !epid_out_traffic(epid)) { + /* Nothing to see on this epid. Only check valid Out Intr epids */ + continue; + } + + urb = activeUrbList[epid]; + if(urb == 0) { + intr_warn("Found Out Intr epid:%d with no active URB\n", epid); + continue; + } + + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_INTERRUPT); + ASSERT(usb_pipeout(urb->pipe)); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + /* Go through EPs between first and second sof-EP. It's here Out Intr EPs + are inserted.*/ + curr_ep = &TxIntrEPList[0]; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if(next_ep == urb_priv->intr_ep_pool[0]) { + /* We found the Out Intr EP for this epid */ + + /* Disable it so it doesn't get processed again */ + next_ep->command &= ~IO_MASK(USB_EP_command, enable); + + /* Finish the active Out Intr URB with status OK */ + tc_finish_urb(hcd, urb, 0); + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != &TxIntrEPList[1]); + + } + local_irq_restore(flags); +} + +/* Interrupt handler for DMA8/IRQ24 with subchannels (called from hardware intr) */ +static irqreturn_t tc_dma_tx_interrupt(int irq, void *vhc) { + struct usb_hcd *hcd = (struct usb_hcd*)vhc; + ASSERT(hcd); + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub0_descr)) { + /* Clear this interrupt */ + *R_DMA_CH8_SUB0_CLR_INTR = IO_STATE(R_DMA_CH8_SUB0_CLR_INTR, clr_descr, do); + restart_dma8_sub0(); + } + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub1_descr)) { + /* Clear this interrupt */ + *R_DMA_CH8_SUB1_CLR_INTR = IO_STATE(R_DMA_CH8_SUB1_CLR_INTR, clr_descr, do); + check_finished_ctrl_tx_epids(hcd); + } + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub2_descr)) { + /* Clear this interrupt */ + *R_DMA_CH8_SUB2_CLR_INTR = IO_STATE(R_DMA_CH8_SUB2_CLR_INTR, clr_descr, do); + check_finished_intr_tx_epids(hcd); + } + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub3_descr)) { + struct crisv10_isoc_complete_data* comp_data; + + /* Flag done Out Isoc for later completion */ + check_finished_isoc_tx_epids(); + + /* Clear this interrupt */ + *R_DMA_CH8_SUB3_CLR_INTR = IO_STATE(R_DMA_CH8_SUB3_CLR_INTR, clr_descr, do); + /* Schedule bottom half of Out Isoc completion function. This function + finishes the URBs marked with isoc_out_done */ + comp_data = (struct crisv10_isoc_complete_data*) + kmem_cache_alloc(isoc_compl_cache, GFP_ATOMIC); + ASSERT(comp_data != NULL); + comp_data ->hcd = hcd; + + INIT_WORK(&comp_data->usb_bh, complete_isoc_bottom_half); + schedule_work(&comp_data->usb_bh); + } + + return IRQ_HANDLED; +} + +/* Interrupt handler for DMA9/IRQ25 (called from hardware intr) */ +static irqreturn_t tc_dma_rx_interrupt(int irq, void *vhc) { + unsigned long flags; + struct urb *urb; + struct usb_hcd *hcd = (struct usb_hcd*)vhc; + struct crisv10_urb_priv *urb_priv; + int epid = 0; + int real_error; + + ASSERT(hcd); + + /* Clear this interrupt. */ + *R_DMA_CH9_CLR_INTR = IO_STATE(R_DMA_CH9_CLR_INTR, clr_eop, do); + + /* Custom clear interrupt for this interrupt */ + /* The reason we cli here is that we call the driver's callback functions. */ + local_irq_save(flags); + + /* Note that this while loop assumes that all packets span only + one rx descriptor. */ + while(myNextRxDesc->status & IO_MASK(USB_IN_status, eop)) { + epid = IO_EXTRACT(USB_IN_status, epid, myNextRxDesc->status); + /* Get the active URB for this epid */ + urb = activeUrbList[epid]; + + ASSERT(epid_inuse(epid)); + if (!urb) { + dma_err("No urb for epid %d in rx interrupt\n", epid); + goto skip_out; + } + + /* Check if any errors on epid */ + real_error = 0; + if (myNextRxDesc->status & IO_MASK(USB_IN_status, error)) { + __u32 r_usb_ept_data; + + if (usb_pipeisoc(urb->pipe)) { + r_usb_ept_data = etrax_epid_iso_get(epid); + if((r_usb_ept_data & IO_MASK(R_USB_EPT_DATA_ISO, valid)) && + (IO_EXTRACT(R_USB_EPT_DATA_ISO, error_code, r_usb_ept_data) == 0) && + (myNextRxDesc->status & IO_MASK(USB_IN_status, nodata))) { + /* Not an error, just a failure to receive an expected iso + in packet in this frame. This is not documented + in the designers reference. Continue processing. + */ + } else real_error = 1; + } else real_error = 1; + } + + if(real_error) { + dma_err("Error in RX descr on epid:%d for URB 0x%x", + epid, (unsigned int)urb); + dump_ept_data(epid); + dump_in_desc(myNextRxDesc); + goto skip_out; + } + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + ASSERT(urb_priv->urb_state == STARTED || + urb_priv->urb_state == UNLINK); + + if ((usb_pipetype(urb->pipe) == PIPE_BULK) || + (usb_pipetype(urb->pipe) == PIPE_CONTROL) || + (usb_pipetype(urb->pipe) == PIPE_INTERRUPT)) { + + /* We get nodata for empty data transactions, and the rx descriptor's + hw_len field is not valid in that case. No data to copy in other + words. */ + if (myNextRxDesc->status & IO_MASK(USB_IN_status, nodata)) { + /* No data to copy */ + } else { + /* + dma_dbg("Processing RX for URB:0x%x epid:%d (data:%d ofs:%d)\n", + (unsigned int)urb, epid, myNextRxDesc->hw_len, + urb_priv->rx_offset); + */ + /* Only copy data if URB isn't flaged to be unlinked*/ + if(urb_priv->urb_state != UNLINK) { + /* Make sure the data fits in the buffer. */ + if(urb_priv->rx_offset + myNextRxDesc->hw_len + <= urb->transfer_buffer_length) { + + /* Copy the data to URBs buffer */ + memcpy(urb->transfer_buffer + urb_priv->rx_offset, + phys_to_virt(myNextRxDesc->buf), myNextRxDesc->hw_len); + urb_priv->rx_offset += myNextRxDesc->hw_len; + } else { + /* Signal overflow when returning URB */ + urb->status = -EOVERFLOW; + tc_finish_urb_later(hcd, urb, urb->status); + } + } + } + + /* Check if it was the last packet in the transfer */ + if (myNextRxDesc->status & IO_MASK(USB_IN_status, eot)) { + /* Special handling for In Ctrl URBs. */ + if(usb_pipecontrol(urb->pipe) && usb_pipein(urb->pipe) && + !(urb_priv->ctrl_zout_done)) { + /* Flag that RX part of Ctrl transfer is done. Because zout descr + interrupt hasn't happend yet will the URB be finished in the + TX-Interrupt. */ + urb_priv->ctrl_rx_done = 1; + tc_dbg("Not finishing In Ctrl URB:0x%x from rx_interrupt, waiting" + " for zout\n", (unsigned int)urb); + } else { + tc_finish_urb(hcd, urb, 0); + } + } + } else { /* ISOC RX */ + /* + isoc_dbg("Processing RX for epid:%d (URB:0x%x) ISOC pipe\n", + epid, (unsigned int)urb); + */ + + struct usb_iso_packet_descriptor *packet; + + if (urb_priv->urb_state == UNLINK) { + isoc_warn("Ignoring Isoc Rx data for urb being unlinked.\n"); + goto skip_out; + } else if (urb_priv->urb_state == NOT_STARTED) { + isoc_err("What? Got Rx data for Isoc urb that isn't started?\n"); + goto skip_out; + } + + packet = &urb->iso_frame_desc[urb_priv->isoc_packet_counter]; + ASSERT(packet); + packet->status = 0; + + if (myNextRxDesc->status & IO_MASK(USB_IN_status, nodata)) { + /* We get nodata for empty data transactions, and the rx descriptor's + hw_len field is not valid in that case. We copy 0 bytes however to + stay in synch. */ + packet->actual_length = 0; + } else { + packet->actual_length = myNextRxDesc->hw_len; + /* Make sure the data fits in the buffer. */ + ASSERT(packet->actual_length <= packet->length); + memcpy(urb->transfer_buffer + packet->offset, + phys_to_virt(myNextRxDesc->buf), packet->actual_length); + if(packet->actual_length > 0) + isoc_dbg("Copied %d bytes, packet %d for URB:0x%x[%d]\n", + packet->actual_length, urb_priv->isoc_packet_counter, + (unsigned int)urb, urb_priv->urb_num); + } + + /* Increment the packet counter. */ + urb_priv->isoc_packet_counter++; + + /* Note that we don't care about the eot field in the rx descriptor's + status. It will always be set for isoc traffic. */ + if (urb->number_of_packets == urb_priv->isoc_packet_counter) { + /* Complete the urb with status OK. */ + tc_finish_urb(hcd, urb, 0); + } + } + + skip_out: + myNextRxDesc->status = 0; + myNextRxDesc->command |= IO_MASK(USB_IN_command, eol); + myLastRxDesc->command &= ~IO_MASK(USB_IN_command, eol); + myLastRxDesc = myNextRxDesc; + myNextRxDesc = phys_to_virt(myNextRxDesc->next); + flush_etrax_cache(); + *R_DMA_CH9_CMD = IO_STATE(R_DMA_CH9_CMD, cmd, restart); + } + + local_irq_restore(flags); + + return IRQ_HANDLED; +} + +static void tc_bulk_start_timer_func(unsigned long dummy) { + /* We might enable an EP descriptor behind the current DMA position when + it's about to decide that there are no more bulk traffic and it should + stop the bulk channel. + Therefore we periodically check if the bulk channel is stopped and there + is an enabled bulk EP descriptor, in which case we start the bulk + channel. */ + + if (!(*R_DMA_CH8_SUB0_CMD & IO_MASK(R_DMA_CH8_SUB0_CMD, cmd))) { + int epid; + + timer_dbg("bulk_start_timer: Bulk DMA channel not running.\n"); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if (TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + timer_warn("Found enabled EP for epid %d, starting bulk channel.\n", + epid); + restart_dma8_sub0(); + + /* Restart the bulk eot timer since we just started the bulk channel.*/ + mod_timer(&bulk_eot_timer, jiffies + BULK_EOT_TIMER_INTERVAL); + + /* No need to search any further. */ + break; + } + } + } else { + timer_dbg("bulk_start_timer: Bulk DMA channel running.\n"); + } +} + +static void tc_bulk_eot_timer_func(unsigned long dummy) { + struct usb_hcd *hcd = (struct usb_hcd*)dummy; + ASSERT(hcd); + /* Because of a race condition in the top half, we might miss a bulk eot. + This timer "simulates" a bulk eot if we don't get one for a while, + hopefully correcting the situation. */ + timer_dbg("bulk_eot_timer timed out.\n"); + check_finished_bulk_tx_epids(hcd, 1); +} + + +/*************************************************************/ +/*************************************************************/ +/* Device driver block */ +/*************************************************************/ +/*************************************************************/ + +/* Forward declarations for device driver functions */ +static int devdrv_hcd_probe(struct device *); +static int devdrv_hcd_remove(struct device *); +#ifdef CONFIG_PM +static int devdrv_hcd_suspend(struct device *, u32, u32); +static int devdrv_hcd_resume(struct device *, u32); +#endif /* CONFIG_PM */ + +/* the device */ +static struct platform_device *devdrv_hc_platform_device; + +/* device driver interface */ +static struct device_driver devdrv_hc_device_driver = { + .name = (char *) hc_name, + .bus = &platform_bus_type, + + .probe = devdrv_hcd_probe, + .remove = devdrv_hcd_remove, + +#ifdef CONFIG_PM + .suspend = devdrv_hcd_suspend, + .resume = devdrv_hcd_resume, +#endif /* CONFIG_PM */ +}; + +/* initialize the host controller and driver */ +static int __init_or_module devdrv_hcd_probe(struct device *dev) +{ + struct usb_hcd *hcd; + struct crisv10_hcd *crisv10_hcd; + int retval; + + /* Check DMA burst length */ + if(IO_EXTRACT(R_BUS_CONFIG, dma_burst, *R_BUS_CONFIG) != + IO_STATE(R_BUS_CONFIG, dma_burst, burst32)) { + devdrv_err("Invalid DMA burst length in Etrax 100LX," + " needs to be 32\n"); + return -EPERM; + } + + //XXX: dev->usb_id don't exist, using "" instread? - claudio + hcd = usb_create_hcd(&crisv10_hc_driver, dev, ""); + if (!hcd) + return -ENOMEM; + + crisv10_hcd = hcd_to_crisv10_hcd(hcd); + spin_lock_init(&crisv10_hcd->lock); + crisv10_hcd->num_ports = num_ports(); + crisv10_hcd->running = 0; + + dev_set_drvdata(dev, crisv10_hcd); + + devdrv_dbg("ETRAX USB IRQs HC:%d RX:%d TX:%d\n", ETRAX_USB_HC_IRQ, + ETRAX_USB_RX_IRQ, ETRAX_USB_TX_IRQ); + + /* Print out chip version read from registers */ + int rev_maj = *R_USB_REVISION & IO_MASK(R_USB_REVISION, major); + int rev_min = *R_USB_REVISION & IO_MASK(R_USB_REVISION, minor); + if(rev_min == 0) { + devdrv_info("Etrax 100LX USB Revision %d v1,2\n", rev_maj); + } else { + devdrv_info("Etrax 100LX USB Revision %d v%d\n", rev_maj, rev_min); + } + + devdrv_info("Bulk timer interval, start:%d eot:%d\n", + BULK_START_TIMER_INTERVAL, + BULK_EOT_TIMER_INTERVAL); + + + /* Init root hub data structures */ + if(rh_init()) { + devdrv_err("Failed init data for Root Hub\n"); + retval = -ENOMEM; + } + + if(port_in_use(0)) { + if (cris_request_io_interface(if_usb_1, "ETRAX100LX USB-HCD")) { + printk(KERN_CRIT "usb-host: request IO interface usb1 failed"); + retval = -EBUSY; + goto out; + } + devdrv_info("Claimed interface for USB physical port 1\n"); + } + if(port_in_use(1)) { + if (cris_request_io_interface(if_usb_2, "ETRAX100LX USB-HCD")) { + /* Free first interface if second failed to be claimed */ + if(port_in_use(0)) { + cris_free_io_interface(if_usb_1); + } + printk(KERN_CRIT "usb-host: request IO interface usb2 failed"); + retval = -EBUSY; + goto out; + } + devdrv_info("Claimed interface for USB physical port 2\n"); + } + + /* Init transfer controller structs and locks */ + if((retval = tc_init(hcd)) != 0) { + goto out; + } + + /* Attach interrupt functions for DMA and init DMA controller */ + if((retval = tc_dma_init(hcd)) != 0) { + goto out; + } + + /* Attach the top IRQ handler for USB controller interrupts */ + if (request_irq(ETRAX_USB_HC_IRQ, crisv10_hcd_top_irq, 0, + "ETRAX 100LX built-in USB (HC)", hcd)) { + err("Could not allocate IRQ %d for USB", ETRAX_USB_HC_IRQ); + retval = -EBUSY; + goto out; + } + + /* iso_eof is only enabled when isoc traffic is running. */ + *R_USB_IRQ_MASK_SET = + /* IO_STATE(R_USB_IRQ_MASK_SET, iso_eof, set) | */ + IO_STATE(R_USB_IRQ_MASK_SET, bulk_eot, set) | + IO_STATE(R_USB_IRQ_MASK_SET, epid_attn, set) | + IO_STATE(R_USB_IRQ_MASK_SET, port_status, set) | + IO_STATE(R_USB_IRQ_MASK_SET, ctl_status, set); + + + crisv10_ready_wait(); + /* Reset the USB interface. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, reset); + + /* Designer's Reference, p. 8 - 10 says we should Initate R_USB_FM_PSTART to + 0x2A30 (10800), to guarantee that control traffic gets 10% of the + bandwidth, and periodic transfer may allocate the rest (90%). + This doesn't work though. + The value 11960 is chosen to be just after the SOF token, with a couple + of bit times extra for possible bit stuffing. */ + *R_USB_FM_PSTART = IO_FIELD(R_USB_FM_PSTART, value, 11960); + + crisv10_ready_wait(); + /* Configure the USB interface as a host controller. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_config); + + + /* Check so controller not busy before enabling ports */ + crisv10_ready_wait(); + + /* Enable selected USB ports */ + if(port_in_use(0)) { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, no); + } else { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, yes); + } + if(port_in_use(1)) { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, no); + } else { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, yes); + } + + crisv10_ready_wait(); + /* Start processing of USB traffic. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + + /* Do not continue probing initialization before USB interface is done */ + crisv10_ready_wait(); + + /* Register our Host Controller to USB Core + * Finish the remaining parts of generic HCD initialization: allocate the + * buffers of consistent memory, register the bus + * and call the driver's reset() and start() routines. */ + retval = usb_add_hcd(hcd, ETRAX_USB_HC_IRQ, IRQF_DISABLED); + if (retval != 0) { + devdrv_err("Failed registering HCD driver\n"); + goto out; + } + + return 0; + + out: + devdrv_hcd_remove(dev); + return retval; +} + + +/* cleanup after the host controller and driver */ +static int __init_or_module devdrv_hcd_remove(struct device *dev) +{ + struct crisv10_hcd *crisv10_hcd = dev_get_drvdata(dev); + struct usb_hcd *hcd; + + if (!crisv10_hcd) + return 0; + hcd = crisv10_hcd_to_hcd(crisv10_hcd); + + + /* Stop USB Controller in Etrax 100LX */ + crisv10_hcd_reset(hcd); + + usb_remove_hcd(hcd); + devdrv_dbg("Removed HCD from USB Core\n"); + + /* Free USB Controller IRQ */ + free_irq(ETRAX_USB_HC_IRQ, NULL); + + /* Free resources */ + tc_dma_destroy(); + tc_destroy(); + + + if(port_in_use(0)) { + cris_free_io_interface(if_usb_1); + } + if(port_in_use(1)) { + cris_free_io_interface(if_usb_2); + } + + devdrv_dbg("Freed all claimed resources\n"); + + return 0; +} + + +#ifdef CONFIG_PM + +static int devdrv_hcd_suspend(struct usb_hcd *hcd, u32 state, u32 level) +{ + return 0; /* no-op for now */ +} + +static int devdrv_hcd_resume(struct usb_hcd *hcd, u32 level) +{ + return 0; /* no-op for now */ +} + +#endif /* CONFIG_PM */ + + +/*************************************************************/ +/*************************************************************/ +/* Module block */ +/*************************************************************/ +/*************************************************************/ + +/* register driver */ +static int __init module_hcd_init(void) +{ + + if (usb_disabled()) + return -ENODEV; + + /* Here we select enabled ports by following defines created from + menuconfig */ +#ifndef CONFIG_ETRAX_USB_HOST_PORT1 + ports &= ~(1<<0); +#endif +#ifndef CONFIG_ETRAX_USB_HOST_PORT2 + ports &= ~(1<<1); +#endif + + printk(KERN_INFO "%s version "VERSION" "COPYRIGHT"\n", product_desc); + + devdrv_hc_platform_device = + platform_device_register_simple((char *) hc_name, 0, NULL, 0); + + if (IS_ERR(devdrv_hc_platform_device)) + return PTR_ERR(devdrv_hc_platform_device); + return driver_register(&devdrv_hc_device_driver); + /* + * Note that we do not set the DMA mask for the device, + * i.e. we pretend that we will use PIO, since no specific + * allocation routines are needed for DMA buffers. This will + * cause the HCD buffer allocation routines to fall back to + * kmalloc(). + */ +} + +/* unregister driver */ +static void __exit module_hcd_exit(void) +{ + driver_unregister(&devdrv_hc_device_driver); +} + + +/* Module hooks */ +module_init(module_hcd_init); +module_exit(module_hcd_exit); +/* + * + * ETRAX 100LX USB Host Controller Driver + * + * Copyright (C) 2005 - 2008 Axis Communications AB + * + * Author: Konrad Eriksson + * + */ + +#include +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include + +#include "../core/hcd.h" +#include "../core/hub.h" +#include "hc-crisv10.h" +#include "hc-cris-dbg.h" + + +/***************************************************************************/ +/***************************************************************************/ +/* Host Controller settings */ +/***************************************************************************/ +/***************************************************************************/ + +#define VERSION "1.00-openwrt_diff" +#define COPYRIGHT "(c) 2005, 2006 Axis Communications AB" +#define DESCRIPTION "ETRAX 100LX USB Host Controller" + +#define ETRAX_USB_HC_IRQ USB_HC_IRQ_NBR +#define ETRAX_USB_RX_IRQ USB_DMA_RX_IRQ_NBR +#define ETRAX_USB_TX_IRQ USB_DMA_TX_IRQ_NBR + +/* Number of physical ports in Etrax 100LX */ +#define USB_ROOT_HUB_PORTS 2 + +const char hc_name[] = "hc-crisv10"; +const char product_desc[] = DESCRIPTION; + +/* The number of epids is, among other things, used for pre-allocating + ctrl, bulk and isoc EP descriptors (one for each epid). + Assumed to be > 1 when initiating the DMA lists. */ +#define NBR_OF_EPIDS 32 + +/* Support interrupt traffic intervals up to 128 ms. */ +#define MAX_INTR_INTERVAL 128 + +/* If periodic traffic (intr or isoc) is to be used, then one entry in the EP + table must be "invalid". By this we mean that we shouldn't care about epid + attentions for this epid, or at least handle them differently from epid + attentions for "valid" epids. This define determines which one to use + (don't change it). */ +#define INVALID_EPID 31 +/* A special epid for the bulk dummys. */ +#define DUMMY_EPID 30 + +/* Module settings */ + +MODULE_DESCRIPTION(DESCRIPTION); +MODULE_LICENSE("GPL"); +MODULE_AUTHOR("Konrad Eriksson "); + + +/* Module parameters */ + +/* 0 = No ports enabled + 1 = Only port 1 enabled (on board ethernet on devboard) + 2 = Only port 2 enabled (external connector on devboard) + 3 = Both ports enabled +*/ +static unsigned int ports = 3; +module_param(ports, uint, S_IRUGO); +MODULE_PARM_DESC(ports, "Bitmask indicating USB ports to use"); + + +/***************************************************************************/ +/***************************************************************************/ +/* Shared global variables for this module */ +/***************************************************************************/ +/***************************************************************************/ + +/* EP descriptor lists for non period transfers. Must be 32-bit aligned. */ +static volatile struct USB_EP_Desc TxBulkEPList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); + +static volatile struct USB_EP_Desc TxCtrlEPList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); + +/* EP descriptor lists for period transfers. Must be 32-bit aligned. */ +static volatile struct USB_EP_Desc TxIntrEPList[MAX_INTR_INTERVAL] __attribute__ ((aligned (4))); +static volatile struct USB_SB_Desc TxIntrSB_zout __attribute__ ((aligned (4))); + +static volatile struct USB_EP_Desc TxIsocEPList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); +static volatile struct USB_SB_Desc TxIsocSB_zout __attribute__ ((aligned (4))); + +static volatile struct USB_SB_Desc TxIsocSBList[NBR_OF_EPIDS] __attribute__ ((aligned (4))); + +/* After each enabled bulk EP IN we put two disabled EP descriptors with the eol flag set, + causing the DMA to stop the DMA channel. The first of these two has the intr flag set, which + gives us a dma8_sub0_descr interrupt. When we receive this, we advance the DMA one step in the + EP list and then restart the bulk channel, thus forcing a switch between bulk EP descriptors + in each frame. */ +static volatile struct USB_EP_Desc TxBulkDummyEPList[NBR_OF_EPIDS][2] __attribute__ ((aligned (4))); + +/* List of URB pointers, where each points to the active URB for a epid. + For Bulk, Ctrl and Intr this means which URB that currently is added to + DMA lists (Isoc URBs are all directly added to DMA lists). As soon as + URB has completed is the queue examined and the first URB in queue is + removed and moved to the activeUrbList while its state change to STARTED and + its transfer(s) gets added to DMA list (exception Isoc where URBs enter + state STARTED directly and added transfers added to DMA lists). */ +static struct urb *activeUrbList[NBR_OF_EPIDS]; + +/* Additional software state info for each epid */ +static struct etrax_epid epid_state[NBR_OF_EPIDS]; + +/* Timer handles for bulk traffic timer used to avoid DMA bug where DMA stops + even if there is new data waiting to be processed */ +static struct timer_list bulk_start_timer = TIMER_INITIALIZER(NULL, 0, 0); +static struct timer_list bulk_eot_timer = TIMER_INITIALIZER(NULL, 0, 0); + +/* We want the start timer to expire before the eot timer, because the former + might start traffic, thus making it unnecessary for the latter to time + out. */ +#define BULK_START_TIMER_INTERVAL (HZ/50) /* 20 ms */ +#define BULK_EOT_TIMER_INTERVAL (HZ/16) /* 60 ms */ + +/* Delay before a URB completion happen when it's scheduled to be delayed */ +#define LATER_TIMER_DELAY (HZ/50) /* 20 ms */ + +/* Simplifying macros for checking software state info of a epid */ +/* ----------------------------------------------------------------------- */ +#define epid_inuse(epid) epid_state[epid].inuse +#define epid_out_traffic(epid) epid_state[epid].out_traffic +#define epid_isoc(epid) (epid_state[epid].type == PIPE_ISOCHRONOUS ? 1 : 0) +#define epid_intr(epid) (epid_state[epid].type == PIPE_INTERRUPT ? 1 : 0) + + +/***************************************************************************/ +/***************************************************************************/ +/* DEBUG FUNCTIONS */ +/***************************************************************************/ +/***************************************************************************/ +/* Note that these functions are always available in their "__" variants, + for use in error situations. The "__" missing variants are controlled by + the USB_DEBUG_DESC/USB_DEBUG_URB macros. */ +static void __dump_urb(struct urb* purb) +{ + struct crisv10_urb_priv *urb_priv = purb->hcpriv; + int urb_num = -1; + if(urb_priv) { + urb_num = urb_priv->urb_num; + } + printk("\nURB:0x%x[%d]\n", (unsigned int)purb, urb_num); + printk("dev :0x%08lx\n", (unsigned long)purb->dev); + printk("pipe :0x%08x\n", purb->pipe); + printk("status :%d\n", purb->status); + printk("transfer_flags :0x%08x\n", purb->transfer_flags); + printk("transfer_buffer :0x%08lx\n", (unsigned long)purb->transfer_buffer); + printk("transfer_buffer_length:%d\n", purb->transfer_buffer_length); + printk("actual_length :%d\n", purb->actual_length); + printk("setup_packet :0x%08lx\n", (unsigned long)purb->setup_packet); + printk("start_frame :%d\n", purb->start_frame); + printk("number_of_packets :%d\n", purb->number_of_packets); + printk("interval :%d\n", purb->interval); + printk("error_count :%d\n", purb->error_count); + printk("context :0x%08lx\n", (unsigned long)purb->context); + printk("complete :0x%08lx\n\n", (unsigned long)purb->complete); +} + +static void __dump_in_desc(volatile struct USB_IN_Desc *in) +{ + printk("\nUSB_IN_Desc at 0x%08lx\n", (unsigned long)in); + printk(" sw_len : 0x%04x (%d)\n", in->sw_len, in->sw_len); + printk(" command : 0x%04x\n", in->command); + printk(" next : 0x%08lx\n", in->next); + printk(" buf : 0x%08lx\n", in->buf); + printk(" hw_len : 0x%04x (%d)\n", in->hw_len, in->hw_len); + printk(" status : 0x%04x\n\n", in->status); +} + +static void __dump_sb_desc(volatile struct USB_SB_Desc *sb) +{ + char tt = (sb->command & 0x30) >> 4; + char *tt_string; + + switch (tt) { + case 0: + tt_string = "zout"; + break; + case 1: + tt_string = "in"; + break; + case 2: + tt_string = "out"; + break; + case 3: + tt_string = "setup"; + break; + default: + tt_string = "unknown (weird)"; + } + + printk(" USB_SB_Desc at 0x%08lx ", (unsigned long)sb); + printk(" command:0x%04x (", sb->command); + printk("rem:%d ", (sb->command & 0x3f00) >> 8); + printk("full:%d ", (sb->command & 0x40) >> 6); + printk("tt:%d(%s) ", tt, tt_string); + printk("intr:%d ", (sb->command & 0x8) >> 3); + printk("eot:%d ", (sb->command & 0x2) >> 1); + printk("eol:%d)", sb->command & 0x1); + printk(" sw_len:0x%04x(%d)", sb->sw_len, sb->sw_len); + printk(" next:0x%08lx", sb->next); + printk(" buf:0x%08lx\n", sb->buf); +} + + +static void __dump_ep_desc(volatile struct USB_EP_Desc *ep) +{ + printk("USB_EP_Desc at 0x%08lx ", (unsigned long)ep); + printk(" command:0x%04x (", ep->command); + printk("ep_id:%d ", (ep->command & 0x1f00) >> 8); + printk("enable:%d ", (ep->command & 0x10) >> 4); + printk("intr:%d ", (ep->command & 0x8) >> 3); + printk("eof:%d ", (ep->command & 0x2) >> 1); + printk("eol:%d)", ep->command & 0x1); + printk(" hw_len:0x%04x(%d)", ep->hw_len, ep->hw_len); + printk(" next:0x%08lx", ep->next); + printk(" sub:0x%08lx\n", ep->sub); +} + +static inline void __dump_ep_list(int pipe_type) +{ + volatile struct USB_EP_Desc *ep; + volatile struct USB_EP_Desc *first_ep; + volatile struct USB_SB_Desc *sb; + + switch (pipe_type) + { + case PIPE_BULK: + first_ep = &TxBulkEPList[0]; + break; + case PIPE_CONTROL: + first_ep = &TxCtrlEPList[0]; + break; + case PIPE_INTERRUPT: + first_ep = &TxIntrEPList[0]; + break; + case PIPE_ISOCHRONOUS: + first_ep = &TxIsocEPList[0]; + break; + default: + return; + } + ep = first_ep; + + printk("\n\nDumping EP list...\n\n"); + + do { + __dump_ep_desc(ep); + /* Cannot phys_to_virt on 0 as it turns into 80000000, which is != 0. */ + sb = ep->sub ? phys_to_virt(ep->sub) : 0; + while (sb) { + __dump_sb_desc(sb); + sb = sb->next ? phys_to_virt(sb->next) : 0; + } + ep = (volatile struct USB_EP_Desc *)(phys_to_virt(ep->next)); + + } while (ep != first_ep); +} + +static inline void __dump_ept_data(int epid) +{ + unsigned long flags; + __u32 r_usb_ept_data; + + if (epid < 0 || epid > 31) { + printk("Cannot dump ept data for invalid epid %d\n", epid); + return; + } + + local_irq_save(flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, epid); + nop(); + r_usb_ept_data = *R_USB_EPT_DATA; + local_irq_restore(flags); + + printk(" R_USB_EPT_DATA = 0x%x for epid %d :\n", r_usb_ept_data, epid); + if (r_usb_ept_data == 0) { + /* No need for more detailed printing. */ + return; + } + printk(" valid : %d\n", (r_usb_ept_data & 0x80000000) >> 31); + printk(" hold : %d\n", (r_usb_ept_data & 0x40000000) >> 30); + printk(" error_count_in : %d\n", (r_usb_ept_data & 0x30000000) >> 28); + printk(" t_in : %d\n", (r_usb_ept_data & 0x08000000) >> 27); + printk(" low_speed : %d\n", (r_usb_ept_data & 0x04000000) >> 26); + printk(" port : %d\n", (r_usb_ept_data & 0x03000000) >> 24); + printk(" error_code : %d\n", (r_usb_ept_data & 0x00c00000) >> 22); + printk(" t_out : %d\n", (r_usb_ept_data & 0x00200000) >> 21); + printk(" error_count_out : %d\n", (r_usb_ept_data & 0x00180000) >> 19); + printk(" max_len : %d\n", (r_usb_ept_data & 0x0003f800) >> 11); + printk(" ep : %d\n", (r_usb_ept_data & 0x00000780) >> 7); + printk(" dev : %d\n", (r_usb_ept_data & 0x0000003f)); +} + +static inline void __dump_ept_data_iso(int epid) +{ + unsigned long flags; + __u32 ept_data; + + if (epid < 0 || epid > 31) { + printk("Cannot dump ept data for invalid epid %d\n", epid); + return; + } + + local_irq_save(flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, epid); + nop(); + ept_data = *R_USB_EPT_DATA_ISO; + local_irq_restore(flags); + + printk(" R_USB_EPT_DATA = 0x%x for epid %d :\n", ept_data, epid); + if (ept_data == 0) { + /* No need for more detailed printing. */ + return; + } + printk(" valid : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, valid, + ept_data)); + printk(" port : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, port, + ept_data)); + printk(" error_code : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, error_code, + ept_data)); + printk(" max_len : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, max_len, + ept_data)); + printk(" ep : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, ep, + ept_data)); + printk(" dev : %d\n", IO_EXTRACT(R_USB_EPT_DATA_ISO, dev, + ept_data)); +} + +static inline void __dump_ept_data_list(void) +{ + int i; + + printk("Dumping the whole R_USB_EPT_DATA list\n"); + + for (i = 0; i < 32; i++) { + __dump_ept_data(i); + } +} + +static void debug_epid(int epid) { + int i; + + if(epid_isoc(epid)) { + __dump_ept_data_iso(epid); + } else { + __dump_ept_data(epid); + } + + printk("Bulk:\n"); + for(i = 0; i < 32; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxBulkEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxBulkEPList[i])); + } + } + + printk("Ctrl:\n"); + for(i = 0; i < 32; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxCtrlEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxCtrlEPList[i])); + } + } + + printk("Intr:\n"); + for(i = 0; i < MAX_INTR_INTERVAL; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxIntrEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxIntrEPList[i])); + } + } + + printk("Isoc:\n"); + for(i = 0; i < 32; i++) { + if(IO_EXTRACT(USB_EP_command, epid, TxIsocEPList[i].command) == + epid) { + printk("%d: ", i); __dump_ep_desc(&(TxIsocEPList[i])); + } + } + + __dump_ept_data_list(); + __dump_ep_list(PIPE_INTERRUPT); + printk("\n\n"); +} + + + +char* hcd_status_to_str(__u8 bUsbStatus) { + static char hcd_status_str[128]; + hcd_status_str[0] = '\0'; + if(bUsbStatus & IO_STATE(R_USB_STATUS, ourun, yes)) { + strcat(hcd_status_str, "ourun "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, perror, yes)) { + strcat(hcd_status_str, "perror "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, device_mode, yes)) { + strcat(hcd_status_str, "device_mode "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, host_mode, yes)) { + strcat(hcd_status_str, "host_mode "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, started, yes)) { + strcat(hcd_status_str, "started "); + } + if(bUsbStatus & IO_STATE(R_USB_STATUS, running, yes)) { + strcat(hcd_status_str, "running "); + } + return hcd_status_str; +} + + +char* sblist_to_str(struct USB_SB_Desc* sb_desc) { + static char sblist_to_str_buff[128]; + char tmp[32], tmp2[32]; + sblist_to_str_buff[0] = '\0'; + while(sb_desc != NULL) { + switch(IO_EXTRACT(USB_SB_command, tt, sb_desc->command)) { + case 0: sprintf(tmp, "zout"); break; + case 1: sprintf(tmp, "in"); break; + case 2: sprintf(tmp, "out"); break; + case 3: sprintf(tmp, "setup"); break; + } + sprintf(tmp2, "(%s %d)", tmp, sb_desc->sw_len); + strcat(sblist_to_str_buff, tmp2); + if(sb_desc->next != 0) { + sb_desc = phys_to_virt(sb_desc->next); + } else { + sb_desc = NULL; + } + } + return sblist_to_str_buff; +} + +char* port_status_to_str(__u16 wPortStatus) { + static char port_status_str[128]; + port_status_str[0] = '\0'; + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, connected, yes)) { + strcat(port_status_str, "connected "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) { + strcat(port_status_str, "enabled "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, suspended, yes)) { + strcat(port_status_str, "suspended "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, reset, yes)) { + strcat(port_status_str, "reset "); + } + if(wPortStatus & IO_STATE(R_USB_RH_PORT_STATUS_1, speed, full)) { + strcat(port_status_str, "full-speed "); + } else { + strcat(port_status_str, "low-speed "); + } + return port_status_str; +} + + +char* endpoint_to_str(struct usb_endpoint_descriptor *ed) { + static char endpoint_to_str_buff[128]; + char tmp[32]; + int epnum = ed->bEndpointAddress & 0x0F; + int dir = ed->bEndpointAddress & 0x80; + int type = ed->bmAttributes & 0x03; + endpoint_to_str_buff[0] = '\0'; + sprintf(endpoint_to_str_buff, "ep:%d ", epnum); + switch(type) { + case 0: + sprintf(tmp, " ctrl"); + break; + case 1: + sprintf(tmp, " isoc"); + break; + case 2: + sprintf(tmp, " bulk"); + break; + case 3: + sprintf(tmp, " intr"); + break; + } + strcat(endpoint_to_str_buff, tmp); + if(dir) { + sprintf(tmp, " in"); + } else { + sprintf(tmp, " out"); + } + strcat(endpoint_to_str_buff, tmp); + + return endpoint_to_str_buff; +} + +/* Debug helper functions for Transfer Controller */ +char* pipe_to_str(unsigned int pipe) { + static char pipe_to_str_buff[128]; + char tmp[64]; + sprintf(pipe_to_str_buff, "dir:%s", str_dir(pipe)); + sprintf(tmp, " type:%s", str_type(pipe)); + strcat(pipe_to_str_buff, tmp); + + sprintf(tmp, " dev:%d", usb_pipedevice(pipe)); + strcat(pipe_to_str_buff, tmp); + sprintf(tmp, " ep:%d", usb_pipeendpoint(pipe)); + strcat(pipe_to_str_buff, tmp); + return pipe_to_str_buff; +} + + +#define USB_DEBUG_DESC 1 + +#ifdef USB_DEBUG_DESC +#define dump_in_desc(x) __dump_in_desc(x) +#define dump_sb_desc(...) __dump_sb_desc(...) +#define dump_ep_desc(x) __dump_ep_desc(x) +#define dump_ept_data(x) __dump_ept_data(x) +#else +#define dump_in_desc(...) do {} while (0) +#define dump_sb_desc(...) do {} while (0) +#define dump_ep_desc(...) do {} while (0) +#endif + + +/* Uncomment this to enable massive function call trace + #define USB_DEBUG_TRACE */ + +#ifdef USB_DEBUG_TRACE +#define DBFENTER (printk(": Entering: %s\n", __FUNCTION__)) +#define DBFEXIT (printk(": Exiting: %s\n", __FUNCTION__)) +#else +#define DBFENTER do {} while (0) +#define DBFEXIT do {} while (0) +#endif + +#define CHECK_ALIGN(x) if (((__u32)(x)) & 0x00000003) \ +{panic("Alignment check (DWORD) failed at %s:%s:%d\n", __FILE__, __FUNCTION__, __LINE__);} + +/* Most helpful debugging aid */ +#define ASSERT(expr) ((void) ((expr) ? 0 : (err("assert failed at: %s %d",__FUNCTION__, __LINE__)))) + + +/***************************************************************************/ +/***************************************************************************/ +/* Forward declarations */ +/***************************************************************************/ +/***************************************************************************/ +void crisv10_hcd_epid_attn_irq(struct crisv10_irq_reg *reg); +void crisv10_hcd_port_status_irq(struct crisv10_irq_reg *reg); +void crisv10_hcd_ctl_status_irq(struct crisv10_irq_reg *reg); +void crisv10_hcd_isoc_eof_irq(struct crisv10_irq_reg *reg); + +void rh_port_status_change(__u16[]); +int rh_clear_port_feature(__u8, __u16); +int rh_set_port_feature(__u8, __u16); +static void rh_disable_port(unsigned int port); + +static void check_finished_bulk_tx_epids(struct usb_hcd *hcd, + int timer); + +static int tc_setup_epid(struct usb_host_endpoint *ep, struct urb *urb, + int mem_flags); +static void tc_free_epid(struct usb_host_endpoint *ep); +static int tc_allocate_epid(void); +static void tc_finish_urb(struct usb_hcd *hcd, struct urb *urb, int status); +static void tc_finish_urb_later(struct usb_hcd *hcd, struct urb *urb, + int status); + +static int urb_priv_create(struct usb_hcd *hcd, struct urb *urb, int epid, + int mem_flags); +static void urb_priv_free(struct usb_hcd *hcd, struct urb *urb); + +static int crisv10_usb_check_bandwidth(struct usb_device *dev,struct urb *urb); +static void crisv10_usb_claim_bandwidth( + struct usb_device *dev, struct urb *urb, int bustime, int isoc); +static void crisv10_usb_release_bandwidth( + struct usb_hcd *hcd, int isoc, int bandwidth); + +static inline struct urb *urb_list_first(int epid); +static inline void urb_list_add(struct urb *urb, int epid, + int mem_flags); +static inline urb_entry_t *urb_list_entry(struct urb *urb, int epid); +static inline void urb_list_del(struct urb *urb, int epid); +static inline void urb_list_move_last(struct urb *urb, int epid); +static inline struct urb *urb_list_next(struct urb *urb, int epid); + +int create_sb_for_urb(struct urb *urb, int mem_flags); +int init_intr_urb(struct urb *urb, int mem_flags); + +static inline void etrax_epid_set(__u8 index, __u32 data); +static inline void etrax_epid_clear_error(__u8 index); +static inline void etrax_epid_set_toggle(__u8 index, __u8 dirout, + __u8 toggle); +static inline __u8 etrax_epid_get_toggle(__u8 index, __u8 dirout); +static inline __u32 etrax_epid_get(__u8 index); + +/* We're accessing the same register position in Etrax so + when we do full access the internal difference doesn't matter */ +#define etrax_epid_iso_set(index, data) etrax_epid_set(index, data) +#define etrax_epid_iso_get(index) etrax_epid_get(index) + + +static void tc_dma_process_isoc_urb(struct urb *urb); +static void tc_dma_process_queue(int epid); +static void tc_dma_unlink_intr_urb(struct urb *urb); +static irqreturn_t tc_dma_tx_interrupt(int irq, void *vhc); +static irqreturn_t tc_dma_rx_interrupt(int irq, void *vhc); + +static void tc_bulk_start_timer_func(unsigned long dummy); +static void tc_bulk_eot_timer_func(unsigned long dummy); + + +/*************************************************************/ +/*************************************************************/ +/* Host Controler Driver block */ +/*************************************************************/ +/*************************************************************/ + +/* HCD operations */ +static irqreturn_t crisv10_hcd_top_irq(int irq, void*); +static int crisv10_hcd_reset(struct usb_hcd *); +static int crisv10_hcd_start(struct usb_hcd *); +static void crisv10_hcd_stop(struct usb_hcd *); +#ifdef CONFIG_PM +static int crisv10_hcd_suspend(struct device *, u32, u32); +static int crisv10_hcd_resume(struct device *, u32); +#endif /* CONFIG_PM */ +static int crisv10_hcd_get_frame(struct usb_hcd *); + +static int tc_urb_enqueue(struct usb_hcd *, struct urb *, gfp_t mem_flags); +static int tc_urb_dequeue(struct usb_hcd *, struct urb *, int); +static void tc_endpoint_disable(struct usb_hcd *, struct usb_host_endpoint *ep); + +static int rh_status_data_request(struct usb_hcd *, char *); +static int rh_control_request(struct usb_hcd *, u16, u16, u16, char*, u16); + +#ifdef CONFIG_PM +static int crisv10_hcd_hub_suspend(struct usb_hcd *); +static int crisv10_hcd_hub_resume(struct usb_hcd *); +#endif /* CONFIG_PM */ +#ifdef CONFIG_USB_OTG +static int crisv10_hcd_start_port_reset(struct usb_hcd *, unsigned); +#endif /* CONFIG_USB_OTG */ + +/* host controller driver interface */ +static const struct hc_driver crisv10_hc_driver = + { + .description = hc_name, + .product_desc = product_desc, + .hcd_priv_size = sizeof(struct crisv10_hcd), + + /* Attaching IRQ handler manualy in probe() */ + /* .irq = crisv10_hcd_irq, */ + + .flags = HCD_USB11, + + /* called to init HCD and root hub */ + .reset = crisv10_hcd_reset, + .start = crisv10_hcd_start, + + /* cleanly make HCD stop writing memory and doing I/O */ + .stop = crisv10_hcd_stop, + + /* return current frame number */ + .get_frame_number = crisv10_hcd_get_frame, + + + /* Manage i/o requests via the Transfer Controller */ + .urb_enqueue = tc_urb_enqueue, + .urb_dequeue = tc_urb_dequeue, + + /* hw synch, freeing endpoint resources that urb_dequeue can't */ + .endpoint_disable = tc_endpoint_disable, + + + /* Root Hub support */ + .hub_status_data = rh_status_data_request, + .hub_control = rh_control_request, +#ifdef CONFIG_PM + .hub_suspend = rh_suspend_request, + .hub_resume = rh_resume_request, +#endif /* CONFIG_PM */ +#ifdef CONFIG_USB_OTG + .start_port_reset = crisv10_hcd_start_port_reset, +#endif /* CONFIG_USB_OTG */ + }; + + +/* + * conversion between pointers to a hcd and the corresponding + * crisv10_hcd + */ + +static inline struct crisv10_hcd *hcd_to_crisv10_hcd(struct usb_hcd *hcd) +{ + return (struct crisv10_hcd *) hcd->hcd_priv; +} + +static inline struct usb_hcd *crisv10_hcd_to_hcd(struct crisv10_hcd *hcd) +{ + return container_of((void *) hcd, struct usb_hcd, hcd_priv); +} + +/* check if specified port is in use */ +static inline int port_in_use(unsigned int port) +{ + return ports & (1 << port); +} + +/* number of ports in use */ +static inline unsigned int num_ports(void) +{ + unsigned int i, num = 0; + for (i = 0; i < USB_ROOT_HUB_PORTS; i++) + if (port_in_use(i)) + num++; + return num; +} + +/* map hub port number to the port number used internally by the HC */ +static inline unsigned int map_port(unsigned int port) +{ + unsigned int i, num = 0; + for (i = 0; i < USB_ROOT_HUB_PORTS; i++) + if (port_in_use(i)) + if (++num == port) + return i; + return -1; +} + +/* size of descriptors in slab cache */ +#ifndef MAX +#define MAX(x, y) ((x) > (y) ? (x) : (y)) +#endif + + +/******************************************************************/ +/* Hardware Interrupt functions */ +/******************************************************************/ + +/* Fast interrupt handler for HC */ +static irqreturn_t crisv10_hcd_top_irq(int irq, void *vcd) +{ + struct usb_hcd *hcd = vcd; + struct crisv10_irq_reg reg; + __u32 irq_mask; + unsigned long flags; + + DBFENTER; + + ASSERT(hcd != NULL); + reg.hcd = hcd; + + /* Turn of other interrupts while handling these sensitive cases */ + local_irq_save(flags); + + /* Read out which interrupts that are flaged */ + irq_mask = *R_USB_IRQ_MASK_READ; + reg.r_usb_irq_mask_read = irq_mask; + + /* Reading R_USB_STATUS clears the ctl_status interrupt. Note that + R_USB_STATUS must be read before R_USB_EPID_ATTN since reading the latter + clears the ourun and perror fields of R_USB_STATUS. */ + reg.r_usb_status = *R_USB_STATUS; + + /* Reading R_USB_EPID_ATTN clears the iso_eof, bulk_eot and epid_attn + interrupts. */ + reg.r_usb_epid_attn = *R_USB_EPID_ATTN; + + /* Reading R_USB_RH_PORT_STATUS_1 and R_USB_RH_PORT_STATUS_2 clears the + port_status interrupt. */ + reg.r_usb_rh_port_status_1 = *R_USB_RH_PORT_STATUS_1; + reg.r_usb_rh_port_status_2 = *R_USB_RH_PORT_STATUS_2; + + /* Reading R_USB_FM_NUMBER clears the sof interrupt. */ + /* Note: the lower 11 bits contain the actual frame number, sent with each + sof. */ + reg.r_usb_fm_number = *R_USB_FM_NUMBER; + + /* Interrupts are handled in order of priority. */ + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, port_status)) { + crisv10_hcd_port_status_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, epid_attn)) { + crisv10_hcd_epid_attn_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, ctl_status)) { + crisv10_hcd_ctl_status_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, iso_eof)) { + crisv10_hcd_isoc_eof_irq(®); + } + if (irq_mask & IO_MASK(R_USB_IRQ_MASK_READ, bulk_eot)) { + /* Update/restart the bulk start timer since obviously the channel is + running. */ + mod_timer(&bulk_start_timer, jiffies + BULK_START_TIMER_INTERVAL); + /* Update/restart the bulk eot timer since we just received an bulk eot + interrupt. */ + mod_timer(&bulk_eot_timer, jiffies + BULK_EOT_TIMER_INTERVAL); + + /* Check for finished bulk transfers on epids */ + check_finished_bulk_tx_epids(hcd, 0); + } + local_irq_restore(flags); + + DBFEXIT; + return IRQ_HANDLED; +} + + +void crisv10_hcd_epid_attn_irq(struct crisv10_irq_reg *reg) { + struct usb_hcd *hcd = reg->hcd; + struct crisv10_urb_priv *urb_priv; + int epid; + DBFENTER; + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if (test_bit(epid, (void *)®->r_usb_epid_attn)) { + struct urb *urb; + __u32 ept_data; + int error_code; + + if (epid == DUMMY_EPID || epid == INVALID_EPID) { + /* We definitely don't care about these ones. Besides, they are + always disabled, so any possible disabling caused by the + epid attention interrupt is irrelevant. */ + continue; + } + + if(!epid_inuse(epid)) { + irq_err("Epid attention on epid:%d that isn't in use\n", epid); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + debug_epid(epid); + continue; + } + + /* Note that although there are separate R_USB_EPT_DATA and + R_USB_EPT_DATA_ISO registers, they are located at the same address and + are of the same size. In other words, this read should be ok for isoc + also. */ + ept_data = etrax_epid_get(epid); + error_code = IO_EXTRACT(R_USB_EPT_DATA, error_code, ept_data); + + /* Get the active URB for this epid. We blatantly assume + that only this URB could have caused the epid attention. */ + urb = activeUrbList[epid]; + if (urb == NULL) { + irq_err("Attention on epid:%d error:%d with no active URB.\n", + epid, error_code); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + debug_epid(epid); + continue; + } + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + /* Using IO_STATE_VALUE on R_USB_EPT_DATA should be ok for isoc also. */ + if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + + /* Isoc traffic doesn't have error_count_in/error_count_out. */ + if ((usb_pipetype(urb->pipe) != PIPE_ISOCHRONOUS) && + (IO_EXTRACT(R_USB_EPT_DATA, error_count_in, ept_data) == 3 || + IO_EXTRACT(R_USB_EPT_DATA, error_count_out, ept_data) == 3)) { + /* Check if URB allready is marked for late-finish, we can get + several 3rd error for Intr traffic when a device is unplugged */ + if(urb_priv->later_data == NULL) { + /* 3rd error. */ + irq_warn("3rd error for epid:%d (%s %s) URB:0x%x[%d]\n", epid, + str_dir(urb->pipe), str_type(urb->pipe), + (unsigned int)urb, urb_priv->urb_num); + + tc_finish_urb_later(hcd, urb, -EPROTO); + } + + } else if (reg->r_usb_status & IO_MASK(R_USB_STATUS, perror)) { + irq_warn("Perror for epid:%d\n", epid); + printk("FM_NUMBER: %d\n", reg->r_usb_fm_number & 0x7ff); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + __dump_urb(urb); + debug_epid(epid); + + if (!(ept_data & IO_MASK(R_USB_EPT_DATA, valid))) { + /* invalid ep_id */ + panic("Perror because of invalid epid." + " Deconfigured too early?"); + } else { + /* past eof1, near eof, zout transfer, setup transfer */ + /* Dump the urb and the relevant EP descriptor. */ + panic("Something wrong with DMA descriptor contents." + " Too much traffic inserted?"); + } + } else if (reg->r_usb_status & IO_MASK(R_USB_STATUS, ourun)) { + /* buffer ourun */ + printk("FM_NUMBER: %d\n", reg->r_usb_fm_number & 0x7ff); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + __dump_urb(urb); + debug_epid(epid); + + panic("Buffer overrun/underrun for epid:%d. DMA too busy?", epid); + } else { + irq_warn("Attention on epid:%d (%s %s) with no error code\n", epid, + str_dir(urb->pipe), str_type(urb->pipe)); + printk("R_USB_STATUS: 0x%x\n", reg->r_usb_status); + __dump_urb(urb); + debug_epid(epid); + } + + } else if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, + stall)) { + /* Not really a protocol error, just says that the endpoint gave + a stall response. Note that error_code cannot be stall for isoc. */ + if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) { + panic("Isoc traffic cannot stall"); + } + + tc_dbg("Stall for epid:%d (%s %s) URB:0x%x\n", epid, + str_dir(urb->pipe), str_type(urb->pipe), (unsigned int)urb); + tc_finish_urb(hcd, urb, -EPIPE); + + } else if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, + bus_error)) { + /* Two devices responded to a transaction request. Must be resolved + by software. FIXME: Reset ports? */ + panic("Bus error for epid %d." + " Two devices responded to transaction request\n", + epid); + + } else if (error_code == IO_STATE_VALUE(R_USB_EPT_DATA, error_code, + buffer_error)) { + /* DMA overrun or underrun. */ + irq_warn("Buffer overrun/underrun for epid:%d (%s %s)\n", epid, + str_dir(urb->pipe), str_type(urb->pipe)); + + /* It seems that error_code = buffer_error in + R_USB_EPT_DATA/R_USB_EPT_DATA_ISO and ourun = yes in R_USB_STATUS + are the same error. */ + tc_finish_urb(hcd, urb, -EPROTO); + } else { + irq_warn("Unknown attention on epid:%d (%s %s)\n", epid, + str_dir(urb->pipe), str_type(urb->pipe)); + dump_ept_data(epid); + } + } + } + DBFEXIT; +} + +void crisv10_hcd_port_status_irq(struct crisv10_irq_reg *reg) +{ + __u16 port_reg[USB_ROOT_HUB_PORTS]; + DBFENTER; + port_reg[0] = reg->r_usb_rh_port_status_1; + port_reg[1] = reg->r_usb_rh_port_status_2; + rh_port_status_change(port_reg); + DBFEXIT; +} + +void crisv10_hcd_isoc_eof_irq(struct crisv10_irq_reg *reg) +{ + int epid; + struct urb *urb; + struct crisv10_urb_priv *urb_priv; + + DBFENTER; + + for (epid = 0; epid < NBR_OF_EPIDS - 1; epid++) { + + /* Only check epids that are in use, is valid and has SB list */ + if (!epid_inuse(epid) || epid == INVALID_EPID || + TxIsocEPList[epid].sub == 0 || epid == DUMMY_EPID) { + /* Nothing here to see. */ + continue; + } + ASSERT(epid_isoc(epid)); + + /* Get the active URB for this epid (if any). */ + urb = activeUrbList[epid]; + if (urb == 0) { + isoc_warn("Ignoring NULL urb for epid:%d\n", epid); + continue; + } + if(!epid_out_traffic(epid)) { + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + if (urb_priv->urb_state == NOT_STARTED) { + /* If ASAP is not set and urb->start_frame is the current frame, + start the transfer. */ + if (!(urb->transfer_flags & URB_ISO_ASAP) && + (urb->start_frame == (*R_USB_FM_NUMBER & 0x7ff))) { + /* EP should not be enabled if we're waiting for start_frame */ + ASSERT((TxIsocEPList[epid].command & + IO_STATE(USB_EP_command, enable, yes)) == 0); + + isoc_warn("Enabling isoc IN EP descr for epid %d\n", epid); + TxIsocEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + + /* This urb is now active. */ + urb_priv->urb_state = STARTED; + continue; + } + } + } + } + + DBFEXIT; +} + +void crisv10_hcd_ctl_status_irq(struct crisv10_irq_reg *reg) +{ + struct crisv10_hcd* crisv10_hcd = hcd_to_crisv10_hcd(reg->hcd); + + DBFENTER; + ASSERT(crisv10_hcd); + +/* irq_dbg("ctr_status_irq, controller status: %s\n", + hcd_status_to_str(reg->r_usb_status));*/ + + /* FIXME: What should we do if we get ourun or perror? Dump the EP and SB + list for the corresponding epid? */ + if (reg->r_usb_status & IO_MASK(R_USB_STATUS, ourun)) { + panic("USB controller got ourun."); + } + if (reg->r_usb_status & IO_MASK(R_USB_STATUS, perror)) { + + /* Before, etrax_usb_do_intr_recover was called on this epid if it was + an interrupt pipe. I don't see how re-enabling all EP descriptors + will help if there was a programming error. */ + panic("USB controller got perror."); + } + + /* Keep track of USB Controller, if it's running or not */ + if(reg->r_usb_status & IO_STATE(R_USB_STATUS, running, yes)) { + crisv10_hcd->running = 1; + } else { + crisv10_hcd->running = 0; + } + + if (reg->r_usb_status & IO_MASK(R_USB_STATUS, device_mode)) { + /* We should never operate in device mode. */ + panic("USB controller in device mode."); + } + + /* Set the flag to avoid getting "Unlink after no-IRQ? Controller is probably + using the wrong IRQ" from hcd_unlink_urb() in drivers/usb/core/hcd.c */ + set_bit(HCD_FLAG_SAW_IRQ, ®->hcd->flags); + + DBFEXIT; +} + + +/******************************************************************/ +/* Host Controller interface functions */ +/******************************************************************/ + +static inline void crisv10_ready_wait(void) { + volatile int timeout = 10000; + /* Check the busy bit of USB controller in Etrax */ + while((*R_USB_COMMAND & IO_MASK(R_USB_COMMAND, busy)) && + (timeout-- > 0)); +} + +/* reset host controller */ +static int crisv10_hcd_reset(struct usb_hcd *hcd) +{ + DBFENTER; + hcd_dbg(hcd, "reset\n"); + + + /* Reset the USB interface. */ + /* + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, reset); + nop(); + */ + DBFEXIT; + return 0; +} + +/* start host controller */ +static int crisv10_hcd_start(struct usb_hcd *hcd) +{ + DBFENTER; + hcd_dbg(hcd, "start\n"); + + crisv10_ready_wait(); + + /* Start processing of USB traffic. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + + nop(); + + hcd->state = HC_STATE_RUNNING; + + DBFEXIT; + return 0; +} + +/* stop host controller */ +static void crisv10_hcd_stop(struct usb_hcd *hcd) +{ + DBFENTER; + hcd_dbg(hcd, "stop\n"); + crisv10_hcd_reset(hcd); + DBFEXIT; +} + +/* return the current frame number */ +static int crisv10_hcd_get_frame(struct usb_hcd *hcd) +{ + DBFENTER; + DBFEXIT; + return (*R_USB_FM_NUMBER & 0x7ff); +} + +#ifdef CONFIG_USB_OTG + +static int crisv10_hcd_start_port_reset(struct usb_hcd *hcd, unsigned port) +{ + return 0; /* no-op for now */ +} + +#endif /* CONFIG_USB_OTG */ + + +/******************************************************************/ +/* Root Hub functions */ +/******************************************************************/ + +/* root hub status */ +static const struct usb_hub_status rh_hub_status = + { + .wHubStatus = 0, + .wHubChange = 0, + }; + +/* root hub descriptor */ +static const u8 rh_hub_descr[] = + { + 0x09, /* bDescLength */ + 0x29, /* bDescriptorType */ + USB_ROOT_HUB_PORTS, /* bNbrPorts */ + 0x00, /* wHubCharacteristics */ + 0x00, + 0x01, /* bPwrOn2pwrGood */ + 0x00, /* bHubContrCurrent */ + 0x00, /* DeviceRemovable */ + 0xff /* PortPwrCtrlMask */ + }; + +/* Actual holder of root hub status*/ +struct crisv10_rh rh; + +/* Initialize root hub data structures (called from dvdrv_hcd_probe()) */ +int rh_init(void) { + int i; + /* Reset port status flags */ + for (i = 0; i < USB_ROOT_HUB_PORTS; i++) { + rh.wPortChange[i] = 0; + rh.wPortStatusPrev[i] = 0; + } + return 0; +} + +#define RH_FEAT_MASK ((1<lock); + for (i = 1; i <= crisv10_hcd->num_ports; i++) { + if (rh.wPortChange[map_port(i)]) { + *buf |= (1 << i); + rh_dbg("rh_status_data_request, change on port %d: %s Current Status: %s\n", i, + port_status_to_str(rh.wPortChange[map_port(i)]), + port_status_to_str(rh.wPortStatusPrev[map_port(i)])); + } + } + spin_unlock(&crisv10_hcd->lock); + DBFEXIT; + return *buf == 0 ? 0 : 1; +} + +/* Handle a control request for the root hub (called from hcd_driver) */ +static int rh_control_request(struct usb_hcd *hcd, + u16 typeReq, + u16 wValue, + u16 wIndex, + char *buf, + u16 wLength) { + + struct crisv10_hcd *crisv10_hcd = hcd_to_crisv10_hcd(hcd); + int retval = 0; + int len; + DBFENTER; + + switch (typeReq) { + case GetHubDescriptor: + rh_dbg("GetHubDescriptor\n"); + len = min_t(unsigned int, sizeof rh_hub_descr, wLength); + memcpy(buf, rh_hub_descr, len); + buf[2] = crisv10_hcd->num_ports; + break; + case GetHubStatus: + rh_dbg("GetHubStatus\n"); + len = min_t(unsigned int, sizeof rh_hub_status, wLength); + memcpy(buf, &rh_hub_status, len); + break; + case GetPortStatus: + if (!wIndex || wIndex > crisv10_hcd->num_ports) + goto error; + rh_dbg("GetportStatus, port:%d change:%s status:%s\n", wIndex, + port_status_to_str(rh.wPortChange[map_port(wIndex)]), + port_status_to_str(rh.wPortStatusPrev[map_port(wIndex)])); + *(u16 *) buf = cpu_to_le16(rh.wPortStatusPrev[map_port(wIndex)]); + *(u16 *) (buf + 2) = cpu_to_le16(rh.wPortChange[map_port(wIndex)]); + break; + case SetHubFeature: + rh_dbg("SetHubFeature\n"); + case ClearHubFeature: + rh_dbg("ClearHubFeature\n"); + switch (wValue) { + case C_HUB_OVER_CURRENT: + case C_HUB_LOCAL_POWER: + rh_warn("Not implemented hub request:%d \n", typeReq); + /* not implemented */ + break; + default: + goto error; + } + break; + case SetPortFeature: + if (!wIndex || wIndex > crisv10_hcd->num_ports) + goto error; + if(rh_set_port_feature(map_port(wIndex), wValue)) + goto error; + break; + case ClearPortFeature: + if (!wIndex || wIndex > crisv10_hcd->num_ports) + goto error; + if(rh_clear_port_feature(map_port(wIndex), wValue)) + goto error; + break; + default: + rh_warn("Unknown hub request: %d\n", typeReq); + error: + retval = -EPIPE; + } + DBFEXIT; + return retval; +} + +int rh_set_port_feature(__u8 bPort, __u16 wFeature) { + __u8 bUsbCommand = 0; + __u8 reset_cnt; + switch(wFeature) { + case USB_PORT_FEAT_RESET: + rh_dbg("SetPortFeature: reset\n"); + + if (rh.wPortStatusPrev[bPort] & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) + { + __u8 restart_controller = 0; + + if ( (rh.wPortStatusPrev[0] & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) && + (rh.wPortStatusPrev[1] & + IO_STATE(R_USB_RH_PORT_STATUS_2, enabled, yes)) ) + { + /* Both ports is enabled. The USB controller will not change state. */ + restart_controller = 0; + } + else + { + /* Only ports is enabled. The USB controller will change state and + must be restarted. */ + restart_controller = 1; + } + /* + In ETRAX 100LX it's not possible to reset an enabled root hub port. + The workaround is to disable and enable the port before resetting it. + Disabling the port can, if both ports are disabled at once, cause the + USB controller to change state to HOST_MODE state. + The USB controller state transition causes a lot of unwanted + interrupts that must be avoided. + Disabling the USB controller status and port status interrupts before + disabling/resetting the port stops these interrupts. + + These actions are performed: + 1. Disable USB controller status and port status interrupts. + 2. Disable the port + 3. Wait for the port to be disabled. + 4. Enable the port. + 5. Wait for the port to be enabled. + 6. Reset the port. + 7. Wait for for the reset to end. + 8. Wait for the USB controller entering started state. + 9. Order the USB controller to running state. + 10. Wait for the USB controller reaching running state. + 11. Clear all interrupts generated during the disable/enable/reset + procedure. + 12. Enable the USB controller status and port status interrupts. + */ + + /* 1. Disable USB controller status and USB port status interrupts. */ + *R_USB_IRQ_MASK_CLR = IO_STATE(R_USB_IRQ_MASK_CLR, ctl_status, clr); + __asm__ __volatile__ (" nop"); + *R_USB_IRQ_MASK_CLR = IO_STATE(R_USB_IRQ_MASK_CLR, port_status, clr); + __asm__ __volatile__ (" nop"); + + { + + /* Since an root hub port reset shall be 50 ms and the ETRAX 100LX + root hub port reset is 10 ms we must perform 5 port resets to + achieve a proper root hub port reset. */ + for (reset_cnt = 0; reset_cnt < 5; reset_cnt ++) + { + rh_dbg("Disable Port %d\n", bPort + 1); + + /* 2. Disable the port*/ + if (bPort == 0) + { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, yes); + } + else + { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, yes); + } + + /* 3. Wait for the port to be disabled. */ + while ( (bPort == 0) ? + *R_USB_RH_PORT_STATUS_1 & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes) : + *R_USB_RH_PORT_STATUS_2 & + IO_STATE(R_USB_RH_PORT_STATUS_2, enabled, yes) ) {} + + rh_dbg("Port %d is disabled. Enable it!\n", bPort + 1); + + /* 4. Enable the port. */ + if (bPort == 0) + { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, no); + } + else + { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, no); + } + + /* 5. Wait for the port to be enabled again. */ + while (!( (bPort == 0) ? + *R_USB_RH_PORT_STATUS_1 & + IO_STATE(R_USB_RH_PORT_STATUS_1, connected, yes) : + *R_USB_RH_PORT_STATUS_2 & + IO_STATE(R_USB_RH_PORT_STATUS_2, connected, yes) ) ) {} + + rh_dbg("Port %d is enabled.\n", bPort + 1); + + /* 6. Reset the port */ + crisv10_ready_wait(); + *R_USB_COMMAND = + ( (bPort == 0) ? + IO_STATE(R_USB_COMMAND, port_sel, port1): + IO_STATE(R_USB_COMMAND, port_sel, port2) ) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, busy, no) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, nop); + rh_dbg("Port %d is resetting.\n", bPort + 1); + + /* 7. The USB specification says that we should wait for at least + 10ms for device recover */ + udelay(10500); /* 10,5ms blocking wait */ + + crisv10_ready_wait(); + } + } + + + /* Check if the USB controller needs to be restarted. */ + if (restart_controller) + { + /* 8. Wait for the USB controller entering started state. */ + while (!(*R_USB_STATUS & IO_STATE(R_USB_STATUS, started, yes))) {} + + /* 9. Order the USB controller to running state. */ + crisv10_ready_wait(); + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, busy, no) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + + /* 10. Wait for the USB controller reaching running state. */ + while (!(*R_USB_STATUS & IO_STATE(R_USB_STATUS, running, yes))) {} + } + + /* 11. Clear any controller or port satus interrupts before enabling + the interrupts. */ + { + u16 dummy; + + /* Clear the port status interrupt of the reset port. */ + if (bPort == 0) + { + rh_dbg("Clearing port 1 interrupts\n"); + dummy = *R_USB_RH_PORT_STATUS_1; + } + else + { + rh_dbg("Clearing port 2 interrupts\n"); + dummy = *R_USB_RH_PORT_STATUS_2; + } + + if (restart_controller) + { + /* The USB controller is restarted. Clear all interupts. */ + rh_dbg("Clearing all interrupts\n"); + dummy = *R_USB_STATUS; + dummy = *R_USB_RH_PORT_STATUS_1; + dummy = *R_USB_RH_PORT_STATUS_2; + } + } + + /* 12. Enable USB controller status and USB port status interrupts. */ + *R_USB_IRQ_MASK_SET = IO_STATE(R_USB_IRQ_MASK_SET, ctl_status, set); + __asm__ __volatile__ (" nop"); + *R_USB_IRQ_MASK_SET = IO_STATE(R_USB_IRQ_MASK_SET, port_status, set); + __asm__ __volatile__ (" nop"); + + } + else + { + + bUsbCommand |= IO_STATE(R_USB_COMMAND, port_cmd, reset); + /* Select which port via the port_sel field */ + bUsbCommand |= IO_FIELD(R_USB_COMMAND, port_sel, bPort+1); + + /* Make sure the controller isn't busy. */ + crisv10_ready_wait(); + /* Send out the actual command to the USB controller */ + *R_USB_COMMAND = bUsbCommand; + + /* Wait a while for controller to first become started after port reset */ + udelay(12000); /* 12ms blocking wait */ + + /* Make sure the controller isn't busy. */ + crisv10_ready_wait(); + + /* If all enabled ports were disabled the host controller goes down into + started mode, so we need to bring it back into the running state. + (This is safe even if it's already in the running state.) */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + } + + break; + case USB_PORT_FEAT_SUSPEND: + rh_dbg("SetPortFeature: suspend\n"); + bUsbCommand |= IO_STATE(R_USB_COMMAND, port_cmd, suspend); + goto set; + break; + case USB_PORT_FEAT_POWER: + rh_dbg("SetPortFeature: power\n"); + break; + case USB_PORT_FEAT_C_CONNECTION: + rh_dbg("SetPortFeature: c_connection\n"); + break; + case USB_PORT_FEAT_C_RESET: + rh_dbg("SetPortFeature: c_reset\n"); + break; + case USB_PORT_FEAT_C_OVER_CURRENT: + rh_dbg("SetPortFeature: c_over_current\n"); + break; + + set: + /* Select which port via the port_sel field */ + bUsbCommand |= IO_FIELD(R_USB_COMMAND, port_sel, bPort+1); + + /* Make sure the controller isn't busy. */ + crisv10_ready_wait(); + /* Send out the actual command to the USB controller */ + *R_USB_COMMAND = bUsbCommand; + break; + default: + rh_dbg("SetPortFeature: unknown feature\n"); + return -1; + } + return 0; +} + +int rh_clear_port_feature(__u8 bPort, __u16 wFeature) { + switch(wFeature) { + case USB_PORT_FEAT_ENABLE: + rh_dbg("ClearPortFeature: enable\n"); + rh_disable_port(bPort); + break; + case USB_PORT_FEAT_SUSPEND: + rh_dbg("ClearPortFeature: suspend\n"); + break; + case USB_PORT_FEAT_POWER: + rh_dbg("ClearPortFeature: power\n"); + break; + + case USB_PORT_FEAT_C_ENABLE: + rh_dbg("ClearPortFeature: c_enable\n"); + goto clear; + case USB_PORT_FEAT_C_SUSPEND: + rh_dbg("ClearPortFeature: c_suspend\n"); + goto clear; + case USB_PORT_FEAT_C_CONNECTION: + rh_dbg("ClearPortFeature: c_connection\n"); + goto clear; + case USB_PORT_FEAT_C_OVER_CURRENT: + rh_dbg("ClearPortFeature: c_over_current\n"); + goto clear; + case USB_PORT_FEAT_C_RESET: + rh_dbg("ClearPortFeature: c_reset\n"); + goto clear; + clear: + rh.wPortChange[bPort] &= ~(1 << (wFeature - 16)); + break; + default: + rh_dbg("ClearPortFeature: unknown feature\n"); + return -1; + } + return 0; +} + + +#ifdef CONFIG_PM +/* Handle a suspend request for the root hub (called from hcd_driver) */ +static int rh_suspend_request(struct usb_hcd *hcd) +{ + return 0; /* no-op for now */ +} + +/* Handle a resume request for the root hub (called from hcd_driver) */ +static int rh_resume_request(struct usb_hcd *hcd) +{ + return 0; /* no-op for now */ +} +#endif /* CONFIG_PM */ + + + +/* Wrapper function for workaround port disable registers in USB controller */ +static void rh_disable_port(unsigned int port) { + volatile int timeout = 10000; + volatile char* usb_portx_disable; + switch(port) { + case 0: + usb_portx_disable = R_USB_PORT1_DISABLE; + break; + case 1: + usb_portx_disable = R_USB_PORT2_DISABLE; + break; + default: + /* Invalid port index */ + return; + } + /* Set disable flag in special register */ + *usb_portx_disable = IO_STATE(R_USB_PORT1_DISABLE, disable, yes); + /* Wait until not enabled anymore */ + while((rh.wPortStatusPrev[port] & + IO_STATE(R_USB_RH_PORT_STATUS_1, enabled, yes)) && + (timeout-- > 0)); + + /* clear disable flag in special register */ + *usb_portx_disable = IO_STATE(R_USB_PORT1_DISABLE, disable, no); + rh_info("Physical port %d disabled\n", port+1); +} + + +/******************************************************************/ +/* Transfer Controller (TC) functions */ +/******************************************************************/ + +/* FIXME: Should RX_BUF_SIZE be a config option, or maybe we should adjust it + dynamically? + To adjust it dynamically we would have to get an interrupt when we reach + the end of the rx descriptor list, or when we get close to the end, and + then allocate more descriptors. */ +#define NBR_OF_RX_DESC 512 +#define RX_DESC_BUF_SIZE 1024 +#define RX_BUF_SIZE (NBR_OF_RX_DESC * RX_DESC_BUF_SIZE) + + +/* Local variables for Transfer Controller */ +/* --------------------------------------- */ + +/* This is a circular (double-linked) list of the active urbs for each epid. + The head is never removed, and new urbs are linked onto the list as + urb_entry_t elements. Don't reference urb_list directly; use the wrapper + functions instead (which includes spin_locks) */ +static struct list_head urb_list[NBR_OF_EPIDS]; + +/* Read about the need and usage of this lock in submit_ctrl_urb. */ +/* Lock for URB lists for each EPID */ +static spinlock_t urb_list_lock; + +/* Lock for EPID array register (R_USB_EPT_x) in Etrax */ +static spinlock_t etrax_epid_lock; + +/* Lock for dma8 sub0 handling */ +static spinlock_t etrax_dma8_sub0_lock; + +/* DMA IN cache bug. Align the DMA IN buffers to 32 bytes, i.e. a cache line. + Since RX_DESC_BUF_SIZE is 1024 is a multiple of 32, all rx buffers will be + cache aligned. */ +static volatile unsigned char RxBuf[RX_BUF_SIZE] __attribute__ ((aligned (32))); +static volatile struct USB_IN_Desc RxDescList[NBR_OF_RX_DESC] __attribute__ ((aligned (4))); + +/* Pointers into RxDescList. */ +static volatile struct USB_IN_Desc *myNextRxDesc; +static volatile struct USB_IN_Desc *myLastRxDesc; + +/* A zout transfer makes a memory access at the address of its buf pointer, + which means that setting this buf pointer to 0 will cause an access to the + flash. In addition to this, setting sw_len to 0 results in a 16/32 bytes + (depending on DMA burst size) transfer. + Instead, we set it to 1, and point it to this buffer. */ +static int zout_buffer[4] __attribute__ ((aligned (4))); + +/* Cache for allocating new EP and SB descriptors. */ +static struct kmem_cache *usb_desc_cache; + +/* Cache for the data allocated in the isoc descr top half. */ +static struct kmem_cache *isoc_compl_cache; + +/* Cache for the data allocated when delayed finishing of URBs */ +static struct kmem_cache *later_data_cache; + + +/* Counter to keep track of how many Isoc EP we have sat up. Used to enable + and disable iso_eof interrupt. We only need these interrupts when we have + Isoc data endpoints (consumes CPU cycles). + FIXME: This could be more fine granular, so this interrupt is only enabled + when we have a In Isoc URB not URB_ISO_ASAP flaged queued. */ +static int isoc_epid_counter; + +/* Protecting wrapper functions for R_USB_EPT_x */ +/* -------------------------------------------- */ +static inline void etrax_epid_set(__u8 index, __u32 data) { + unsigned long flags; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + *R_USB_EPT_DATA = data; + spin_unlock_irqrestore(&etrax_epid_lock, flags); +} + +static inline void etrax_epid_clear_error(__u8 index) { + unsigned long flags; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + *R_USB_EPT_DATA &= + ~(IO_MASK(R_USB_EPT_DATA, error_count_in) | + IO_MASK(R_USB_EPT_DATA, error_count_out) | + IO_MASK(R_USB_EPT_DATA, error_code)); + spin_unlock_irqrestore(&etrax_epid_lock, flags); +} + +static inline void etrax_epid_set_toggle(__u8 index, __u8 dirout, + __u8 toggle) { + unsigned long flags; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + if(dirout) { + *R_USB_EPT_DATA &= ~IO_MASK(R_USB_EPT_DATA, t_out); + *R_USB_EPT_DATA |= IO_FIELD(R_USB_EPT_DATA, t_out, toggle); + } else { + *R_USB_EPT_DATA &= ~IO_MASK(R_USB_EPT_DATA, t_in); + *R_USB_EPT_DATA |= IO_FIELD(R_USB_EPT_DATA, t_in, toggle); + } + spin_unlock_irqrestore(&etrax_epid_lock, flags); +} + +static inline __u8 etrax_epid_get_toggle(__u8 index, __u8 dirout) { + unsigned long flags; + __u8 toggle; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + if (dirout) { + toggle = IO_EXTRACT(R_USB_EPT_DATA, t_out, *R_USB_EPT_DATA); + } else { + toggle = IO_EXTRACT(R_USB_EPT_DATA, t_in, *R_USB_EPT_DATA); + } + spin_unlock_irqrestore(&etrax_epid_lock, flags); + return toggle; +} + + +static inline __u32 etrax_epid_get(__u8 index) { + unsigned long flags; + __u32 data; + spin_lock_irqsave(&etrax_epid_lock, flags); + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, index); + nop(); + data = *R_USB_EPT_DATA; + spin_unlock_irqrestore(&etrax_epid_lock, flags); + return data; +} + + + + +/* Main functions for Transfer Controller */ +/* -------------------------------------- */ + +/* Init structs, memories and lists used by Transfer Controller */ +int tc_init(struct usb_hcd *hcd) { + int i; + /* Clear software state info for all epids */ + memset(epid_state, 0, sizeof(struct etrax_epid) * NBR_OF_EPIDS); + + /* Set Invalid and Dummy as being in use and disabled */ + epid_state[INVALID_EPID].inuse = 1; + epid_state[DUMMY_EPID].inuse = 1; + epid_state[INVALID_EPID].disabled = 1; + epid_state[DUMMY_EPID].disabled = 1; + + /* Clear counter for how many Isoc epids we have sat up */ + isoc_epid_counter = 0; + + /* Initialize the urb list by initiating a head for each list. + Also reset list hodling active URB for each epid */ + for (i = 0; i < NBR_OF_EPIDS; i++) { + INIT_LIST_HEAD(&urb_list[i]); + activeUrbList[i] = NULL; + } + + /* Init lock for URB lists */ + spin_lock_init(&urb_list_lock); + /* Init lock for Etrax R_USB_EPT register */ + spin_lock_init(&etrax_epid_lock); + /* Init lock for Etrax dma8 sub0 handling */ + spin_lock_init(&etrax_dma8_sub0_lock); + + /* We use kmem_cache_* to make sure that all DMA desc. are dword aligned */ + + /* Note that we specify sizeof(struct USB_EP_Desc) as the size, but also + allocate SB descriptors from this cache. This is ok since + sizeof(struct USB_EP_Desc) == sizeof(struct USB_SB_Desc). */ + usb_desc_cache = kmem_cache_create("usb_desc_cache", + sizeof(struct USB_EP_Desc), 0, + SLAB_HWCACHE_ALIGN, 0); + if(usb_desc_cache == NULL) { + return -ENOMEM; + } + + /* Create slab cache for speedy allocation of memory for isoc bottom-half + interrupt handling */ + isoc_compl_cache = + kmem_cache_create("isoc_compl_cache", + sizeof(struct crisv10_isoc_complete_data), + 0, SLAB_HWCACHE_ALIGN, 0); + if(isoc_compl_cache == NULL) { + return -ENOMEM; + } + + /* Create slab cache for speedy allocation of memory for later URB finish + struct */ + later_data_cache = + kmem_cache_create("later_data_cache", + sizeof(struct urb_later_data), + 0, SLAB_HWCACHE_ALIGN, 0); + if(later_data_cache == NULL) { + return -ENOMEM; + } + + + /* Initiate the bulk start timer. */ + init_timer(&bulk_start_timer); + bulk_start_timer.expires = jiffies + BULK_START_TIMER_INTERVAL; + bulk_start_timer.function = tc_bulk_start_timer_func; + add_timer(&bulk_start_timer); + + + /* Initiate the bulk eot timer. */ + init_timer(&bulk_eot_timer); + bulk_eot_timer.expires = jiffies + BULK_EOT_TIMER_INTERVAL; + bulk_eot_timer.function = tc_bulk_eot_timer_func; + bulk_eot_timer.data = (unsigned long)hcd; + add_timer(&bulk_eot_timer); + + return 0; +} + +/* Uninitialize all resources used by Transfer Controller */ +void tc_destroy(void) { + + /* Destroy all slab cache */ + kmem_cache_destroy(usb_desc_cache); + kmem_cache_destroy(isoc_compl_cache); + kmem_cache_destroy(later_data_cache); + + /* Remove timers */ + del_timer(&bulk_start_timer); + del_timer(&bulk_eot_timer); +} + +static void restart_dma8_sub0(void) { + unsigned long flags; + spin_lock_irqsave(&etrax_dma8_sub0_lock, flags); + /* Verify that the dma is not running */ + if ((*R_DMA_CH8_SUB0_CMD & IO_MASK(R_DMA_CH8_SUB0_CMD, cmd)) == 0) { + struct USB_EP_Desc *ep = (struct USB_EP_Desc *)phys_to_virt(*R_DMA_CH8_SUB0_EP); + while (DUMMY_EPID == IO_EXTRACT(USB_EP_command, epid, ep->command)) { + ep = (struct USB_EP_Desc *)phys_to_virt(ep->next); + } + /* Advance the DMA to the next EP descriptor that is not a DUMMY_EPID. */ + *R_DMA_CH8_SUB0_EP = virt_to_phys(ep); + /* Restart the DMA */ + *R_DMA_CH8_SUB0_CMD = IO_STATE(R_DMA_CH8_SUB0_CMD, cmd, start); + } + spin_unlock_irqrestore(&etrax_dma8_sub0_lock, flags); +} + +/* queue an URB with the transfer controller (called from hcd_driver) */ +static int tc_urb_enqueue(struct usb_hcd *hcd, + struct urb *urb, + gfp_t mem_flags) { + int epid; + int retval; + int bustime = 0; + int maxpacket; + unsigned long flags; + struct crisv10_urb_priv *urb_priv; + struct crisv10_hcd* crisv10_hcd = hcd_to_crisv10_hcd(hcd); + DBFENTER; + + if(!(crisv10_hcd->running)) { + /* The USB Controller is not running, probably because no device is + attached. No idea to enqueue URBs then */ + tc_warn("Rejected enqueueing of URB:0x%x because no dev attached\n", + (unsigned int)urb); + return -ENOENT; + } + + maxpacket = usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe)); + /* Special case check for In Isoc transfers. Specification states that each + In Isoc transfer consists of one packet and therefore it should fit into + the transfer-buffer of an URB. + We do the check here to be sure (an invalid scenario can be produced with + parameters to the usbtest suite) */ + if(usb_pipeisoc(urb->pipe) && usb_pipein(urb->pipe) && + (urb->transfer_buffer_length < maxpacket)) { + tc_err("Submit In Isoc URB with buffer length:%d to pipe with maxpacketlen: %d\n", urb->transfer_buffer_length, maxpacket); + return -EMSGSIZE; + } + + /* Check if there is a epid for URBs destination, if not this function + set up one. */ + epid = tc_setup_epid(urb->ep, urb, mem_flags); + if (epid < 0) { + tc_err("Failed setup epid:%d for URB:0x%x\n", epid, (unsigned int)urb); + DBFEXIT; + return -ENOMEM; + } + + if(urb == activeUrbList[epid]) { + tc_err("Resubmition of allready active URB:0x%x\n", (unsigned int)urb); + return -ENXIO; + } + + if(urb_list_entry(urb, epid)) { + tc_err("Resubmition of allready queued URB:0x%x\n", (unsigned int)urb); + return -ENXIO; + } + + /* If we actively have flaged endpoint as disabled then refuse submition */ + if(epid_state[epid].disabled) { + return -ENOENT; + } + + /* Allocate and init HC-private data for URB */ + if(urb_priv_create(hcd, urb, epid, mem_flags) != 0) { + DBFEXIT; + return -ENOMEM; + } + urb_priv = urb->hcpriv; + + /* Check if there is enough bandwidth for periodic transfer */ + if(usb_pipeint(urb->pipe) || usb_pipeisoc(urb->pipe)) { + /* only check (and later claim) if not already claimed */ + if (urb_priv->bandwidth == 0) { + bustime = crisv10_usb_check_bandwidth(urb->dev, urb); + if (bustime < 0) { + tc_err("Not enough periodic bandwidth\n"); + urb_priv_free(hcd, urb); + DBFEXIT; + return -ENOSPC; + } + } + } + + tc_dbg("Enqueue URB:0x%x[%d] epid:%d (%s) bufflen:%d\n", + (unsigned int)urb, urb_priv->urb_num, epid, + pipe_to_str(urb->pipe), urb->transfer_buffer_length); + + /* Create and link SBs required for this URB */ + retval = create_sb_for_urb(urb, mem_flags); + if(retval != 0) { + tc_err("Failed to create SBs for URB:0x%x[%d]\n", (unsigned int)urb, + urb_priv->urb_num); + urb_priv_free(hcd, urb); + DBFEXIT; + return retval; + } + + /* Init intr EP pool if this URB is a INTR transfer. This pool is later + used when inserting EPs in the TxIntrEPList. We do the alloc here + so we can't run out of memory later */ + if(usb_pipeint(urb->pipe)) { + retval = init_intr_urb(urb, mem_flags); + if(retval != 0) { + tc_warn("Failed to init Intr URB\n"); + urb_priv_free(hcd, urb); + DBFEXIT; + return retval; + } + } + + /* Disable other access when inserting USB */ + local_irq_save(flags); + + /* Claim bandwidth, if needed */ + if(bustime) { + crisv10_usb_claim_bandwidth(urb->dev, + urb, + bustime, + (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS)); + } + + /* Add URB to EP queue */ + urb_list_add(urb, epid, mem_flags); + + if(usb_pipeisoc(urb->pipe)) { + /* Special processing of Isoc URBs. */ + tc_dma_process_isoc_urb(urb); + } else { + /* Process EP queue for rest of the URB types (Bulk, Ctrl, Intr) */ + tc_dma_process_queue(epid); + } + + local_irq_restore(flags); + + DBFEXIT; + return 0; +} + +/* remove an URB from the transfer controller queues (called from hcd_driver)*/ +static int tc_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status) { + struct crisv10_urb_priv *urb_priv; + unsigned long flags; + int epid; + + DBFENTER; + /* Disable interrupts here since a descriptor interrupt for the isoc epid + will modify the sb list. This could possibly be done more granular, but + urb_dequeue should not be used frequently anyway. + */ + local_irq_save(flags); + + urb->status = status; + urb_priv = urb->hcpriv; + + if (!urb_priv) { + /* This happens if a device driver calls unlink on an urb that + was never submitted (lazy driver) or if the urb was completed + while dequeue was being called. */ + tc_warn("Dequeing of not enqueued URB:0x%x\n", (unsigned int)urb); + local_irq_restore(flags); + return 0; + } + epid = urb_priv->epid; + + tc_warn("Dequeing %s URB:0x%x[%d] (%s %s epid:%d) status:%d %s\n", + (urb == activeUrbList[epid]) ? "active" : "queued", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), epid, urb->status, + (urb_priv->later_data) ? "later-sched" : ""); + + /* For Bulk, Ctrl and Intr are only one URB active at a time. So any URB + that isn't active can be dequeued by just removing it from the queue */ + if(usb_pipebulk(urb->pipe) || usb_pipecontrol(urb->pipe) || + usb_pipeint(urb->pipe)) { + + /* Check if URB haven't gone further than the queue */ + if(urb != activeUrbList[epid]) { + ASSERT(urb_priv->later_data == NULL); + tc_warn("Dequeing URB:0x%x[%d] (%s %s epid:%d) from queue" + " (not active)\n", (unsigned int)urb, urb_priv->urb_num, + str_dir(urb->pipe), str_type(urb->pipe), epid); + + /* Finish the URB with error status from USB core */ + tc_finish_urb(hcd, urb, urb->status); + local_irq_restore(flags); + return 0; + } + } + + /* Set URB status to Unlink for handling when interrupt comes. */ + urb_priv->urb_state = UNLINK; + + /* Differentiate dequeing of Bulk and Ctrl from Isoc and Intr */ + switch(usb_pipetype(urb->pipe)) { + case PIPE_BULK: + /* Check if EP still is enabled */ + if (TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxBulkEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + } + /* Kicking dummy list out of the party. */ + TxBulkEPList[epid].next = virt_to_phys(&TxBulkEPList[(epid + 1) % NBR_OF_EPIDS]); + break; + case PIPE_CONTROL: + /* Check if EP still is enabled */ + if (TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxCtrlEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + } + break; + case PIPE_ISOCHRONOUS: + /* Disabling, busy-wait and unlinking of Isoc SBs will be done in + finish_isoc_urb(). Because there might the case when URB is dequeued + but there are other valid URBs waiting */ + + /* Check if In Isoc EP still is enabled */ + if (TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxIsocEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + } + break; + case PIPE_INTERRUPT: + /* Special care is taken for interrupt URBs. EPs are unlinked in + tc_finish_urb */ + break; + default: + break; + } + + /* Asynchronous unlink, finish the URB later from scheduled or other + event (data finished, error) */ + tc_finish_urb_later(hcd, urb, urb->status); + + local_irq_restore(flags); + DBFEXIT; + return 0; +} + + +static void tc_sync_finish_epid(struct usb_hcd *hcd, int epid) { + volatile int timeout = 10000; + struct urb* urb; + struct crisv10_urb_priv* urb_priv; + unsigned long flags; + + volatile struct USB_EP_Desc *first_ep; /* First EP in the list. */ + volatile struct USB_EP_Desc *curr_ep; /* Current EP, the iterator. */ + volatile struct USB_EP_Desc *next_ep; /* The EP after current. */ + + int type = epid_state[epid].type; + + /* Setting this flag will cause enqueue() to return -ENOENT for new + submitions on this endpoint and finish_urb() wont process queue further */ + epid_state[epid].disabled = 1; + + switch(type) { + case PIPE_BULK: + /* Check if EP still is enabled */ + if (TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxBulkEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + tc_warn("sync_finish: Disabling EP for epid:%d\n", epid); + + /* Do busy-wait until DMA not using this EP descriptor anymore */ + while((*R_DMA_CH8_SUB0_EP == + virt_to_phys(&TxBulkEPList[epid])) && + (timeout-- > 0)); + + } + break; + + case PIPE_CONTROL: + /* Check if EP still is enabled */ + if (TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + /* The EP was enabled, disable it. */ + TxCtrlEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + tc_warn("sync_finish: Disabling EP for epid:%d\n", epid); + + /* Do busy-wait until DMA not using this EP descriptor anymore */ + while((*R_DMA_CH8_SUB1_EP == + virt_to_phys(&TxCtrlEPList[epid])) && + (timeout-- > 0)); + } + break; + + case PIPE_INTERRUPT: + local_irq_save(flags); + /* Disable all Intr EPs belonging to epid */ + first_ep = &TxIntrEPList[0]; + curr_ep = first_ep; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if (IO_EXTRACT(USB_EP_command, epid, next_ep->command) == epid) { + /* Disable EP */ + next_ep->command &= ~IO_MASK(USB_EP_command, enable); + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != first_ep); + + local_irq_restore(flags); + break; + + case PIPE_ISOCHRONOUS: + /* Check if EP still is enabled */ + if (TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + tc_warn("sync_finish: Disabling Isoc EP for epid:%d\n", epid); + /* The EP was enabled, disable it. */ + TxIsocEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + + while((*R_DMA_CH8_SUB3_EP == virt_to_phys(&TxIsocEPList[epid])) && + (timeout-- > 0)); + } + break; + } + + local_irq_save(flags); + + /* Finish if there is active URB for this endpoint */ + if(activeUrbList[epid] != NULL) { + urb = activeUrbList[epid]; + urb_priv = urb->hcpriv; + ASSERT(urb_priv); + tc_warn("Sync finish %s URB:0x%x[%d] (%s %s epid:%d) status:%d %s\n", + (urb == activeUrbList[epid]) ? "active" : "queued", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), epid, urb->status, + (urb_priv->later_data) ? "later-sched" : ""); + + tc_finish_urb(hcd, activeUrbList[epid], -ENOENT); + ASSERT(activeUrbList[epid] == NULL); + } + + /* Finish any queued URBs for this endpoint. There won't be any resubmitions + because epid_disabled causes enqueue() to fail for this endpoint */ + while((urb = urb_list_first(epid)) != NULL) { + urb_priv = urb->hcpriv; + ASSERT(urb_priv); + + tc_warn("Sync finish %s URB:0x%x[%d] (%s %s epid:%d) status:%d %s\n", + (urb == activeUrbList[epid]) ? "active" : "queued", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), epid, urb->status, + (urb_priv->later_data) ? "later-sched" : ""); + + tc_finish_urb(hcd, urb, -ENOENT); + } + epid_state[epid].disabled = 0; + local_irq_restore(flags); +} + +/* free resources associated with an endpoint (called from hcd_driver) */ +static void tc_endpoint_disable(struct usb_hcd *hcd, + struct usb_host_endpoint *ep) { + DBFENTER; + /* Only free epid if it has been allocated. We get two endpoint_disable + requests for ctrl endpoints so ignore the second one */ + if(ep->hcpriv != NULL) { + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + int epid = ep_priv->epid; + tc_warn("endpoint_disable ep:0x%x ep-priv:0x%x (%s) (epid:%d freed)\n", + (unsigned int)ep, (unsigned int)ep->hcpriv, + endpoint_to_str(&(ep->desc)), epid); + + tc_sync_finish_epid(hcd, epid); + + ASSERT(activeUrbList[epid] == NULL); + ASSERT(list_empty(&urb_list[epid])); + + tc_free_epid(ep); + } else { + tc_dbg("endpoint_disable ep:0x%x ep-priv:0x%x (%s)\n", (unsigned int)ep, + (unsigned int)ep->hcpriv, endpoint_to_str(&(ep->desc))); + } + DBFEXIT; +} + +static void tc_finish_urb_later_proc(struct work_struct* work) { + unsigned long flags; + struct urb_later_data* uld; + + local_irq_save(flags); + uld = container_of(work, struct urb_later_data, dws.work); + if(uld->urb == NULL) { + late_dbg("Later finish of URB = NULL (allready finished)\n"); + } else { + struct crisv10_urb_priv* urb_priv = uld->urb->hcpriv; + ASSERT(urb_priv); + if(urb_priv->urb_num == uld->urb_num) { + late_dbg("Later finish of URB:0x%x[%d]\n", (unsigned int)(uld->urb), + urb_priv->urb_num); + if(uld->status != uld->urb->status) { + errno_dbg("Later-finish URB with status:%d, later-status:%d\n", + uld->urb->status, uld->status); + } + if(uld != urb_priv->later_data) { + panic("Scheduled uld not same as URBs uld\n"); + } + tc_finish_urb(uld->hcd, uld->urb, uld->status); + } else { + late_warn("Ignoring later finish of URB:0x%x[%d]" + ", urb_num doesn't match current URB:0x%x[%d]", + (unsigned int)(uld->urb), uld->urb_num, + (unsigned int)(uld->urb), urb_priv->urb_num); + } + } + local_irq_restore(flags); + kmem_cache_free(later_data_cache, uld); +} + +static void tc_finish_urb_later(struct usb_hcd *hcd, struct urb *urb, + int status) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + struct urb_later_data* uld; + + ASSERT(urb_priv); + + if(urb_priv->later_data != NULL) { + /* Later-finish allready scheduled for this URB, just update status to + return when finishing later */ + errno_dbg("Later-finish schedule change URB status:%d with new" + " status:%d\n", urb_priv->later_data->status, status); + + urb_priv->later_data->status = status; + return; + } + + uld = kmem_cache_alloc(later_data_cache, GFP_ATOMIC); + ASSERT(uld); + + uld->hcd = hcd; + uld->urb = urb; + uld->urb_num = urb_priv->urb_num; + uld->status = status; + + INIT_DELAYED_WORK(&uld->dws, tc_finish_urb_later_proc); + urb_priv->later_data = uld; + + /* Schedule the finishing of the URB to happen later */ + schedule_delayed_work(&uld->dws, LATER_TIMER_DELAY); +} + +static void tc_finish_isoc_urb(struct usb_hcd *hcd, struct urb *urb, + int status); + +static void tc_finish_urb(struct usb_hcd *hcd, struct urb *urb, int status) { + struct crisv10_hcd* crisv10_hcd = hcd_to_crisv10_hcd(hcd); + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + int epid; + char toggle; + int urb_num; + + DBFENTER; + ASSERT(urb_priv != NULL); + epid = urb_priv->epid; + urb_num = urb_priv->urb_num; + + if(urb != activeUrbList[epid]) { + if(urb_list_entry(urb, epid)) { + /* Remove this URB from the list. Only happens when URB are finished + before having been processed (dequeing) */ + urb_list_del(urb, epid); + } else { + tc_warn("Finishing of URB:0x%x[%d] neither active or in queue for" + " epid:%d\n", (unsigned int)urb, urb_num, epid); + } + } + + /* Cancel any pending later-finish of this URB */ + if(urb_priv->later_data) { + urb_priv->later_data->urb = NULL; + } + + /* For an IN pipe, we always set the actual length, regardless of whether + there was an error or not (which means the device driver can use the data + if it wants to). */ + if(usb_pipein(urb->pipe)) { + urb->actual_length = urb_priv->rx_offset; + } else { + /* Set actual_length for OUT urbs also; the USB mass storage driver seems + to want that. */ + if (status == 0 && urb->status == -EINPROGRESS) { + urb->actual_length = urb->transfer_buffer_length; + } else { + /* We wouldn't know of any partial writes if there was an error. */ + urb->actual_length = 0; + } + } + + + /* URB status mangling */ + if(urb->status == -EINPROGRESS) { + /* The USB core hasn't changed the status, let's set our finish status */ + urb->status = status; + + if ((status == 0) && (urb->transfer_flags & URB_SHORT_NOT_OK) && + usb_pipein(urb->pipe) && + (urb->actual_length != urb->transfer_buffer_length)) { + /* URB_SHORT_NOT_OK means that short reads (shorter than the endpoint's + max length) is to be treated as an error. */ + errno_dbg("Finishing URB:0x%x[%d] with SHORT_NOT_OK flag and short" + " data:%d\n", (unsigned int)urb, urb_num, + urb->actual_length); + urb->status = -EREMOTEIO; + } + + if(urb_priv->urb_state == UNLINK) { + /* URB has been requested to be unlinked asynchronously */ + urb->status = -ECONNRESET; + errno_dbg("Fixing unlink status of URB:0x%x[%d] to:%d\n", + (unsigned int)urb, urb_num, urb->status); + } + } else { + /* The USB Core wants to signal some error via the URB, pass it through */ + } + + /* use completely different finish function for Isoc URBs */ + if(usb_pipeisoc(urb->pipe)) { + tc_finish_isoc_urb(hcd, urb, status); + return; + } + + /* Do special unlinking of EPs for Intr traffic */ + if(usb_pipeint(urb->pipe)) { + tc_dma_unlink_intr_urb(urb); + } + + /* Release allocated bandwidth for periodic transfers */ + if(usb_pipeint(urb->pipe) || usb_pipeisoc(urb->pipe)) + crisv10_usb_release_bandwidth(hcd, + usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS, + urb_priv->bandwidth); + + /* This URB is active on EP */ + if(urb == activeUrbList[epid]) { + /* We need to fiddle with the toggle bits because the hardware doesn't do + it for us. */ + toggle = etrax_epid_get_toggle(epid, usb_pipeout(urb->pipe)); + usb_settoggle(urb->dev, usb_pipeendpoint(urb->pipe), + usb_pipeout(urb->pipe), toggle); + + /* Checks for Ctrl and Bulk EPs */ + switch(usb_pipetype(urb->pipe)) { + case PIPE_BULK: + /* Check so Bulk EP realy is disabled before finishing active URB */ + ASSERT((TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) == + IO_STATE(USB_EP_command, enable, no)); + /* Disable sub-pointer for EP to avoid next tx_interrupt() to + process Bulk EP. */ + TxBulkEPList[epid].sub = 0; + /* No need to wait for the DMA before changing the next pointer. + The modulo NBR_OF_EPIDS isn't actually necessary, since we will never use + the last one (INVALID_EPID) for actual traffic. */ + TxBulkEPList[epid].next = + virt_to_phys(&TxBulkEPList[(epid + 1) % NBR_OF_EPIDS]); + break; + case PIPE_CONTROL: + /* Check so Ctrl EP realy is disabled before finishing active URB */ + ASSERT((TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) == + IO_STATE(USB_EP_command, enable, no)); + /* Disable sub-pointer for EP to avoid next tx_interrupt() to + process Ctrl EP. */ + TxCtrlEPList[epid].sub = 0; + break; + } + } + + /* Free HC-private URB data*/ + urb_priv_free(hcd, urb); + + if(urb->status) { + errno_dbg("finish_urb (URB:0x%x[%d] %s %s) (data:%d) status:%d\n", + (unsigned int)urb, urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb->actual_length, urb->status); + } else { + tc_dbg("finish_urb (URB:0x%x[%d] %s %s) (data:%d) status:%d\n", + (unsigned int)urb, urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb->actual_length, urb->status); + } + + /* If we just finished an active URB, clear active pointer. */ + if (urb == activeUrbList[epid]) { + /* Make URB not active on EP anymore */ + activeUrbList[epid] = NULL; + + if(urb->status == 0) { + /* URB finished sucessfully, process queue to see if there are any more + URBs waiting before we call completion function.*/ + if(crisv10_hcd->running) { + /* Only process queue if USB controller is running */ + tc_dma_process_queue(epid); + } else { + tc_warn("No processing of queue for epid:%d, USB Controller not" + " running\n", epid); + } + } + } + + /* Hand the URB from HCD to its USB device driver, using its completion + functions */ + usb_hcd_giveback_urb (hcd, urb, status); + + /* Check the queue once more if the URB returned with error, because we + didn't do it before the completion function because the specification + states that the queue should not restart until all it's unlinked + URBs have been fully retired, with the completion functions run */ + if(crisv10_hcd->running) { + /* Only process queue if USB controller is running */ + tc_dma_process_queue(epid); + } else { + tc_warn("No processing of queue for epid:%d, USB Controller not running\n", + epid); + } + + DBFEXIT; +} + +static void tc_finish_isoc_urb(struct usb_hcd *hcd, struct urb *urb, + int status) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + int epid, i; + volatile int timeout = 10000; + int bandwidth = 0; + + ASSERT(urb_priv); + epid = urb_priv->epid; + + ASSERT(usb_pipeisoc(urb->pipe)); + + /* Set that all isoc packets have status and length set before + completing the urb. */ + for (i = urb_priv->isoc_packet_counter; i < urb->number_of_packets; i++){ + urb->iso_frame_desc[i].actual_length = 0; + urb->iso_frame_desc[i].status = -EPROTO; + } + + /* Check if the URB is currently active (done or error) */ + if(urb == activeUrbList[epid]) { + /* Check if there are another In Isoc URB queued for this epid */ + if (!list_empty(&urb_list[epid])&& !epid_state[epid].disabled) { + /* Move it from queue to active and mark it started so Isoc transfers + won't be interrupted. + All Isoc URBs data transfers are already added to DMA lists so we + don't have to insert anything in DMA lists here. */ + activeUrbList[epid] = urb_list_first(epid); + ((struct crisv10_urb_priv *)(activeUrbList[epid]->hcpriv))->urb_state = + STARTED; + urb_list_del(activeUrbList[epid], epid); + + if(urb->status) { + errno_dbg("finish_isoc_urb (URB:0x%x[%d] %s %s) (%d of %d packets)" + " status:%d, new waiting URB:0x%x[%d]\n", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb_priv->isoc_packet_counter, + urb->number_of_packets, urb->status, + (unsigned int)activeUrbList[epid], + ((struct crisv10_urb_priv *)(activeUrbList[epid]->hcpriv))->urb_num); + } + + } else { /* No other URB queued for this epid */ + if(urb->status) { + errno_dbg("finish_isoc_urb (URB:0x%x[%d] %s %s) (%d of %d packets)" + " status:%d, no new URB waiting\n", + (unsigned int)urb, urb_priv->urb_num, str_dir(urb->pipe), + str_type(urb->pipe), urb_priv->isoc_packet_counter, + urb->number_of_packets, urb->status); + } + + /* Check if EP is still enabled, then shut it down. */ + if (TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + isoc_dbg("Isoc EP enabled for epid:%d, disabling it\n", epid); + + /* Should only occur for In Isoc EPs where SB isn't consumed. */ + ASSERT(usb_pipein(urb->pipe)); + + /* Disable it and wait for it to stop */ + TxIsocEPList[epid].command &= ~IO_MASK(USB_EP_command, enable); + + /* Ah, the luxury of busy-wait. */ + while((*R_DMA_CH8_SUB3_EP == virt_to_phys(&TxIsocEPList[epid])) && + (timeout-- > 0)); + } + + /* Unlink SB to say that epid is finished. */ + TxIsocEPList[epid].sub = 0; + TxIsocEPList[epid].hw_len = 0; + + /* No URB active for EP anymore */ + activeUrbList[epid] = NULL; + } + } else { /* Finishing of not active URB (queued up with SBs thought) */ + isoc_warn("finish_isoc_urb (URB:0x%x %s) (%d of %d packets) status:%d," + " SB queued but not active\n", + (unsigned int)urb, str_dir(urb->pipe), + urb_priv->isoc_packet_counter, urb->number_of_packets, + urb->status); + if(usb_pipeout(urb->pipe)) { + /* Finishing of not yet active Out Isoc URB needs unlinking of SBs. */ + struct USB_SB_Desc *iter_sb, *prev_sb, *next_sb; + + iter_sb = TxIsocEPList[epid].sub ? + phys_to_virt(TxIsocEPList[epid].sub) : 0; + prev_sb = 0; + + /* SB that is linked before this URBs first SB */ + while (iter_sb && (iter_sb != urb_priv->first_sb)) { + prev_sb = iter_sb; + iter_sb = iter_sb->next ? phys_to_virt(iter_sb->next) : 0; + } + + if (iter_sb == 0) { + /* Unlink of the URB currently being transmitted. */ + prev_sb = 0; + iter_sb = TxIsocEPList[epid].sub ? phys_to_virt(TxIsocEPList[epid].sub) : 0; + } + + while (iter_sb && (iter_sb != urb_priv->last_sb)) { + iter_sb = iter_sb->next ? phys_to_virt(iter_sb->next) : 0; + } + + if (iter_sb) { + next_sb = iter_sb->next ? phys_to_virt(iter_sb->next) : 0; + } else { + /* This should only happen if the DMA has completed + processing the SB list for this EP while interrupts + are disabled. */ + isoc_dbg("Isoc urb not found, already sent?\n"); + next_sb = 0; + } + if (prev_sb) { + prev_sb->next = next_sb ? virt_to_phys(next_sb) : 0; + } else { + TxIsocEPList[epid].sub = next_sb ? virt_to_phys(next_sb) : 0; + } + } + } + + /* Free HC-private URB data*/ + bandwidth = urb_priv->bandwidth; + urb_priv_free(hcd, urb); + + crisv10_usb_release_bandwidth(hcd, usb_pipeisoc(urb->pipe), bandwidth); + + /* Hand the URB from HCD to its USB device driver, using its completion + functions */ + usb_hcd_giveback_urb (hcd, urb, status); +} + +static __u32 urb_num = 0; + +/* allocate and initialize URB private data */ +static int urb_priv_create(struct usb_hcd *hcd, struct urb *urb, int epid, + int mem_flags) { + struct crisv10_urb_priv *urb_priv; + + urb_priv = kmalloc(sizeof *urb_priv, mem_flags); + if (!urb_priv) + return -ENOMEM; + memset(urb_priv, 0, sizeof *urb_priv); + + urb_priv->epid = epid; + urb_priv->urb_state = NOT_STARTED; + + urb->hcpriv = urb_priv; + /* Assign URB a sequence number, and increment counter */ + urb_priv->urb_num = urb_num; + urb_num++; + urb_priv->bandwidth = 0; + return 0; +} + +/* free URB private data */ +static void urb_priv_free(struct usb_hcd *hcd, struct urb *urb) { + int i; + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + ASSERT(urb_priv != 0); + + /* Check it has any SBs linked that needs to be freed*/ + if(urb_priv->first_sb != NULL) { + struct USB_SB_Desc *next_sb, *first_sb, *last_sb; + int i = 0; + first_sb = urb_priv->first_sb; + last_sb = urb_priv->last_sb; + ASSERT(last_sb); + while(first_sb != last_sb) { + next_sb = (struct USB_SB_Desc *)phys_to_virt(first_sb->next); + kmem_cache_free(usb_desc_cache, first_sb); + first_sb = next_sb; + i++; + } + kmem_cache_free(usb_desc_cache, last_sb); + i++; + } + + /* Check if it has any EPs in its Intr pool that also needs to be freed */ + if(urb_priv->intr_ep_pool_length > 0) { + for(i = 0; i < urb_priv->intr_ep_pool_length; i++) { + kfree(urb_priv->intr_ep_pool[i]); + } + /* + tc_dbg("Freed %d EPs from URB:0x%x EP pool\n", + urb_priv->intr_ep_pool_length, (unsigned int)urb); + */ + } + + kfree(urb_priv); + urb->hcpriv = NULL; +} + +static int ep_priv_create(struct usb_host_endpoint *ep, int mem_flags) { + struct crisv10_ep_priv *ep_priv; + + ep_priv = kmalloc(sizeof *ep_priv, mem_flags); + if (!ep_priv) + return -ENOMEM; + memset(ep_priv, 0, sizeof *ep_priv); + + ep->hcpriv = ep_priv; + return 0; +} + +static void ep_priv_free(struct usb_host_endpoint *ep) { + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + ASSERT(ep_priv); + kfree(ep_priv); + ep->hcpriv = NULL; +} + +/* + * usb_check_bandwidth(): + * + * old_alloc is from host_controller->bandwidth_allocated in microseconds; + * bustime is from calc_bus_time(), but converted to microseconds. + * + * returns if successful, + * or -ENOSPC if bandwidth request fails. + * + * FIXME: + * This initial implementation does not use Endpoint.bInterval + * in managing bandwidth allocation. + * It probably needs to be expanded to use Endpoint.bInterval. + * This can be done as a later enhancement (correction). + * + * This will also probably require some kind of + * frame allocation tracking...meaning, for example, + * that if multiple drivers request interrupts every 10 USB frames, + * they don't all have to be allocated at + * frame numbers N, N+10, N+20, etc. Some of them could be at + * N+11, N+21, N+31, etc., and others at + * N+12, N+22, N+32, etc. + * + * Similarly for isochronous transfers... + * + * Individual HCDs can schedule more directly ... this logic + * is not correct for high speed transfers. + */ +static int crisv10_usb_check_bandwidth( + struct usb_device *dev, + struct urb *urb) +{ + unsigned int pipe = urb->pipe; + long bustime; + int is_in = usb_pipein (pipe); + int is_iso = usb_pipeisoc (pipe); + int old_alloc = dev->bus->bandwidth_allocated; + int new_alloc; + + bustime = NS_TO_US (usb_calc_bus_time (dev->speed, is_in, is_iso, + usb_maxpacket (dev, pipe, !is_in))); + if (is_iso) + bustime /= urb->number_of_packets; + + new_alloc = old_alloc + (int) bustime; + if (new_alloc > FRAME_TIME_MAX_USECS_ALLOC) { + dev_dbg (&dev->dev, "usb_check_bandwidth FAILED: %d + %ld = %d usec\n", + old_alloc, bustime, new_alloc); + bustime = -ENOSPC; /* report error */ + } + + return bustime; +} + +/** + * usb_claim_bandwidth - records bandwidth for a periodic transfer + * @dev: source/target of request + * @urb: request (urb->dev == dev) + * @bustime: bandwidth consumed, in (average) microseconds per frame + * @isoc: true iff the request is isochronous + * + * HCDs are expected not to overcommit periodic bandwidth, and to record such + * reservations whenever endpoints are added to the periodic schedule. + * + * FIXME averaging per-frame is suboptimal. Better to sum over the HCD's + * entire periodic schedule ... 32 frames for OHCI, 1024 for UHCI, settable + * for EHCI (256/512/1024 frames, default 1024) and have the bus expose how + * large its periodic schedule is. + */ +static void crisv10_usb_claim_bandwidth( + struct usb_device *dev, + struct urb *urb, int bustime, int isoc) +{ + dev->bus->bandwidth_allocated += bustime; + if (isoc) + dev->bus->bandwidth_isoc_reqs++; + else + dev->bus->bandwidth_int_reqs++; + struct crisv10_urb_priv *urb_priv; + urb_priv = urb->hcpriv; + urb_priv->bandwidth = bustime; +} + +/** + * usb_release_bandwidth - reverses effect of usb_claim_bandwidth() + * @hcd: host controller + * @isoc: true iff the request is isochronous + * @bandwidth: bandwidth returned + * + * This records that previously allocated bandwidth has been released. + * Bandwidth is released when endpoints are removed from the host controller's + * periodic schedule. + */ +static void crisv10_usb_release_bandwidth( + struct usb_hcd *hcd, + int isoc, + int bandwidth) +{ + hcd_to_bus(hcd)->bandwidth_allocated -= bandwidth; + if (isoc) + hcd_to_bus(hcd)->bandwidth_isoc_reqs--; + else + hcd_to_bus(hcd)->bandwidth_int_reqs--; +} + + +/* EPID handling functions, managing EP-list in Etrax through wrappers */ +/* ------------------------------------------------------------------- */ + +/* Sets up a new EPID for an endpoint or returns existing if found */ +static int tc_setup_epid(struct usb_host_endpoint *ep, struct urb *urb, + int mem_flags) { + int epid; + char devnum, endpoint, out_traffic, slow; + int maxlen; + __u32 epid_data; + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + + DBFENTER; + + /* Check if a valid epid already is setup for this endpoint */ + if(ep_priv != NULL) { + return ep_priv->epid; + } + + /* We must find and initiate a new epid for this urb. */ + epid = tc_allocate_epid(); + + if (epid == -1) { + /* Failed to allocate a new epid. */ + DBFEXIT; + return epid; + } + + /* We now have a new epid to use. Claim it. */ + epid_state[epid].inuse = 1; + + /* Init private data for new endpoint */ + if(ep_priv_create(ep, mem_flags) != 0) { + return -ENOMEM; + } + ep_priv = ep->hcpriv; + ep_priv->epid = epid; + + devnum = usb_pipedevice(urb->pipe); + endpoint = usb_pipeendpoint(urb->pipe); + slow = (urb->dev->speed == USB_SPEED_LOW); + maxlen = usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe)); + + if (usb_pipetype(urb->pipe) == PIPE_CONTROL) { + /* We want both IN and OUT control traffic to be put on the same + EP/SB list. */ + out_traffic = 1; + } else { + out_traffic = usb_pipeout(urb->pipe); + } + + if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) { + epid_data = IO_STATE(R_USB_EPT_DATA_ISO, valid, yes) | + /* FIXME: Change any to the actual port? */ + IO_STATE(R_USB_EPT_DATA_ISO, port, any) | + IO_FIELD(R_USB_EPT_DATA_ISO, max_len, maxlen) | + IO_FIELD(R_USB_EPT_DATA_ISO, ep, endpoint) | + IO_FIELD(R_USB_EPT_DATA_ISO, dev, devnum); + etrax_epid_iso_set(epid, epid_data); + } else { + epid_data = IO_STATE(R_USB_EPT_DATA, valid, yes) | + IO_FIELD(R_USB_EPT_DATA, low_speed, slow) | + /* FIXME: Change any to the actual port? */ + IO_STATE(R_USB_EPT_DATA, port, any) | + IO_FIELD(R_USB_EPT_DATA, max_len, maxlen) | + IO_FIELD(R_USB_EPT_DATA, ep, endpoint) | + IO_FIELD(R_USB_EPT_DATA, dev, devnum); + etrax_epid_set(epid, epid_data); + } + + epid_state[epid].out_traffic = out_traffic; + epid_state[epid].type = usb_pipetype(urb->pipe); + + tc_warn("Setting up ep:0x%x epid:%d (addr:%d endp:%d max_len:%d %s %s %s)\n", + (unsigned int)ep, epid, devnum, endpoint, maxlen, + str_type(urb->pipe), out_traffic ? "out" : "in", + slow ? "low" : "full"); + + /* Enable Isoc eof interrupt if we set up the first Isoc epid */ + if(usb_pipeisoc(urb->pipe)) { + isoc_epid_counter++; + if(isoc_epid_counter == 1) { + isoc_warn("Enabled Isoc eof interrupt\n"); + *R_USB_IRQ_MASK_SET = IO_STATE(R_USB_IRQ_MASK_SET, iso_eof, set); + } + } + + DBFEXIT; + return epid; +} + +static void tc_free_epid(struct usb_host_endpoint *ep) { + unsigned long flags; + struct crisv10_ep_priv *ep_priv = ep->hcpriv; + int epid; + volatile int timeout = 10000; + + DBFENTER; + + if (ep_priv == NULL) { + tc_warn("Trying to free unused epid on ep:0x%x\n", (unsigned int)ep); + DBFEXIT; + return; + } + + epid = ep_priv->epid; + + /* Disable Isoc eof interrupt if we free the last Isoc epid */ + if(epid_isoc(epid)) { + ASSERT(isoc_epid_counter > 0); + isoc_epid_counter--; + if(isoc_epid_counter == 0) { + *R_USB_IRQ_MASK_CLR = IO_STATE(R_USB_IRQ_MASK_CLR, iso_eof, clr); + isoc_warn("Disabled Isoc eof interrupt\n"); + } + } + + /* Take lock manualy instead of in epid_x_x wrappers, + because we need to be polling here */ + spin_lock_irqsave(&etrax_epid_lock, flags); + + *R_USB_EPT_INDEX = IO_FIELD(R_USB_EPT_INDEX, value, epid); + nop(); + while((*R_USB_EPT_DATA & IO_MASK(R_USB_EPT_DATA, hold)) && + (timeout-- > 0)); + /* This will, among other things, set the valid field to 0. */ + *R_USB_EPT_DATA = 0; + spin_unlock_irqrestore(&etrax_epid_lock, flags); + + /* Free resource in software state info list */ + epid_state[epid].inuse = 0; + + /* Free private endpoint data */ + ep_priv_free(ep); + + DBFEXIT; +} + +static int tc_allocate_epid(void) { + int i; + DBFENTER; + for (i = 0; i < NBR_OF_EPIDS; i++) { + if (!epid_inuse(i)) { + DBFEXIT; + return i; + } + } + + tc_warn("Found no free epids\n"); + DBFEXIT; + return -1; +} + + +/* Wrappers around the list functions (include/linux/list.h). */ +/* ---------------------------------------------------------- */ +static inline int __urb_list_empty(int epid) { + int retval; + retval = list_empty(&urb_list[epid]); + return retval; +} + +/* Returns first urb for this epid, or NULL if list is empty. */ +static inline struct urb *urb_list_first(int epid) { + unsigned long flags; + struct urb *first_urb = 0; + spin_lock_irqsave(&urb_list_lock, flags); + if (!__urb_list_empty(epid)) { + /* Get the first urb (i.e. head->next). */ + urb_entry_t *urb_entry = list_entry((&urb_list[epid])->next, urb_entry_t, list); + first_urb = urb_entry->urb; + } + spin_unlock_irqrestore(&urb_list_lock, flags); + return first_urb; +} + +/* Adds an urb_entry last in the list for this epid. */ +static inline void urb_list_add(struct urb *urb, int epid, int mem_flags) { + unsigned long flags; + urb_entry_t *urb_entry = (urb_entry_t *)kmalloc(sizeof(urb_entry_t), mem_flags); + ASSERT(urb_entry); + + urb_entry->urb = urb; + spin_lock_irqsave(&urb_list_lock, flags); + list_add_tail(&urb_entry->list, &urb_list[epid]); + spin_unlock_irqrestore(&urb_list_lock, flags); +} + +/* Search through the list for an element that contains this urb. (The list + is expected to be short and the one we are about to delete will often be + the first in the list.) + Should be protected by spin_locks in calling function */ +static inline urb_entry_t *__urb_list_entry(struct urb *urb, int epid) { + struct list_head *entry; + struct list_head *tmp; + urb_entry_t *urb_entry; + + list_for_each_safe(entry, tmp, &urb_list[epid]) { + urb_entry = list_entry(entry, urb_entry_t, list); + ASSERT(urb_entry); + ASSERT(urb_entry->urb); + + if (urb_entry->urb == urb) { + return urb_entry; + } + } + return 0; +} + +/* Same function as above but for global use. Protects list by spinlock */ +static inline urb_entry_t *urb_list_entry(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + spin_unlock_irqrestore(&urb_list_lock, flags); + return (urb_entry); +} + +/* Delete an urb from the list. */ +static inline void urb_list_del(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + + /* Delete entry and free. */ + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + ASSERT(urb_entry); + + list_del(&urb_entry->list); + spin_unlock_irqrestore(&urb_list_lock, flags); + kfree(urb_entry); +} + +/* Move an urb to the end of the list. */ +static inline void urb_list_move_last(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + ASSERT(urb_entry); + + list_del(&urb_entry->list); + list_add_tail(&urb_entry->list, &urb_list[epid]); + spin_unlock_irqrestore(&urb_list_lock, flags); +} + +/* Get the next urb in the list. */ +static inline struct urb *urb_list_next(struct urb *urb, int epid) { + unsigned long flags; + urb_entry_t *urb_entry; + + spin_lock_irqsave(&urb_list_lock, flags); + urb_entry = __urb_list_entry(urb, epid); + ASSERT(urb_entry); + + if (urb_entry->list.next != &urb_list[epid]) { + struct list_head *elem = urb_entry->list.next; + urb_entry = list_entry(elem, urb_entry_t, list); + spin_unlock_irqrestore(&urb_list_lock, flags); + return urb_entry->urb; + } else { + spin_unlock_irqrestore(&urb_list_lock, flags); + return NULL; + } +} + +struct USB_EP_Desc* create_ep(int epid, struct USB_SB_Desc* sb_desc, + int mem_flags) { + struct USB_EP_Desc *ep_desc; + ep_desc = (struct USB_EP_Desc *) kmem_cache_alloc(usb_desc_cache, mem_flags); + if(ep_desc == NULL) + return NULL; + memset(ep_desc, 0, sizeof(struct USB_EP_Desc)); + + ep_desc->hw_len = 0; + ep_desc->command = (IO_FIELD(USB_EP_command, epid, epid) | + IO_STATE(USB_EP_command, enable, yes)); + if(sb_desc == NULL) { + ep_desc->sub = 0; + } else { + ep_desc->sub = virt_to_phys(sb_desc); + } + return ep_desc; +} + +#define TT_ZOUT 0 +#define TT_IN 1 +#define TT_OUT 2 +#define TT_SETUP 3 + +#define CMD_EOL IO_STATE(USB_SB_command, eol, yes) +#define CMD_INTR IO_STATE(USB_SB_command, intr, yes) +#define CMD_FULL IO_STATE(USB_SB_command, full, yes) + +/* Allocation and setup of a generic SB. Used to create SETUP, OUT and ZOUT + SBs. Also used by create_sb_in() to avoid same allocation procedure at two + places */ +struct USB_SB_Desc* create_sb(struct USB_SB_Desc* sb_prev, int tt, void* data, + int datalen, int mem_flags) { + struct USB_SB_Desc *sb_desc; + sb_desc = (struct USB_SB_Desc*)kmem_cache_alloc(usb_desc_cache, mem_flags); + if(sb_desc == NULL) + return NULL; + memset(sb_desc, 0, sizeof(struct USB_SB_Desc)); + + sb_desc->command = IO_FIELD(USB_SB_command, tt, tt) | + IO_STATE(USB_SB_command, eot, yes); + + sb_desc->sw_len = datalen; + if(data != NULL) { + sb_desc->buf = virt_to_phys(data); + } else { + sb_desc->buf = 0; + } + if(sb_prev != NULL) { + sb_prev->next = virt_to_phys(sb_desc); + } + return sb_desc; +} + +/* Creates a copy of an existing SB by allocation space for it and copy + settings */ +struct USB_SB_Desc* create_sb_copy(struct USB_SB_Desc* sb_orig, int mem_flags) { + struct USB_SB_Desc *sb_desc; + sb_desc = (struct USB_SB_Desc*)kmem_cache_alloc(usb_desc_cache, mem_flags); + if(sb_desc == NULL) + return NULL; + + memcpy(sb_desc, sb_orig, sizeof(struct USB_SB_Desc)); + return sb_desc; +} + +/* A specific create_sb function for creation of in SBs. This is due to + that datalen in In SBs shows how many packets we are expecting. It also + sets up the rem field to show if how many bytes we expect in last packet + if it's not a full one */ +struct USB_SB_Desc* create_sb_in(struct USB_SB_Desc* sb_prev, int datalen, + int maxlen, int mem_flags) { + struct USB_SB_Desc *sb_desc; + sb_desc = create_sb(sb_prev, TT_IN, NULL, + datalen ? (datalen - 1) / maxlen + 1 : 0, mem_flags); + if(sb_desc == NULL) + return NULL; + sb_desc->command |= IO_FIELD(USB_SB_command, rem, datalen % maxlen); + return sb_desc; +} + +void set_sb_cmds(struct USB_SB_Desc *sb_desc, __u16 flags) { + sb_desc->command |= flags; +} + +int create_sb_for_urb(struct urb *urb, int mem_flags) { + int is_out = !usb_pipein(urb->pipe); + int type = usb_pipetype(urb->pipe); + int maxlen = usb_maxpacket(urb->dev, urb->pipe, is_out); + int buf_len = urb->transfer_buffer_length; + void *buf = buf_len > 0 ? urb->transfer_buffer : NULL; + struct USB_SB_Desc *sb_desc = NULL; + + struct crisv10_urb_priv *urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv != NULL); + + switch(type) { + case PIPE_CONTROL: + /* Setup stage */ + sb_desc = create_sb(NULL, TT_SETUP, urb->setup_packet, 8, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + set_sb_cmds(sb_desc, CMD_FULL); + + /* Attach first SB to URB */ + urb_priv->first_sb = sb_desc; + + if (is_out) { /* Out Control URB */ + /* If this Control OUT transfer has an optional data stage we add + an OUT token before the mandatory IN (status) token */ + if ((buf_len > 0) && buf) { + sb_desc = create_sb(sb_desc, TT_OUT, buf, buf_len, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + set_sb_cmds(sb_desc, CMD_FULL); + } + + /* Status stage */ + /* The data length has to be exactly 1. This is due to a requirement + of the USB specification that a host must be prepared to receive + data in the status phase */ + sb_desc = create_sb(sb_desc, TT_IN, NULL, 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + } else { /* In control URB */ + /* Data stage */ + sb_desc = create_sb_in(sb_desc, buf_len, maxlen, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + + /* Status stage */ + /* Read comment at zout_buffer declaration for an explanation to this. */ + sb_desc = create_sb(sb_desc, TT_ZOUT, &zout_buffer[0], 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + /* Set descriptor interrupt flag for in URBs so we can finish URB after + zout-packet has been sent */ + set_sb_cmds(sb_desc, CMD_INTR | CMD_FULL); + } + /* Set end-of-list flag in last SB */ + set_sb_cmds(sb_desc, CMD_EOL); + /* Attach last SB to URB */ + urb_priv->last_sb = sb_desc; + break; + + case PIPE_BULK: + if (is_out) { /* Out Bulk URB */ + sb_desc = create_sb(NULL, TT_OUT, buf, buf_len, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + /* The full field is set to yes, even if we don't actually check that + this is a full-length transfer (i.e., that transfer_buffer_length % + maxlen = 0). + Setting full prevents the USB controller from sending an empty packet + in that case. However, if URB_ZERO_PACKET was set we want that. */ + if (!(urb->transfer_flags & URB_ZERO_PACKET)) { + set_sb_cmds(sb_desc, CMD_FULL); + } + } else { /* In Bulk URB */ + sb_desc = create_sb_in(NULL, buf_len, maxlen, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + } + /* Set end-of-list flag for last SB */ + set_sb_cmds(sb_desc, CMD_EOL); + + /* Attach SB to URB */ + urb_priv->first_sb = sb_desc; + urb_priv->last_sb = sb_desc; + break; + + case PIPE_INTERRUPT: + if(is_out) { /* Out Intr URB */ + sb_desc = create_sb(NULL, TT_OUT, buf, buf_len, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + + /* The full field is set to yes, even if we don't actually check that + this is a full-length transfer (i.e., that transfer_buffer_length % + maxlen = 0). + Setting full prevents the USB controller from sending an empty packet + in that case. However, if URB_ZERO_PACKET was set we want that. */ + if (!(urb->transfer_flags & URB_ZERO_PACKET)) { + set_sb_cmds(sb_desc, CMD_FULL); + } + /* Only generate TX interrupt if it's a Out URB*/ + set_sb_cmds(sb_desc, CMD_INTR); + + } else { /* In Intr URB */ + sb_desc = create_sb_in(NULL, buf_len, maxlen, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + } + /* Set end-of-list flag for last SB */ + set_sb_cmds(sb_desc, CMD_EOL); + + /* Attach SB to URB */ + urb_priv->first_sb = sb_desc; + urb_priv->last_sb = sb_desc; + + break; + case PIPE_ISOCHRONOUS: + if(is_out) { /* Out Isoc URB */ + int i; + if(urb->number_of_packets == 0) { + tc_err("Can't create SBs for Isoc URB with zero packets\n"); + return -EPIPE; + } + /* Create one SB descriptor for each packet and link them together. */ + for(i = 0; i < urb->number_of_packets; i++) { + if (urb->iso_frame_desc[i].length > 0) { + + sb_desc = create_sb(sb_desc, TT_OUT, urb->transfer_buffer + + urb->iso_frame_desc[i].offset, + urb->iso_frame_desc[i].length, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + + /* Check if it's a full length packet */ + if (urb->iso_frame_desc[i].length == + usb_maxpacket(urb->dev, urb->pipe, usb_pipeout(urb->pipe))) { + set_sb_cmds(sb_desc, CMD_FULL); + } + + } else { /* zero length packet */ + sb_desc = create_sb(sb_desc, TT_ZOUT, &zout_buffer[0], 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + set_sb_cmds(sb_desc, CMD_FULL); + } + /* Attach first SB descriptor to URB */ + if (i == 0) { + urb_priv->first_sb = sb_desc; + } + } + /* Set interrupt and end-of-list flags in last SB */ + set_sb_cmds(sb_desc, CMD_INTR | CMD_EOL); + /* Attach last SB descriptor to URB */ + urb_priv->last_sb = sb_desc; + tc_dbg("Created %d out SBs for Isoc URB:0x%x\n", + urb->number_of_packets, (unsigned int)urb); + } else { /* In Isoc URB */ + /* Actual number of packets is not relevant for periodic in traffic as + long as it is more than zero. Set to 1 always. */ + sb_desc = create_sb(sb_desc, TT_IN, NULL, 1, mem_flags); + if(sb_desc == NULL) + return -ENOMEM; + /* Set end-of-list flags for SB */ + set_sb_cmds(sb_desc, CMD_EOL); + + /* Attach SB to URB */ + urb_priv->first_sb = sb_desc; + urb_priv->last_sb = sb_desc; + } + break; + default: + tc_err("Unknown pipe-type\n"); + return -EPIPE; + break; + } + return 0; +} + +int init_intr_urb(struct urb *urb, int mem_flags) { + struct crisv10_urb_priv *urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + struct USB_EP_Desc* ep_desc; + int interval; + int i; + int ep_count; + + ASSERT(urb_priv != NULL); + ASSERT(usb_pipeint(urb->pipe)); + /* We can't support interval longer than amount of eof descriptors in + TxIntrEPList */ + if(urb->interval > MAX_INTR_INTERVAL) { + tc_err("Interrupt interval %dms too big (max: %dms)\n", urb->interval, + MAX_INTR_INTERVAL); + return -EINVAL; + } + + /* We assume that the SB descriptors already have been setup */ + ASSERT(urb_priv->first_sb != NULL); + + /* Round of the interval to 2^n, it is obvious that this code favours + smaller numbers, but that is actually a good thing */ + /* FIXME: The "rounding error" for larger intervals will be quite + large. For in traffic this shouldn't be a problem since it will only + mean that we "poll" more often. */ + interval = urb->interval; + for (i = 0; interval; i++) { + interval = interval >> 1; + } + urb_priv->interval = 1 << (i - 1); + + /* We can only have max interval for Out Interrupt due to that we can only + handle one linked in EP for a certain epid in the Intr descr array at the + time. The USB Controller in the Etrax 100LX continues to process Intr EPs + so we have no way of knowing which one that caused the actual transfer if + we have several linked in. */ + if(usb_pipeout(urb->pipe)) { + urb_priv->interval = MAX_INTR_INTERVAL; + } + + /* Calculate amount of EPs needed */ + ep_count = MAX_INTR_INTERVAL / urb_priv->interval; + + for(i = 0; i < ep_count; i++) { + ep_desc = create_ep(urb_priv->epid, urb_priv->first_sb, mem_flags); + if(ep_desc == NULL) { + /* Free any descriptors that we may have allocated before failure */ + while(i > 0) { + i--; + kfree(urb_priv->intr_ep_pool[i]); + } + return -ENOMEM; + } + urb_priv->intr_ep_pool[i] = ep_desc; + } + urb_priv->intr_ep_pool_length = ep_count; + return 0; +} + +/* DMA RX/TX functions */ +/* ----------------------- */ + +static void tc_dma_init_rx_list(void) { + int i; + + /* Setup descriptor list except last one */ + for (i = 0; i < (NBR_OF_RX_DESC - 1); i++) { + RxDescList[i].sw_len = RX_DESC_BUF_SIZE; + RxDescList[i].command = 0; + RxDescList[i].next = virt_to_phys(&RxDescList[i + 1]); + RxDescList[i].buf = virt_to_phys(RxBuf + (i * RX_DESC_BUF_SIZE)); + RxDescList[i].hw_len = 0; + RxDescList[i].status = 0; + + /* DMA IN cache bug. (struct etrax_dma_descr has the same layout as + USB_IN_Desc for the relevant fields.) */ + prepare_rx_descriptor((struct etrax_dma_descr*)&RxDescList[i]); + + } + /* Special handling of last descriptor */ + RxDescList[i].sw_len = RX_DESC_BUF_SIZE; + RxDescList[i].command = IO_STATE(USB_IN_command, eol, yes); + RxDescList[i].next = virt_to_phys(&RxDescList[0]); + RxDescList[i].buf = virt_to_phys(RxBuf + (i * RX_DESC_BUF_SIZE)); + RxDescList[i].hw_len = 0; + RxDescList[i].status = 0; + + /* Setup list pointers that show progress in list */ + myNextRxDesc = &RxDescList[0]; + myLastRxDesc = &RxDescList[NBR_OF_RX_DESC - 1]; + + flush_etrax_cache(); + /* Point DMA to first descriptor in list and start it */ + *R_DMA_CH9_FIRST = virt_to_phys(myNextRxDesc); + *R_DMA_CH9_CMD = IO_STATE(R_DMA_CH9_CMD, cmd, start); +} + + +static void tc_dma_init_tx_bulk_list(void) { + int i; + volatile struct USB_EP_Desc *epDescr; + + for (i = 0; i < (NBR_OF_EPIDS - 1); i++) { + epDescr = &(TxBulkEPList[i]); + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxBulkEPList[i + 1]); + + /* Initiate two EPs, disabled and with the eol flag set. No need for any + preserved epid. */ + + /* The first one has the intr flag set so we get an interrupt when the DMA + channel is about to become disabled. */ + CHECK_ALIGN(&TxBulkDummyEPList[i][0]); + TxBulkDummyEPList[i][0].hw_len = 0; + TxBulkDummyEPList[i][0].command = (IO_FIELD(USB_EP_command, epid, DUMMY_EPID) | + IO_STATE(USB_EP_command, eol, yes) | + IO_STATE(USB_EP_command, intr, yes)); + TxBulkDummyEPList[i][0].sub = 0; + TxBulkDummyEPList[i][0].next = virt_to_phys(&TxBulkDummyEPList[i][1]); + + /* The second one. */ + CHECK_ALIGN(&TxBulkDummyEPList[i][1]); + TxBulkDummyEPList[i][1].hw_len = 0; + TxBulkDummyEPList[i][1].command = (IO_FIELD(USB_EP_command, epid, DUMMY_EPID) | + IO_STATE(USB_EP_command, eol, yes)); + TxBulkDummyEPList[i][1].sub = 0; + /* The last dummy's next pointer is the same as the current EP's next pointer. */ + TxBulkDummyEPList[i][1].next = virt_to_phys(&TxBulkEPList[i + 1]); + } + + /* Special handling of last descr in list, make list circular */ + epDescr = &TxBulkEPList[i]; + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_STATE(USB_EP_command, eol, yes) | + IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxBulkEPList[0]); + + /* Init DMA sub-channel pointers to last item in each list */ + *R_DMA_CH8_SUB0_EP = virt_to_phys(&TxBulkEPList[i]); + /* No point in starting the bulk channel yet. + *R_DMA_CH8_SUB0_CMD = IO_STATE(R_DMA_CH8_SUB0_CMD, cmd, start); */ +} + +static void tc_dma_init_tx_ctrl_list(void) { + int i; + volatile struct USB_EP_Desc *epDescr; + + for (i = 0; i < (NBR_OF_EPIDS - 1); i++) { + epDescr = &(TxCtrlEPList[i]); + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxCtrlEPList[i + 1]); + } + /* Special handling of last descr in list, make list circular */ + epDescr = &TxCtrlEPList[i]; + CHECK_ALIGN(epDescr); + epDescr->hw_len = 0; + epDescr->command = IO_STATE(USB_EP_command, eol, yes) | + IO_FIELD(USB_EP_command, epid, i); + epDescr->sub = 0; + epDescr->next = virt_to_phys(&TxCtrlEPList[0]); + + /* Init DMA sub-channel pointers to last item in each list */ + *R_DMA_CH8_SUB1_EP = virt_to_phys(&TxCtrlEPList[i]); + /* No point in starting the ctrl channel yet. + *R_DMA_CH8_SUB1_CMD = IO_STATE(R_DMA_CH8_SUB0_CMD, cmd, start); */ +} + + +static void tc_dma_init_tx_intr_list(void) { + int i; + + TxIntrSB_zout.sw_len = 1; + TxIntrSB_zout.next = 0; + TxIntrSB_zout.buf = virt_to_phys(&zout_buffer[0]); + TxIntrSB_zout.command = (IO_FIELD(USB_SB_command, rem, 0) | + IO_STATE(USB_SB_command, tt, zout) | + IO_STATE(USB_SB_command, full, yes) | + IO_STATE(USB_SB_command, eot, yes) | + IO_STATE(USB_SB_command, eol, yes)); + + for (i = 0; i < (MAX_INTR_INTERVAL - 1); i++) { + CHECK_ALIGN(&TxIntrEPList[i]); + TxIntrEPList[i].hw_len = 0; + TxIntrEPList[i].command = + (IO_STATE(USB_EP_command, eof, yes) | + IO_STATE(USB_EP_command, enable, yes) | + IO_FIELD(USB_EP_command, epid, INVALID_EPID)); + TxIntrEPList[i].sub = virt_to_phys(&TxIntrSB_zout); + TxIntrEPList[i].next = virt_to_phys(&TxIntrEPList[i + 1]); + } + + /* Special handling of last descr in list, make list circular */ + CHECK_ALIGN(&TxIntrEPList[i]); + TxIntrEPList[i].hw_len = 0; + TxIntrEPList[i].command = + (IO_STATE(USB_EP_command, eof, yes) | + IO_STATE(USB_EP_command, eol, yes) | + IO_STATE(USB_EP_command, enable, yes) | + IO_FIELD(USB_EP_command, epid, INVALID_EPID)); + TxIntrEPList[i].sub = virt_to_phys(&TxIntrSB_zout); + TxIntrEPList[i].next = virt_to_phys(&TxIntrEPList[0]); + + intr_dbg("Initiated Intr EP descriptor list\n"); + + + /* Connect DMA 8 sub-channel 2 to first in list */ + *R_DMA_CH8_SUB2_EP = virt_to_phys(&TxIntrEPList[0]); +} + +static void tc_dma_init_tx_isoc_list(void) { + int i; + + DBFENTER; + + /* Read comment at zout_buffer declaration for an explanation to this. */ + TxIsocSB_zout.sw_len = 1; + TxIsocSB_zout.next = 0; + TxIsocSB_zout.buf = virt_to_phys(&zout_buffer[0]); + TxIsocSB_zout.command = (IO_FIELD(USB_SB_command, rem, 0) | + IO_STATE(USB_SB_command, tt, zout) | + IO_STATE(USB_SB_command, full, yes) | + IO_STATE(USB_SB_command, eot, yes) | + IO_STATE(USB_SB_command, eol, yes)); + + /* The last isochronous EP descriptor is a dummy. */ + for (i = 0; i < (NBR_OF_EPIDS - 1); i++) { + CHECK_ALIGN(&TxIsocEPList[i]); + TxIsocEPList[i].hw_len = 0; + TxIsocEPList[i].command = IO_FIELD(USB_EP_command, epid, i); + TxIsocEPList[i].sub = 0; + TxIsocEPList[i].next = virt_to_phys(&TxIsocEPList[i + 1]); + } + + CHECK_ALIGN(&TxIsocEPList[i]); + TxIsocEPList[i].hw_len = 0; + + /* Must enable the last EP descr to get eof interrupt. */ + TxIsocEPList[i].command = (IO_STATE(USB_EP_command, enable, yes) | + IO_STATE(USB_EP_command, eof, yes) | + IO_STATE(USB_EP_command, eol, yes) | + IO_FIELD(USB_EP_command, epid, INVALID_EPID)); + TxIsocEPList[i].sub = virt_to_phys(&TxIsocSB_zout); + TxIsocEPList[i].next = virt_to_phys(&TxIsocEPList[0]); + + *R_DMA_CH8_SUB3_EP = virt_to_phys(&TxIsocEPList[0]); + *R_DMA_CH8_SUB3_CMD = IO_STATE(R_DMA_CH8_SUB3_CMD, cmd, start); +} + +static int tc_dma_init(struct usb_hcd *hcd) { + tc_dma_init_rx_list(); + tc_dma_init_tx_bulk_list(); + tc_dma_init_tx_ctrl_list(); + tc_dma_init_tx_intr_list(); + tc_dma_init_tx_isoc_list(); + + if (cris_request_dma(USB_TX_DMA_NBR, + "ETRAX 100LX built-in USB (Tx)", + DMA_VERBOSE_ON_ERROR, + dma_usb)) { + err("Could not allocate DMA ch 8 for USB"); + return -EBUSY; + } + + if (cris_request_dma(USB_RX_DMA_NBR, + "ETRAX 100LX built-in USB (Rx)", + DMA_VERBOSE_ON_ERROR, + dma_usb)) { + err("Could not allocate DMA ch 9 for USB"); + return -EBUSY; + } + + *R_IRQ_MASK2_SET = + /* Note that these interrupts are not used. */ + IO_STATE(R_IRQ_MASK2_SET, dma8_sub0_descr, set) | + /* Sub channel 1 (ctrl) descr. interrupts are used. */ + IO_STATE(R_IRQ_MASK2_SET, dma8_sub1_descr, set) | + IO_STATE(R_IRQ_MASK2_SET, dma8_sub2_descr, set) | + /* Sub channel 3 (isoc) descr. interrupts are used. */ + IO_STATE(R_IRQ_MASK2_SET, dma8_sub3_descr, set); + + /* Note that the dma9_descr interrupt is not used. */ + *R_IRQ_MASK2_SET = + IO_STATE(R_IRQ_MASK2_SET, dma9_eop, set) | + IO_STATE(R_IRQ_MASK2_SET, dma9_descr, set); + + if (request_irq(ETRAX_USB_RX_IRQ, tc_dma_rx_interrupt, 0, + "ETRAX 100LX built-in USB (Rx)", hcd)) { + err("Could not allocate IRQ %d for USB", ETRAX_USB_RX_IRQ); + return -EBUSY; + } + + if (request_irq(ETRAX_USB_TX_IRQ, tc_dma_tx_interrupt, 0, + "ETRAX 100LX built-in USB (Tx)", hcd)) { + err("Could not allocate IRQ %d for USB", ETRAX_USB_TX_IRQ); + return -EBUSY; + } + + return 0; +} + +static void tc_dma_destroy(void) { + free_irq(ETRAX_USB_RX_IRQ, NULL); + free_irq(ETRAX_USB_TX_IRQ, NULL); + + cris_free_dma(USB_TX_DMA_NBR, "ETRAX 100LX built-in USB (Tx)"); + cris_free_dma(USB_RX_DMA_NBR, "ETRAX 100LX built-in USB (Rx)"); + +} + +static void tc_dma_link_intr_urb(struct urb *urb); + +/* Handle processing of Bulk, Ctrl and Intr queues */ +static void tc_dma_process_queue(int epid) { + struct urb *urb; + struct crisv10_urb_priv *urb_priv; + unsigned long flags; + char toggle; + + if(epid_state[epid].disabled) { + /* Don't process any URBs on a disabled endpoint */ + return; + } + + /* Do not disturb us while fiddling with EPs and epids */ + local_irq_save(flags); + + /* For bulk, Ctrl and Intr can we only have one URB active at a time for + a specific EP. */ + if(activeUrbList[epid] != NULL) { + /* An URB is already active on EP, skip checking queue */ + local_irq_restore(flags); + return; + } + + urb = urb_list_first(epid); + if(urb == NULL) { + /* No URB waiting in EP queue. Nothing do to */ + local_irq_restore(flags); + return; + } + + urb_priv = urb->hcpriv; + ASSERT(urb_priv != NULL); + ASSERT(urb_priv->urb_state == NOT_STARTED); + ASSERT(!usb_pipeisoc(urb->pipe)); + + /* Remove this URB from the queue and move it to active */ + activeUrbList[epid] = urb; + urb_list_del(urb, epid); + + urb_priv->urb_state = STARTED; + + /* Reset error counters (regardless of which direction this traffic is). */ + etrax_epid_clear_error(epid); + + /* Special handling of Intr EP lists */ + if(usb_pipeint(urb->pipe)) { + tc_dma_link_intr_urb(urb); + local_irq_restore(flags); + return; + } + + /* Software must preset the toggle bits for Bulk and Ctrl */ + if(usb_pipecontrol(urb->pipe)) { + /* Toggle bits are initialized only during setup transaction in a + CTRL transfer */ + etrax_epid_set_toggle(epid, 0, 0); + etrax_epid_set_toggle(epid, 1, 0); + } else { + toggle = usb_gettoggle(urb->dev, usb_pipeendpoint(urb->pipe), + usb_pipeout(urb->pipe)); + etrax_epid_set_toggle(epid, usb_pipeout(urb->pipe), toggle); + } + + tc_dbg("Added SBs from (URB:0x%x %s %s) to epid %d: %s\n", + (unsigned int)urb, str_dir(urb->pipe), str_type(urb->pipe), epid, + sblist_to_str(urb_priv->first_sb)); + + /* We start the DMA sub channel without checking if it's running or not, + because: + 1) If it's already running, issuing the start command is a nop. + 2) We avoid a test-and-set race condition. */ + switch(usb_pipetype(urb->pipe)) { + case PIPE_BULK: + /* Assert that the EP descriptor is disabled. */ + ASSERT(!(TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable))); + + /* Set up and enable the EP descriptor. */ + TxBulkEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + TxBulkEPList[epid].hw_len = 0; + TxBulkEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + + /* Check if the dummy list is already with us (if several urbs were queued). */ + if (usb_pipein(urb->pipe) && (TxBulkEPList[epid].next != virt_to_phys(&TxBulkDummyEPList[epid][0]))) { + tc_dbg("Inviting dummy list to the party for urb 0x%lx, epid %d", + (unsigned long)urb, epid); + + /* We don't need to check if the DMA is at this EP or not before changing the + next pointer, since we will do it in one 32-bit write (EP descriptors are + 32-bit aligned). */ + TxBulkEPList[epid].next = virt_to_phys(&TxBulkDummyEPList[epid][0]); + } + + restart_dma8_sub0(); + + /* Update/restart the bulk start timer since we just started the channel.*/ + mod_timer(&bulk_start_timer, jiffies + BULK_START_TIMER_INTERVAL); + /* Update/restart the bulk eot timer since we just inserted traffic. */ + mod_timer(&bulk_eot_timer, jiffies + BULK_EOT_TIMER_INTERVAL); + break; + case PIPE_CONTROL: + /* Assert that the EP descriptor is disabled. */ + ASSERT(!(TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable))); + + /* Set up and enable the EP descriptor. */ + TxCtrlEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + TxCtrlEPList[epid].hw_len = 0; + TxCtrlEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + + *R_DMA_CH8_SUB1_CMD = IO_STATE(R_DMA_CH8_SUB1_CMD, cmd, start); + break; + } + local_irq_restore(flags); +} + +static void tc_dma_link_intr_urb(struct urb *urb) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + volatile struct USB_EP_Desc *tmp_ep; + struct USB_EP_Desc *ep_desc; + int i = 0, epid; + int pool_idx = 0; + + ASSERT(urb_priv != NULL); + epid = urb_priv->epid; + ASSERT(urb_priv->interval > 0); + ASSERT(urb_priv->intr_ep_pool_length > 0); + + tmp_ep = &TxIntrEPList[0]; + + /* Only insert one EP descriptor in list for Out Intr URBs. + We can only handle Out Intr with interval of 128ms because + it's not possible to insert several Out Intr EPs because they + are not consumed by the DMA. */ + if(usb_pipeout(urb->pipe)) { + ep_desc = urb_priv->intr_ep_pool[0]; + ASSERT(ep_desc); + ep_desc->next = tmp_ep->next; + tmp_ep->next = virt_to_phys(ep_desc); + i++; + } else { + /* Loop through Intr EP descriptor list and insert EP for URB at + specified interval */ + do { + /* Each EP descriptor with eof flag sat signals a new frame */ + if (tmp_ep->command & IO_MASK(USB_EP_command, eof)) { + /* Insert a EP from URBs EP pool at correct interval */ + if ((i % urb_priv->interval) == 0) { + ep_desc = urb_priv->intr_ep_pool[pool_idx]; + ASSERT(ep_desc); + ep_desc->next = tmp_ep->next; + tmp_ep->next = virt_to_phys(ep_desc); + pool_idx++; + ASSERT(pool_idx <= urb_priv->intr_ep_pool_length); + } + i++; + } + tmp_ep = (struct USB_EP_Desc *)phys_to_virt(tmp_ep->next); + } while(tmp_ep != &TxIntrEPList[0]); + } + + intr_dbg("Added SBs to intr epid %d: %s interval:%d (%d EP)\n", epid, + sblist_to_str(urb_priv->first_sb), urb_priv->interval, pool_idx); + + /* We start the DMA sub channel without checking if it's running or not, + because: + 1) If it's already running, issuing the start command is a nop. + 2) We avoid a test-and-set race condition. */ + *R_DMA_CH8_SUB2_CMD = IO_STATE(R_DMA_CH8_SUB2_CMD, cmd, start); +} + +static void tc_dma_process_isoc_urb(struct urb *urb) { + unsigned long flags; + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + int epid; + + /* Do not disturb us while fiddling with EPs and epids */ + local_irq_save(flags); + + ASSERT(urb_priv); + ASSERT(urb_priv->first_sb); + epid = urb_priv->epid; + + if(activeUrbList[epid] == NULL) { + /* EP is idle, so make this URB active */ + activeUrbList[epid] = urb; + urb_list_del(urb, epid); + ASSERT(TxIsocEPList[epid].sub == 0); + ASSERT(!(TxIsocEPList[epid].command & + IO_STATE(USB_EP_command, enable, yes))); + + /* Differentiate between In and Out Isoc. Because In SBs are not consumed*/ + if(usb_pipein(urb->pipe)) { + /* Each EP for In Isoc will have only one SB descriptor, setup when + submitting the first active urb. We do it here by copying from URBs + pre-allocated SB. */ + memcpy((void *)&(TxIsocSBList[epid]), urb_priv->first_sb, + sizeof(TxIsocSBList[epid])); + TxIsocEPList[epid].hw_len = 0; + TxIsocEPList[epid].sub = virt_to_phys(&(TxIsocSBList[epid])); + } else { + /* For Out Isoc we attach the pre-allocated list of SBs for the URB */ + TxIsocEPList[epid].hw_len = 0; + TxIsocEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + + isoc_dbg("Attached first URB:0x%x[%d] to epid:%d first_sb:0x%x" + " last_sb::0x%x\n", + (unsigned int)urb, urb_priv->urb_num, epid, + (unsigned int)(urb_priv->first_sb), + (unsigned int)(urb_priv->last_sb)); + } + + if (urb->transfer_flags & URB_ISO_ASAP) { + /* The isoc transfer should be started as soon as possible. The + start_frame field is a return value if URB_ISO_ASAP was set. Comparing + R_USB_FM_NUMBER with a USB Chief trace shows that the first isoc IN + token is sent 2 frames later. I'm not sure how this affects usage of + the start_frame field by the device driver, or how it affects things + when USB_ISO_ASAP is not set, so therefore there's no compensation for + the 2 frame "lag" here. */ + urb->start_frame = (*R_USB_FM_NUMBER & 0x7ff); + TxIsocEPList[epid].command |= IO_STATE(USB_EP_command, enable, yes); + urb_priv->urb_state = STARTED; + isoc_dbg("URB_ISO_ASAP set, urb->start_frame set to %d\n", + urb->start_frame); + } else { + /* Not started yet. */ + urb_priv->urb_state = NOT_STARTED; + isoc_warn("urb_priv->urb_state set to NOT_STARTED for URB:0x%x\n", + (unsigned int)urb); + } + + } else { + /* An URB is already active on the EP. Leave URB in queue and let + finish_isoc_urb process it after current active URB */ + ASSERT(TxIsocEPList[epid].sub != 0); + + if(usb_pipein(urb->pipe)) { + /* Because there already is a active In URB on this epid we do nothing + and the finish_isoc_urb() function will handle switching to next URB*/ + + } else { /* For Out Isoc, insert new URBs traffic last in SB-list. */ + struct USB_SB_Desc *temp_sb_desc; + + /* Set state STARTED to all Out Isoc URBs added to SB list because we + don't know how many of them that are finished before descr interrupt*/ + urb_priv->urb_state = STARTED; + + /* Find end of current SB list by looking for SB with eol flag sat */ + temp_sb_desc = phys_to_virt(TxIsocEPList[epid].sub); + while ((temp_sb_desc->command & IO_MASK(USB_SB_command, eol)) != + IO_STATE(USB_SB_command, eol, yes)) { + ASSERT(temp_sb_desc->next); + temp_sb_desc = phys_to_virt(temp_sb_desc->next); + } + + isoc_dbg("Appended URB:0x%x[%d] (first:0x%x last:0x%x) to epid:%d" + " sub:0x%x eol:0x%x\n", + (unsigned int)urb, urb_priv->urb_num, + (unsigned int)(urb_priv->first_sb), + (unsigned int)(urb_priv->last_sb), epid, + (unsigned int)phys_to_virt(TxIsocEPList[epid].sub), + (unsigned int)temp_sb_desc); + + /* Next pointer must be set before eol is removed. */ + temp_sb_desc->next = virt_to_phys(urb_priv->first_sb); + /* Clear the previous end of list flag since there is a new in the + added SB descriptor list. */ + temp_sb_desc->command &= ~IO_MASK(USB_SB_command, eol); + + if (!(TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable))) { + __u32 epid_data; + /* 8.8.5 in Designer's Reference says we should check for and correct + any errors in the EP here. That should not be necessary if + epid_attn is handled correctly, so we assume all is ok. */ + epid_data = etrax_epid_iso_get(epid); + if (IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data) != + IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + isoc_err("Disabled Isoc EP with error:%d on epid:%d when appending" + " URB:0x%x[%d]\n", + IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data), epid, + (unsigned int)urb, urb_priv->urb_num); + } + + /* The SB list was exhausted. */ + if (virt_to_phys(urb_priv->last_sb) != TxIsocEPList[epid].sub) { + /* The new sublist did not get processed before the EP was + disabled. Setup the EP again. */ + + if(virt_to_phys(temp_sb_desc) == TxIsocEPList[epid].sub) { + isoc_dbg("EP for epid:%d stoped at SB:0x%x before newly inserted" + ", restarting from this URBs SB:0x%x\n", + epid, (unsigned int)temp_sb_desc, + (unsigned int)(urb_priv->first_sb)); + TxIsocEPList[epid].hw_len = 0; + TxIsocEPList[epid].sub = virt_to_phys(urb_priv->first_sb); + urb->start_frame = (*R_USB_FM_NUMBER & 0x7ff); + /* Enable the EP again so data gets processed this time */ + TxIsocEPList[epid].command |= + IO_STATE(USB_EP_command, enable, yes); + + } else { + /* The EP has been disabled but not at end this URB (god knows + where). This should generate an epid_attn so we should not be + here */ + isoc_warn("EP was disabled on sb:0x%x before SB list for" + " URB:0x%x[%d] got processed\n", + (unsigned int)phys_to_virt(TxIsocEPList[epid].sub), + (unsigned int)urb, urb_priv->urb_num); + } + } else { + /* This might happend if we are slow on this function and isn't + an error. */ + isoc_dbg("EP was disabled and finished with SBs from appended" + " URB:0x%x[%d]\n", (unsigned int)urb, urb_priv->urb_num); + } + } + } + } + + /* Start the DMA sub channel */ + *R_DMA_CH8_SUB3_CMD = IO_STATE(R_DMA_CH8_SUB3_CMD, cmd, start); + + local_irq_restore(flags); +} + +static void tc_dma_unlink_intr_urb(struct urb *urb) { + struct crisv10_urb_priv *urb_priv = urb->hcpriv; + volatile struct USB_EP_Desc *first_ep; /* First EP in the list. */ + volatile struct USB_EP_Desc *curr_ep; /* Current EP, the iterator. */ + volatile struct USB_EP_Desc *next_ep; /* The EP after current. */ + volatile struct USB_EP_Desc *unlink_ep; /* The one we should remove from + the list. */ + int count = 0; + volatile int timeout = 10000; + int epid; + + /* Read 8.8.4 in Designer's Reference, "Removing an EP Descriptor from the + List". */ + ASSERT(urb_priv); + ASSERT(urb_priv->intr_ep_pool_length > 0); + epid = urb_priv->epid; + + /* First disable all Intr EPs belonging to epid for this URB */ + first_ep = &TxIntrEPList[0]; + curr_ep = first_ep; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if (IO_EXTRACT(USB_EP_command, epid, next_ep->command) == epid) { + /* Disable EP */ + next_ep->command &= ~IO_MASK(USB_EP_command, enable); + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != first_ep); + + + /* Now unlink all EPs belonging to this epid from Descr list */ + first_ep = &TxIntrEPList[0]; + curr_ep = first_ep; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if (IO_EXTRACT(USB_EP_command, epid, next_ep->command) == epid) { + /* This is the one we should unlink. */ + unlink_ep = next_ep; + + /* Actually unlink the EP from the DMA list. */ + curr_ep->next = unlink_ep->next; + + /* Wait until the DMA is no longer at this descriptor. */ + while((*R_DMA_CH8_SUB2_EP == virt_to_phys(unlink_ep)) && + (timeout-- > 0)); + + count++; + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != first_ep); + + if(count != urb_priv->intr_ep_pool_length) { + intr_warn("Unlinked %d of %d Intr EPs for URB:0x%x[%d]\n", count, + urb_priv->intr_ep_pool_length, (unsigned int)urb, + urb_priv->urb_num); + } else { + intr_dbg("Unlinked %d of %d interrupt EPs for URB:0x%x\n", count, + urb_priv->intr_ep_pool_length, (unsigned int)urb); + } +} + +static void check_finished_bulk_tx_epids(struct usb_hcd *hcd, + int timer) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + __u32 epid_data; + + /* Protect TxEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + /* A finished EP descriptor is disabled and has a valid sub pointer */ + if (!(TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) && + (TxBulkEPList[epid].sub != 0)) { + + /* Get the active URB for this epid */ + urb = activeUrbList[epid]; + /* Sanity checks */ + ASSERT(urb); + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + /* Only handle finished out Bulk EPs here, + and let RX interrupt take care of the rest */ + if(!epid_out_traffic(epid)) { + continue; + } + + if(timer) { + tc_warn("Found finished %s Bulk epid:%d URB:0x%x[%d] from timeout\n", + epid_out_traffic(epid) ? "Out" : "In", epid, (unsigned int)urb, + urb_priv->urb_num); + } else { + tc_dbg("Found finished %s Bulk epid:%d URB:0x%x[%d] from interrupt\n", + epid_out_traffic(epid) ? "Out" : "In", epid, (unsigned int)urb, + urb_priv->urb_num); + } + + if(urb_priv->urb_state == UNLINK) { + /* This Bulk URB is requested to be unlinked, that means that the EP + has been disabled and we might not have sent all data */ + tc_finish_urb(hcd, urb, urb->status); + continue; + } + + ASSERT(urb_priv->urb_state == STARTED); + if (phys_to_virt(TxBulkEPList[epid].sub) != urb_priv->last_sb) { + tc_err("Endpoint got disabled before reaching last sb\n"); + } + + epid_data = etrax_epid_get(epid); + if (IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data) == + IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + /* This means that the endpoint has no error, is disabled + and had inserted traffic, i.e. transfer successfully completed. */ + tc_finish_urb(hcd, urb, 0); + } else { + /* Shouldn't happen. We expect errors to be caught by epid + attention. */ + tc_err("Found disabled bulk EP desc (epid:%d error:%d)\n", + epid, IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data)); + } + } else { + tc_dbg("Ignoring In Bulk epid:%d, let RX interrupt handle it\n", epid); + } + } + + local_irq_restore(flags); +} + +static void check_finished_ctrl_tx_epids(struct usb_hcd *hcd) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + __u32 epid_data; + + /* Protect TxEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if(epid == DUMMY_EPID) + continue; + + /* A finished EP descriptor is disabled and has a valid sub pointer */ + if (!(TxCtrlEPList[epid].command & IO_MASK(USB_EP_command, enable)) && + (TxCtrlEPList[epid].sub != 0)) { + + /* Get the active URB for this epid */ + urb = activeUrbList[epid]; + + if(urb == NULL) { + tc_warn("Found finished Ctrl epid:%d with no active URB\n", epid); + continue; + } + + /* Sanity checks */ + ASSERT(usb_pipein(urb->pipe)); + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + if (phys_to_virt(TxCtrlEPList[epid].sub) != urb_priv->last_sb) { + tc_err("Endpoint got disabled before reaching last sb\n"); + } + + epid_data = etrax_epid_get(epid); + if (IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data) == + IO_STATE_VALUE(R_USB_EPT_DATA, error_code, no_error)) { + /* This means that the endpoint has no error, is disabled + and had inserted traffic, i.e. transfer successfully completed. */ + + /* Check if RX-interrupt for In Ctrl has been processed before + finishing the URB */ + if(urb_priv->ctrl_rx_done) { + tc_dbg("Finishing In Ctrl URB:0x%x[%d] in tx_interrupt\n", + (unsigned int)urb, urb_priv->urb_num); + tc_finish_urb(hcd, urb, 0); + } else { + /* If we get zout descriptor interrupt before RX was done for a + In Ctrl transfer, then we flag that and it will be finished + in the RX-Interrupt */ + urb_priv->ctrl_zout_done = 1; + tc_dbg("Got zout descr interrupt before RX interrupt\n"); + } + } else { + /* Shouldn't happen. We expect errors to be caught by epid + attention. */ + tc_err("Found disabled Ctrl EP desc (epid:%d URB:0x%x[%d]) error_code:%d\n", epid, (unsigned int)urb, urb_priv->urb_num, IO_EXTRACT(R_USB_EPT_DATA, error_code, epid_data)); + __dump_ep_desc(&(TxCtrlEPList[epid])); + __dump_ept_data(epid); + } + } + } + local_irq_restore(flags); +} + +/* This function goes through all epids that are setup for Out Isoc transfers + and marks (isoc_out_done) all queued URBs that the DMA has finished + transfer for. + No URB completetion is done here to make interrupt routine return quickly. + URBs are completed later with help of complete_isoc_bottom_half() that + becomes schedules when this functions is finished. */ +static void check_finished_isoc_tx_epids(void) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + struct USB_SB_Desc* sb_desc; + int epid_done; + + /* Protect TxIsocEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if (TxIsocEPList[epid].sub == 0 || epid == INVALID_EPID || + !epid_out_traffic(epid)) { + /* Nothing here to see. */ + continue; + } + ASSERT(epid_inuse(epid)); + ASSERT(epid_isoc(epid)); + + sb_desc = phys_to_virt(TxIsocEPList[epid].sub); + /* Find the last descriptor of the currently active URB for this ep. + This is the first descriptor in the sub list marked for a descriptor + interrupt. */ + while (sb_desc && !IO_EXTRACT(USB_SB_command, intr, sb_desc->command)) { + sb_desc = sb_desc->next ? phys_to_virt(sb_desc->next) : 0; + } + ASSERT(sb_desc); + + isoc_dbg("Descr IRQ checking epid:%d sub:0x%x intr:0x%x\n", + epid, (unsigned int)phys_to_virt(TxIsocEPList[epid].sub), + (unsigned int)sb_desc); + + urb = activeUrbList[epid]; + if(urb == NULL) { + isoc_err("Isoc Descr irq on epid:%d with no active URB\n", epid); + continue; + } + + epid_done = 0; + while(urb && !epid_done) { + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS); + ASSERT(usb_pipeout(urb->pipe)); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + ASSERT(urb_priv->urb_state == STARTED || + urb_priv->urb_state == UNLINK); + + if (sb_desc != urb_priv->last_sb) { + /* This urb has been sent. */ + urb_priv->isoc_out_done = 1; + + } else { /* Found URB that has last_sb as the interrupt reason */ + + /* Check if EP has been disabled, meaning that all transfers are done*/ + if(!(TxIsocEPList[epid].command & IO_MASK(USB_EP_command, enable))) { + ASSERT((sb_desc->command & IO_MASK(USB_SB_command, eol)) == + IO_STATE(USB_SB_command, eol, yes)); + ASSERT(sb_desc->next == 0); + urb_priv->isoc_out_done = 1; + } else { + isoc_dbg("Skipping URB:0x%x[%d] because EP not disabled yet\n", + (unsigned int)urb, urb_priv->urb_num); + } + /* Stop looking any further in queue */ + epid_done = 1; + } + + if (!epid_done) { + if(urb == activeUrbList[epid]) { + urb = urb_list_first(epid); + } else { + urb = urb_list_next(urb, epid); + } + } + } /* END: while(urb && !epid_done) */ + } + + local_irq_restore(flags); +} + + +/* This is where the Out Isoc URBs are realy completed. This function is + scheduled from tc_dma_tx_interrupt() when one or more Out Isoc transfers + are done. This functions completes all URBs earlier marked with + isoc_out_done by fast interrupt routine check_finished_isoc_tx_epids() */ + +static void complete_isoc_bottom_half(struct work_struct* work) { + struct crisv10_isoc_complete_data *comp_data; + struct usb_iso_packet_descriptor *packet; + struct crisv10_urb_priv * urb_priv; + unsigned long flags; + struct urb* urb; + int epid_done; + int epid; + int i; + + comp_data = container_of(work, struct crisv10_isoc_complete_data, usb_bh); + + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS - 1; epid++) { + if(!epid_inuse(epid) || !epid_isoc(epid) || !epid_out_traffic(epid) || epid == DUMMY_EPID) { + /* Only check valid Out Isoc epids */ + continue; + } + + isoc_dbg("Isoc bottom-half checking epid:%d, sub:0x%x\n", epid, + (unsigned int)phys_to_virt(TxIsocEPList[epid].sub)); + + /* The descriptor interrupt handler has marked all transmitted Out Isoc + URBs with isoc_out_done. Now we traverse all epids and for all that + have out Isoc traffic we traverse its URB list and complete the + transmitted URBs. */ + epid_done = 0; + while (!epid_done) { + + /* Get the active urb (if any) */ + urb = activeUrbList[epid]; + if (urb == 0) { + isoc_dbg("No active URB on epid:%d anymore\n", epid); + epid_done = 1; + continue; + } + + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS); + ASSERT(usb_pipeout(urb->pipe)); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + if (!(urb_priv->isoc_out_done)) { + /* We have reached URB that isn't flaged done yet, stop traversing. */ + isoc_dbg("Stoped traversing Out Isoc URBs on epid:%d" + " before not yet flaged URB:0x%x[%d]\n", + epid, (unsigned int)urb, urb_priv->urb_num); + epid_done = 1; + continue; + } + + /* This urb has been sent. */ + isoc_dbg("Found URB:0x%x[%d] that is flaged isoc_out_done\n", + (unsigned int)urb, urb_priv->urb_num); + + /* Set ok on transfered packets for this URB and finish it */ + for (i = 0; i < urb->number_of_packets; i++) { + packet = &urb->iso_frame_desc[i]; + packet->status = 0; + packet->actual_length = packet->length; + } + urb_priv->isoc_packet_counter = urb->number_of_packets; + tc_finish_urb(comp_data->hcd, urb, 0); + + } /* END: while(!epid_done) */ + } /* END: for(epid...) */ + + local_irq_restore(flags); + kmem_cache_free(isoc_compl_cache, comp_data); +} + + +static void check_finished_intr_tx_epids(struct usb_hcd *hcd) { + unsigned long flags; + int epid; + struct urb *urb; + struct crisv10_urb_priv * urb_priv; + volatile struct USB_EP_Desc *curr_ep; /* Current EP, the iterator. */ + volatile struct USB_EP_Desc *next_ep; /* The EP after current. */ + + /* Protect TxintrEPList */ + local_irq_save(flags); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if(!epid_inuse(epid) || !epid_intr(epid) || !epid_out_traffic(epid)) { + /* Nothing to see on this epid. Only check valid Out Intr epids */ + continue; + } + + urb = activeUrbList[epid]; + if(urb == 0) { + intr_warn("Found Out Intr epid:%d with no active URB\n", epid); + continue; + } + + /* Sanity check. */ + ASSERT(usb_pipetype(urb->pipe) == PIPE_INTERRUPT); + ASSERT(usb_pipeout(urb->pipe)); + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + + /* Go through EPs between first and second sof-EP. It's here Out Intr EPs + are inserted.*/ + curr_ep = &TxIntrEPList[0]; + do { + next_ep = (struct USB_EP_Desc *)phys_to_virt(curr_ep->next); + if(next_ep == urb_priv->intr_ep_pool[0]) { + /* We found the Out Intr EP for this epid */ + + /* Disable it so it doesn't get processed again */ + next_ep->command &= ~IO_MASK(USB_EP_command, enable); + + /* Finish the active Out Intr URB with status OK */ + tc_finish_urb(hcd, urb, 0); + } + curr_ep = phys_to_virt(curr_ep->next); + } while (curr_ep != &TxIntrEPList[1]); + + } + local_irq_restore(flags); +} + +/* Interrupt handler for DMA8/IRQ24 with subchannels (called from hardware intr) */ +static irqreturn_t tc_dma_tx_interrupt(int irq, void *vhc) { + struct usb_hcd *hcd = (struct usb_hcd*)vhc; + ASSERT(hcd); + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub0_descr)) { + /* Clear this interrupt */ + *R_DMA_CH8_SUB0_CLR_INTR = IO_STATE(R_DMA_CH8_SUB0_CLR_INTR, clr_descr, do); + restart_dma8_sub0(); + } + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub1_descr)) { + /* Clear this interrupt */ + *R_DMA_CH8_SUB1_CLR_INTR = IO_STATE(R_DMA_CH8_SUB1_CLR_INTR, clr_descr, do); + check_finished_ctrl_tx_epids(hcd); + } + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub2_descr)) { + /* Clear this interrupt */ + *R_DMA_CH8_SUB2_CLR_INTR = IO_STATE(R_DMA_CH8_SUB2_CLR_INTR, clr_descr, do); + check_finished_intr_tx_epids(hcd); + } + + if (*R_IRQ_READ2 & IO_MASK(R_IRQ_READ2, dma8_sub3_descr)) { + struct crisv10_isoc_complete_data* comp_data; + + /* Flag done Out Isoc for later completion */ + check_finished_isoc_tx_epids(); + + /* Clear this interrupt */ + *R_DMA_CH8_SUB3_CLR_INTR = IO_STATE(R_DMA_CH8_SUB3_CLR_INTR, clr_descr, do); + /* Schedule bottom half of Out Isoc completion function. This function + finishes the URBs marked with isoc_out_done */ + comp_data = (struct crisv10_isoc_complete_data*) + kmem_cache_alloc(isoc_compl_cache, GFP_ATOMIC); + ASSERT(comp_data != NULL); + comp_data ->hcd = hcd; + + INIT_WORK(&comp_data->usb_bh, complete_isoc_bottom_half); + schedule_work(&comp_data->usb_bh); + } + + return IRQ_HANDLED; +} + +/* Interrupt handler for DMA9/IRQ25 (called from hardware intr) */ +static irqreturn_t tc_dma_rx_interrupt(int irq, void *vhc) { + unsigned long flags; + struct urb *urb; + struct usb_hcd *hcd = (struct usb_hcd*)vhc; + struct crisv10_urb_priv *urb_priv; + int epid = 0; + int real_error; + + ASSERT(hcd); + + /* Clear this interrupt. */ + *R_DMA_CH9_CLR_INTR = IO_STATE(R_DMA_CH9_CLR_INTR, clr_eop, do); + + /* Custom clear interrupt for this interrupt */ + /* The reason we cli here is that we call the driver's callback functions. */ + local_irq_save(flags); + + /* Note that this while loop assumes that all packets span only + one rx descriptor. */ + while(myNextRxDesc->status & IO_MASK(USB_IN_status, eop)) { + epid = IO_EXTRACT(USB_IN_status, epid, myNextRxDesc->status); + /* Get the active URB for this epid */ + urb = activeUrbList[epid]; + + ASSERT(epid_inuse(epid)); + if (!urb) { + dma_err("No urb for epid %d in rx interrupt\n", epid); + goto skip_out; + } + + /* Check if any errors on epid */ + real_error = 0; + if (myNextRxDesc->status & IO_MASK(USB_IN_status, error)) { + __u32 r_usb_ept_data; + + if (usb_pipeisoc(urb->pipe)) { + r_usb_ept_data = etrax_epid_iso_get(epid); + if((r_usb_ept_data & IO_MASK(R_USB_EPT_DATA_ISO, valid)) && + (IO_EXTRACT(R_USB_EPT_DATA_ISO, error_code, r_usb_ept_data) == 0) && + (myNextRxDesc->status & IO_MASK(USB_IN_status, nodata))) { + /* Not an error, just a failure to receive an expected iso + in packet in this frame. This is not documented + in the designers reference. Continue processing. + */ + } else real_error = 1; + } else real_error = 1; + } + + if(real_error) { + dma_err("Error in RX descr on epid:%d for URB 0x%x", + epid, (unsigned int)urb); + dump_ept_data(epid); + dump_in_desc(myNextRxDesc); + goto skip_out; + } + + urb_priv = (struct crisv10_urb_priv *)urb->hcpriv; + ASSERT(urb_priv); + ASSERT(urb_priv->urb_state == STARTED || + urb_priv->urb_state == UNLINK); + + if ((usb_pipetype(urb->pipe) == PIPE_BULK) || + (usb_pipetype(urb->pipe) == PIPE_CONTROL) || + (usb_pipetype(urb->pipe) == PIPE_INTERRUPT)) { + + /* We get nodata for empty data transactions, and the rx descriptor's + hw_len field is not valid in that case. No data to copy in other + words. */ + if (myNextRxDesc->status & IO_MASK(USB_IN_status, nodata)) { + /* No data to copy */ + } else { + /* + dma_dbg("Processing RX for URB:0x%x epid:%d (data:%d ofs:%d)\n", + (unsigned int)urb, epid, myNextRxDesc->hw_len, + urb_priv->rx_offset); + */ + /* Only copy data if URB isn't flaged to be unlinked*/ + if(urb_priv->urb_state != UNLINK) { + /* Make sure the data fits in the buffer. */ + if(urb_priv->rx_offset + myNextRxDesc->hw_len + <= urb->transfer_buffer_length) { + + /* Copy the data to URBs buffer */ + memcpy(urb->transfer_buffer + urb_priv->rx_offset, + phys_to_virt(myNextRxDesc->buf), myNextRxDesc->hw_len); + urb_priv->rx_offset += myNextRxDesc->hw_len; + } else { + /* Signal overflow when returning URB */ + urb->status = -EOVERFLOW; + tc_finish_urb_later(hcd, urb, urb->status); + } + } + } + + /* Check if it was the last packet in the transfer */ + if (myNextRxDesc->status & IO_MASK(USB_IN_status, eot)) { + /* Special handling for In Ctrl URBs. */ + if(usb_pipecontrol(urb->pipe) && usb_pipein(urb->pipe) && + !(urb_priv->ctrl_zout_done)) { + /* Flag that RX part of Ctrl transfer is done. Because zout descr + interrupt hasn't happend yet will the URB be finished in the + TX-Interrupt. */ + urb_priv->ctrl_rx_done = 1; + tc_dbg("Not finishing In Ctrl URB:0x%x from rx_interrupt, waiting" + " for zout\n", (unsigned int)urb); + } else { + tc_finish_urb(hcd, urb, 0); + } + } + } else { /* ISOC RX */ + /* + isoc_dbg("Processing RX for epid:%d (URB:0x%x) ISOC pipe\n", + epid, (unsigned int)urb); + */ + + struct usb_iso_packet_descriptor *packet; + + if (urb_priv->urb_state == UNLINK) { + isoc_warn("Ignoring Isoc Rx data for urb being unlinked.\n"); + goto skip_out; + } else if (urb_priv->urb_state == NOT_STARTED) { + isoc_err("What? Got Rx data for Isoc urb that isn't started?\n"); + goto skip_out; + } + + packet = &urb->iso_frame_desc[urb_priv->isoc_packet_counter]; + ASSERT(packet); + packet->status = 0; + + if (myNextRxDesc->status & IO_MASK(USB_IN_status, nodata)) { + /* We get nodata for empty data transactions, and the rx descriptor's + hw_len field is not valid in that case. We copy 0 bytes however to + stay in synch. */ + packet->actual_length = 0; + } else { + packet->actual_length = myNextRxDesc->hw_len; + /* Make sure the data fits in the buffer. */ + ASSERT(packet->actual_length <= packet->length); + memcpy(urb->transfer_buffer + packet->offset, + phys_to_virt(myNextRxDesc->buf), packet->actual_length); + if(packet->actual_length > 0) + isoc_dbg("Copied %d bytes, packet %d for URB:0x%x[%d]\n", + packet->actual_length, urb_priv->isoc_packet_counter, + (unsigned int)urb, urb_priv->urb_num); + } + + /* Increment the packet counter. */ + urb_priv->isoc_packet_counter++; + + /* Note that we don't care about the eot field in the rx descriptor's + status. It will always be set for isoc traffic. */ + if (urb->number_of_packets == urb_priv->isoc_packet_counter) { + /* Complete the urb with status OK. */ + tc_finish_urb(hcd, urb, 0); + } + } + + skip_out: + myNextRxDesc->status = 0; + myNextRxDesc->command |= IO_MASK(USB_IN_command, eol); + myLastRxDesc->command &= ~IO_MASK(USB_IN_command, eol); + myLastRxDesc = myNextRxDesc; + myNextRxDesc = phys_to_virt(myNextRxDesc->next); + flush_etrax_cache(); + *R_DMA_CH9_CMD = IO_STATE(R_DMA_CH9_CMD, cmd, restart); + } + + local_irq_restore(flags); + + return IRQ_HANDLED; +} + +static void tc_bulk_start_timer_func(unsigned long dummy) { + /* We might enable an EP descriptor behind the current DMA position when + it's about to decide that there are no more bulk traffic and it should + stop the bulk channel. + Therefore we periodically check if the bulk channel is stopped and there + is an enabled bulk EP descriptor, in which case we start the bulk + channel. */ + + if (!(*R_DMA_CH8_SUB0_CMD & IO_MASK(R_DMA_CH8_SUB0_CMD, cmd))) { + int epid; + + timer_dbg("bulk_start_timer: Bulk DMA channel not running.\n"); + + for (epid = 0; epid < NBR_OF_EPIDS; epid++) { + if (TxBulkEPList[epid].command & IO_MASK(USB_EP_command, enable)) { + timer_warn("Found enabled EP for epid %d, starting bulk channel.\n", + epid); + restart_dma8_sub0(); + + /* Restart the bulk eot timer since we just started the bulk channel.*/ + mod_timer(&bulk_eot_timer, jiffies + BULK_EOT_TIMER_INTERVAL); + + /* No need to search any further. */ + break; + } + } + } else { + timer_dbg("bulk_start_timer: Bulk DMA channel running.\n"); + } +} + +static void tc_bulk_eot_timer_func(unsigned long dummy) { + struct usb_hcd *hcd = (struct usb_hcd*)dummy; + ASSERT(hcd); + /* Because of a race condition in the top half, we might miss a bulk eot. + This timer "simulates" a bulk eot if we don't get one for a while, + hopefully correcting the situation. */ + timer_dbg("bulk_eot_timer timed out.\n"); + check_finished_bulk_tx_epids(hcd, 1); +} + + +/*************************************************************/ +/*************************************************************/ +/* Device driver block */ +/*************************************************************/ +/*************************************************************/ + +/* Forward declarations for device driver functions */ +static int devdrv_hcd_probe(struct device *); +static int devdrv_hcd_remove(struct device *); +#ifdef CONFIG_PM +static int devdrv_hcd_suspend(struct device *, u32, u32); +static int devdrv_hcd_resume(struct device *, u32); +#endif /* CONFIG_PM */ + +/* the device */ +static struct platform_device *devdrv_hc_platform_device; + +/* device driver interface */ +static struct device_driver devdrv_hc_device_driver = { + .name = (char *) hc_name, + .bus = &platform_bus_type, + + .probe = devdrv_hcd_probe, + .remove = devdrv_hcd_remove, + +#ifdef CONFIG_PM + .suspend = devdrv_hcd_suspend, + .resume = devdrv_hcd_resume, +#endif /* CONFIG_PM */ +}; + +/* initialize the host controller and driver */ +static int __init_or_module devdrv_hcd_probe(struct device *dev) +{ + struct usb_hcd *hcd; + struct crisv10_hcd *crisv10_hcd; + int retval; + + /* Check DMA burst length */ + if(IO_EXTRACT(R_BUS_CONFIG, dma_burst, *R_BUS_CONFIG) != + IO_STATE(R_BUS_CONFIG, dma_burst, burst32)) { + devdrv_err("Invalid DMA burst length in Etrax 100LX," + " needs to be 32\n"); + return -EPERM; + } + + //XXX: dev->usb_id don't exist, using "" instread? - claudio + hcd = usb_create_hcd(&crisv10_hc_driver, dev, ""); + if (!hcd) + return -ENOMEM; + + crisv10_hcd = hcd_to_crisv10_hcd(hcd); + spin_lock_init(&crisv10_hcd->lock); + crisv10_hcd->num_ports = num_ports(); + crisv10_hcd->running = 0; + + dev_set_drvdata(dev, crisv10_hcd); + + devdrv_dbg("ETRAX USB IRQs HC:%d RX:%d TX:%d\n", ETRAX_USB_HC_IRQ, + ETRAX_USB_RX_IRQ, ETRAX_USB_TX_IRQ); + + /* Print out chip version read from registers */ + int rev_maj = *R_USB_REVISION & IO_MASK(R_USB_REVISION, major); + int rev_min = *R_USB_REVISION & IO_MASK(R_USB_REVISION, minor); + if(rev_min == 0) { + devdrv_info("Etrax 100LX USB Revision %d v1,2\n", rev_maj); + } else { + devdrv_info("Etrax 100LX USB Revision %d v%d\n", rev_maj, rev_min); + } + + devdrv_info("Bulk timer interval, start:%d eot:%d\n", + BULK_START_TIMER_INTERVAL, + BULK_EOT_TIMER_INTERVAL); + + + /* Init root hub data structures */ + if(rh_init()) { + devdrv_err("Failed init data for Root Hub\n"); + retval = -ENOMEM; + } + + if(port_in_use(0)) { + if (cris_request_io_interface(if_usb_1, "ETRAX100LX USB-HCD")) { + printk(KERN_CRIT "usb-host: request IO interface usb1 failed"); + retval = -EBUSY; + goto out; + } + devdrv_info("Claimed interface for USB physical port 1\n"); + } + if(port_in_use(1)) { + if (cris_request_io_interface(if_usb_2, "ETRAX100LX USB-HCD")) { + /* Free first interface if second failed to be claimed */ + if(port_in_use(0)) { + cris_free_io_interface(if_usb_1); + } + printk(KERN_CRIT "usb-host: request IO interface usb2 failed"); + retval = -EBUSY; + goto out; + } + devdrv_info("Claimed interface for USB physical port 2\n"); + } + + /* Init transfer controller structs and locks */ + if((retval = tc_init(hcd)) != 0) { + goto out; + } + + /* Attach interrupt functions for DMA and init DMA controller */ + if((retval = tc_dma_init(hcd)) != 0) { + goto out; + } + + /* Attach the top IRQ handler for USB controller interrupts */ + if (request_irq(ETRAX_USB_HC_IRQ, crisv10_hcd_top_irq, 0, + "ETRAX 100LX built-in USB (HC)", hcd)) { + err("Could not allocate IRQ %d for USB", ETRAX_USB_HC_IRQ); + retval = -EBUSY; + goto out; + } + + /* iso_eof is only enabled when isoc traffic is running. */ + *R_USB_IRQ_MASK_SET = + /* IO_STATE(R_USB_IRQ_MASK_SET, iso_eof, set) | */ + IO_STATE(R_USB_IRQ_MASK_SET, bulk_eot, set) | + IO_STATE(R_USB_IRQ_MASK_SET, epid_attn, set) | + IO_STATE(R_USB_IRQ_MASK_SET, port_status, set) | + IO_STATE(R_USB_IRQ_MASK_SET, ctl_status, set); + + + crisv10_ready_wait(); + /* Reset the USB interface. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, reset); + + /* Designer's Reference, p. 8 - 10 says we should Initate R_USB_FM_PSTART to + 0x2A30 (10800), to guarantee that control traffic gets 10% of the + bandwidth, and periodic transfer may allocate the rest (90%). + This doesn't work though. + The value 11960 is chosen to be just after the SOF token, with a couple + of bit times extra for possible bit stuffing. */ + *R_USB_FM_PSTART = IO_FIELD(R_USB_FM_PSTART, value, 11960); + + crisv10_ready_wait(); + /* Configure the USB interface as a host controller. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_config); + + + /* Check so controller not busy before enabling ports */ + crisv10_ready_wait(); + + /* Enable selected USB ports */ + if(port_in_use(0)) { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, no); + } else { + *R_USB_PORT1_DISABLE = IO_STATE(R_USB_PORT1_DISABLE, disable, yes); + } + if(port_in_use(1)) { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, no); + } else { + *R_USB_PORT2_DISABLE = IO_STATE(R_USB_PORT2_DISABLE, disable, yes); + } + + crisv10_ready_wait(); + /* Start processing of USB traffic. */ + *R_USB_COMMAND = + IO_STATE(R_USB_COMMAND, port_sel, nop) | + IO_STATE(R_USB_COMMAND, port_cmd, reset) | + IO_STATE(R_USB_COMMAND, ctrl_cmd, host_run); + + /* Do not continue probing initialization before USB interface is done */ + crisv10_ready_wait(); + + /* Register our Host Controller to USB Core + * Finish the remaining parts of generic HCD initialization: allocate the + * buffers of consistent memory, register the bus + * and call the driver's reset() and start() routines. */ + retval = usb_add_hcd(hcd, ETRAX_USB_HC_IRQ, IRQF_DISABLED); + if (retval != 0) { + devdrv_err("Failed registering HCD driver\n"); + goto out; + } + + return 0; + + out: + devdrv_hcd_remove(dev); + return retval; +} + + +/* cleanup after the host controller and driver */ +static int __init_or_module devdrv_hcd_remove(struct device *dev) +{ + struct crisv10_hcd *crisv10_hcd = dev_get_drvdata(dev); + struct usb_hcd *hcd; + + if (!crisv10_hcd) + return 0; + hcd = crisv10_hcd_to_hcd(crisv10_hcd); + + + /* Stop USB Controller in Etrax 100LX */ + crisv10_hcd_reset(hcd); + + usb_remove_hcd(hcd); + devdrv_dbg("Removed HCD from USB Core\n"); + + /* Free USB Controller IRQ */ + free_irq(ETRAX_USB_HC_IRQ, NULL); + + /* Free resources */ + tc_dma_destroy(); + tc_destroy(); + + + if(port_in_use(0)) { + cris_free_io_interface(if_usb_1); + } + if(port_in_use(1)) { + cris_free_io_interface(if_usb_2); + } + + devdrv_dbg("Freed all claimed resources\n"); + + return 0; +} + + +#ifdef CONFIG_PM + +static int devdrv_hcd_suspend(struct usb_hcd *hcd, u32 state, u32 level) +{ + return 0; /* no-op for now */ +} + +static int devdrv_hcd_resume(struct usb_hcd *hcd, u32 level) +{ + return 0; /* no-op for now */ +} + +#endif /* CONFIG_PM */ + + +/*************************************************************/ +/*************************************************************/ +/* Module block */ +/*************************************************************/ +/*************************************************************/ + +/* register driver */ +static int __init module_hcd_init(void) +{ + + if (usb_disabled()) + return -ENODEV; + + /* Here we select enabled ports by following defines created from + menuconfig */ +#ifndef CONFIG_ETRAX_USB_HOST_PORT1 + ports &= ~(1<<0); +#endif +#ifndef CONFIG_ETRAX_USB_HOST_PORT2 + ports &= ~(1<<1); +#endif + + printk(KERN_INFO "%s version "VERSION" "COPYRIGHT"\n", product_desc); + + devdrv_hc_platform_device = + platform_device_register_simple((char *) hc_name, 0, NULL, 0); + + if (IS_ERR(devdrv_hc_platform_device)) + return PTR_ERR(devdrv_hc_platform_device); + return driver_register(&devdrv_hc_device_driver); + /* + * Note that we do not set the DMA mask for the device, + * i.e. we pretend that we will use PIO, since no specific + * allocation routines are needed for DMA buffers. This will + * cause the HCD buffer allocation routines to fall back to + * kmalloc(). + */ +} + +/* unregister driver */ +static void __exit module_hcd_exit(void) +{ + driver_unregister(&devdrv_hc_device_driver); +} + + +/* Module hooks */ +module_init(module_hcd_init); +module_exit(module_hcd_exit); diff --git a/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.h b/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.h new file mode 100644 index 0000000000..19a7e2ee71 --- /dev/null +++ b/target/linux/etrax/files-2.6.30/drivers/usb/host/hc-crisv10.h @@ -0,0 +1,666 @@ +#ifndef __LINUX_ETRAX_USB_H +#define __LINUX_ETRAX_USB_H + +#include +#include + +struct USB_IN_Desc { + volatile __u16 sw_len; + volatile __u16 command; + volatile unsigned long next; + volatile unsigned long buf; + volatile __u16 hw_len; + volatile __u16 status; +}; + +struct USB_SB_Desc { + volatile __u16 sw_len; + volatile __u16 command; + volatile unsigned long next; + volatile unsigned long buf; +}; + +struct USB_EP_Desc { + volatile __u16 hw_len; + volatile __u16 command; + volatile unsigned long sub; + volatile unsigned long next; +}; + + +/* Root Hub port status struct */ +struct crisv10_rh { + volatile __u16 wPortChange[2]; + volatile __u16 wPortStatusPrev[2]; +}; + +/* HCD description */ +struct crisv10_hcd { + spinlock_t lock; + __u8 num_ports; + __u8 running; +}; + + +/* Endpoint HC private data description */ +struct crisv10_ep_priv { + int epid; +}; + +/* Additional software state info for a USB Controller epid */ +struct etrax_epid { + __u8 inuse; /* !0 = setup in Etrax and used for a endpoint */ + __u8 disabled; /* !0 = Temporarly disabled to avoid resubmission */ + __u8 type; /* Setup as: PIPE_BULK, PIPE_CONTROL ... */ + __u8 out_traffic; /* !0 = This epid is for out traffic */ +}; + +/* Struct to hold information of scheduled later URB completion */ +struct urb_later_data { + struct delayed_work dws; + struct usb_hcd *hcd; + struct urb *urb; + int urb_num; + int status; +}; + + +typedef enum { + STARTED, + NOT_STARTED, + UNLINK, +} crisv10_urb_state_t; + + +struct crisv10_urb_priv { + /* Sequence number for this URB. Every new submited URB gets this from + a incrementing counter. Used when a URB is scheduled for later finish to + be sure that the intended URB hasn't already been completed (device + drivers has a tendency to reuse URBs once they are completed, causing us + to not be able to single old ones out only based on the URB pointer.) */ + __u32 urb_num; + + /* The first_sb field is used for freeing all SB descriptors belonging + to an urb. The corresponding ep descriptor's sub pointer cannot be + used for this since the DMA advances the sub pointer as it processes + the sb list. */ + struct USB_SB_Desc *first_sb; + + /* The last_sb field referes to the last SB descriptor that belongs to + this urb. This is important to know so we can free the SB descriptors + that ranges between first_sb and last_sb. */ + struct USB_SB_Desc *last_sb; + + /* The rx_offset field is used in ctrl and bulk traffic to keep track + of the offset in the urb's transfer_buffer where incoming data should be + copied to. */ + __u32 rx_offset; + + /* Counter used in isochronous transfers to keep track of the + number of packets received/transmitted. */ + __u32 isoc_packet_counter; + + /* Flag that marks if this Isoc Out URB has finished it's transfer. Used + because several URBs can be finished before list is processed */ + __u8 isoc_out_done; + + /* This field is used to pass information about the urb's current state + between the various interrupt handlers (thus marked volatile). */ + volatile crisv10_urb_state_t urb_state; + + /* In Ctrl transfers consist of (at least) 3 packets: SETUP, IN and ZOUT. + When DMA8 sub-channel 2 has processed the SB list for this sequence we + get a interrupt. We also get a interrupt for In transfers and which + one of these interrupts that comes first depends of data size and device. + To be sure that we have got both interrupts before we complete the URB + we have these to flags that shows which part that has completed. + We can then check when we get one of the interrupts that if the other has + occured it's safe for us to complete the URB, otherwise we set appropriate + flag and do the completion when we get the other interrupt. */ + volatile unsigned char ctrl_zout_done; + volatile unsigned char ctrl_rx_done; + + /* Connection between the submitted urb and ETRAX epid number */ + __u8 epid; + + /* The rx_data_list field is used for periodic traffic, to hold + received data for later processing in the the complete_urb functions, + where the data us copied to the urb's transfer_buffer. Basically, we + use this intermediate storage because we don't know when it's safe to + reuse the transfer_buffer (FIXME?). */ + struct list_head rx_data_list; + + + /* The interval time rounded up to closest 2^N */ + int interval; + + /* Pool of EP descriptors needed if it's a INTR transfer. + Amount of EPs in pool correspons to how many INTR that should + be inserted in TxIntrEPList (max 128, defined by MAX_INTR_INTERVAL) */ + struct USB_EP_Desc* intr_ep_pool[128]; + + /* The mount of EPs allocated for this INTR URB */ + int intr_ep_pool_length; + + /* Pointer to info struct if URB is scheduled to be finished later */ + struct urb_later_data* later_data; + + /* Allocated bandwidth for isochronous and interrupt traffic */ + int bandwidth; +}; + + +/* This struct is for passing data from the top half to the bottom half irq + handlers */ +struct crisv10_irq_reg { + struct usb_hcd* hcd; + __u32 r_usb_epid_attn; + __u8 r_usb_status; + __u16 r_usb_rh_port_status_1; + __u16 r_usb_rh_port_status_2; + __u32 r_usb_irq_mask_read; + __u32 r_usb_fm_number; + struct work_struct usb_bh; +}; + + +/* This struct is for passing data from the isoc top half to the isoc bottom + half. */ +struct crisv10_isoc_complete_data { + struct usb_hcd *hcd; + struct urb *urb; + struct work_struct usb_bh; +}; + +/* Entry item for URB lists for each endpint */ +typedef struct urb_entry +{ + struct urb *urb; + struct list_head list; +} urb_entry_t; + +/* --------------------------------------------------------------------------- + Virtual Root HUB + ------------------------------------------------------------------------- */ +/* destination of request */ +#define RH_INTERFACE 0x01 +#define RH_ENDPOINT 0x02 +#define RH_OTHER 0x03 + +#define RH_CLASS 0x20 +#define RH_VENDOR 0x40 + +/* Requests: bRequest << 8 | bmRequestType */ +#define RH_GET_STATUS 0x0080 +#define RH_CLEAR_FEATURE 0x0100 +#define RH_SET_FEATURE 0x0300 +#define RH_SET_ADDRESS 0x0500 +#define RH_GET_DESCRIPTOR 0x0680 +#define RH_SET_DESCRIPTOR 0x0700 +#define RH_GET_CONFIGURATION 0x0880 +#define RH_SET_CONFIGURATION 0x0900 +#define RH_GET_STATE 0x0280 +#define RH_GET_INTERFACE 0x0A80 +#define RH_SET_INTERFACE 0x0B00 +#define RH_SYNC_FRAME 0x0C80 +/* Our Vendor Specific Request */ +#define RH_SET_EP 0x2000 + + +/* Hub port features */ +#define RH_PORT_CONNECTION 0x00 +#define RH_PORT_ENABLE 0x01 +#define RH_PORT_SUSPEND 0x02 +#define RH_PORT_OVER_CURRENT 0x03 +#define RH_PORT_RESET 0x04 +#define RH_PORT_POWER 0x08 +#define RH_PORT_LOW_SPEED 0x09 +#define RH_C_PORT_CONNECTION 0x10 +#define RH_C_PORT_ENABLE 0x11 +#define RH_C_PORT_SUSPEND 0x12 +#define RH_C_PORT_OVER_CURRENT 0x13 +#define RH_C_PORT_RESET 0x14 + +/* Hub features */ +#define RH_C_HUB_LOCAL_POWER 0x00 +#define RH_C_HUB_OVER_CURRENT 0x01 + +#define RH_DEVICE_REMOTE_WAKEUP 0x00 +#define RH_ENDPOINT_STALL 0x01 + +/* Our Vendor Specific feature */ +#define RH_REMOVE_EP 0x00 + + +#define RH_ACK 0x01 +#define RH_REQ_ERR -1 +#define RH_NACK 0x00 + +/* Field definitions for */ + +#define USB_IN_command__eol__BITNR 0 /* command macros */ +#define USB_IN_command__eol__WIDTH 1 +#define USB_IN_command__eol__no 0 +#define USB_IN_command__eol__yes 1 + +#define USB_IN_command__intr__BITNR 3 +#define USB_IN_command__intr__WIDTH 1 +#define USB_IN_command__intr__no 0 +#define USB_IN_command__intr__yes 1 + +#define USB_IN_status__eop__BITNR 1 /* status macros. */ +#define USB_IN_status__eop__WIDTH 1 +#define USB_IN_status__eop__no 0 +#define USB_IN_status__eop__yes 1 + +#define USB_IN_status__eot__BITNR 5 +#define USB_IN_status__eot__WIDTH 1 +#define USB_IN_status__eot__no 0 +#define USB_IN_status__eot__yes 1 + +#define USB_IN_status__error__BITNR 6 +#define USB_IN_status__error__WIDTH 1 +#define USB_IN_status__error__no 0 +#define USB_IN_status__error__yes 1 + +#define USB_IN_status__nodata__BITNR 7 +#define USB_IN_status__nodata__WIDTH 1 +#define USB_IN_status__nodata__no 0 +#define USB_IN_status__nodata__yes 1 + +#define USB_IN_status__epid__BITNR 8 +#define USB_IN_status__epid__WIDTH 5 + +#define USB_EP_command__eol__BITNR 0 +#define USB_EP_command__eol__WIDTH 1 +#define USB_EP_command__eol__no 0 +#define USB_EP_command__eol__yes 1 + +#define USB_EP_command__eof__BITNR 1 +#define USB_EP_command__eof__WIDTH 1 +#define USB_EP_command__eof__no 0 +#define USB_EP_command__eof__yes 1 + +#define USB_EP_command__intr__BITNR 3 +#define USB_EP_command__intr__WIDTH 1 +#define USB_EP_command__intr__no 0 +#define USB_EP_command__intr__yes 1 + +#define USB_EP_command__enable__BITNR 4 +#define USB_EP_command__enable__WIDTH 1 +#define USB_EP_command__enable__no 0 +#define USB_EP_command__enable__yes 1 + +#define USB_EP_command__hw_valid__BITNR 5 +#define USB_EP_command__hw_valid__WIDTH 1 +#define USB_EP_command__hw_valid__no 0 +#define USB_EP_command__hw_valid__yes 1 + +#define USB_EP_command__epid__BITNR 8 +#define USB_EP_command__epid__WIDTH 5 + +#define USB_SB_command__eol__BITNR 0 /* command macros. */ +#define USB_SB_command__eol__WIDTH 1 +#define USB_SB_command__eol__no 0 +#define USB_SB_command__eol__yes 1 + +#define USB_SB_command__eot__BITNR 1 +#define USB_SB_command__eot__WIDTH 1 +#define USB_SB_command__eot__no 0 +#define USB_SB_command__eot__yes 1 + +#define USB_SB_command__intr__BITNR 3 +#define USB_SB_command__intr__WIDTH 1 +#define USB_SB_command__intr__no 0 +#define USB_SB_command__intr__yes 1 + +#define USB_SB_command__tt__BITNR 4 +#define USB_SB_command__tt__WIDTH 2 +#define USB_SB_command__tt__zout 0 +#define USB_SB_command__tt__in 1 +#define USB_SB_command__tt__out 2 +#define USB_SB_command__tt__setup 3 + + +#define USB_SB_command__rem__BITNR 8 +#define USB_SB_command__rem__WIDTH 6 + +#define USB_SB_command__full__BITNR 6 +#define USB_SB_command__full__WIDTH 1 +#define USB_SB_command__full__no 0 +#define USB_SB_command__full__yes 1 + +#endif +#ifndef __LINUX_ETRAX_USB_H +#define __LINUX_ETRAX_USB_H + +#include +#include + +struct USB_IN_Desc { + volatile __u16 sw_len; + volatile __u16 command; + volatile unsigned long next; + volatile unsigned long buf; + volatile __u16 hw_len; + volatile __u16 status; +}; + +struct USB_SB_Desc { + volatile __u16 sw_len; + volatile __u16 command; + volatile unsigned long next; + volatile unsigned long buf; +}; + +struct USB_EP_Desc { + volatile __u16 hw_len; + volatile __u16 command; + volatile unsigned long sub; + volatile unsigned long next; +}; + + +/* Root Hub port status struct */ +struct crisv10_rh { + volatile __u16 wPortChange[2]; + volatile __u16 wPortStatusPrev[2]; +}; + +/* HCD description */ +struct crisv10_hcd { + spinlock_t lock; + __u8 num_ports; + __u8 running; +}; + + +/* Endpoint HC private data description */ +struct crisv10_ep_priv { + int epid; +}; + +/* Additional software state info for a USB Controller epid */ +struct etrax_epid { + __u8 inuse; /* !0 = setup in Etrax and used for a endpoint */ + __u8 disabled; /* !0 = Temporarly disabled to avoid resubmission */ + __u8 type; /* Setup as: PIPE_BULK, PIPE_CONTROL ... */ + __u8 out_traffic; /* !0 = This epid is for out traffic */ +}; + +/* Struct to hold information of scheduled later URB completion */ +struct urb_later_data { + struct delayed_work dws; + struct usb_hcd *hcd; + struct urb *urb; + int urb_num; + int status; +}; + + +typedef enum { + STARTED, + NOT_STARTED, + UNLINK, +} crisv10_urb_state_t; + + +struct crisv10_urb_priv { + /* Sequence number for this URB. Every new submited URB gets this from + a incrementing counter. Used when a URB is scheduled for later finish to + be sure that the intended URB hasn't already been completed (device + drivers has a tendency to reuse URBs once they are completed, causing us + to not be able to single old ones out only based on the URB pointer.) */ + __u32 urb_num; + + /* The first_sb field is used for freeing all SB descriptors belonging + to an urb. The corresponding ep descriptor's sub pointer cannot be + used for this since the DMA advances the sub pointer as it processes + the sb list. */ + struct USB_SB_Desc *first_sb; + + /* The last_sb field referes to the last SB descriptor that belongs to + this urb. This is important to know so we can free the SB descriptors + that ranges between first_sb and last_sb. */ + struct USB_SB_Desc *last_sb; + + /* The rx_offset field is used in ctrl and bulk traffic to keep track + of the offset in the urb's transfer_buffer where incoming data should be + copied to. */ + __u32 rx_offset; + + /* Counter used in isochronous transfers to keep track of the + number of packets received/transmitted. */ + __u32 isoc_packet_counter; + + /* Flag that marks if this Isoc Out URB has finished it's transfer. Used + because several URBs can be finished before list is processed */ + __u8 isoc_out_done; + + /* This field is used to pass information about the urb's current state + between the various interrupt handlers (thus marked volatile). */ + volatile crisv10_urb_state_t urb_state; + + /* In Ctrl transfers consist of (at least) 3 packets: SETUP, IN and ZOUT. + When DMA8 sub-channel 2 has processed the SB list for this sequence we + get a interrupt. We also get a interrupt for In transfers and which + one of these interrupts that comes first depends of data size and device. + To be sure that we have got both interrupts before we complete the URB + we have these to flags that shows which part that has completed. + We can then check when we get one of the interrupts that if the other has + occured it's safe for us to complete the URB, otherwise we set appropriate + flag and do the completion when we get the other interrupt. */ + volatile unsigned char ctrl_zout_done; + volatile unsigned char ctrl_rx_done; + + /* Connection between the submitted urb and ETRAX epid number */ + __u8 epid; + + /* The rx_data_list field is used for periodic traffic, to hold + received data for later processing in the the complete_urb functions, + where the data us copied to the urb's transfer_buffer. Basically, we + use this intermediate storage because we don't know when it's safe to + reuse the transfer_buffer (FIXME?). */ + struct list_head rx_data_list; + + + /* The interval time rounded up to closest 2^N */ + int interval; + + /* Pool of EP descriptors needed if it's a INTR transfer. + Amount of EPs in pool correspons to how many INTR that should + be inserted in TxIntrEPList (max 128, defined by MAX_INTR_INTERVAL) */ + struct USB_EP_Desc* intr_ep_pool[128]; + + /* The mount of EPs allocated for this INTR URB */ + int intr_ep_pool_length; + + /* Pointer to info struct if URB is scheduled to be finished later */ + struct urb_later_data* later_data; + + /* Allocated bandwidth for isochronous and interrupt traffic */ + int bandwidth; +}; + + +/* This struct is for passing data from the top half to the bottom half irq + handlers */ +struct crisv10_irq_reg { + struct usb_hcd* hcd; + __u32 r_usb_epid_attn; + __u8 r_usb_status; + __u16 r_usb_rh_port_status_1; + __u16 r_usb_rh_port_status_2; + __u32 r_usb_irq_mask_read; + __u32 r_usb_fm_number; + struct work_struct usb_bh; +}; + + +/* This struct is for passing data from the isoc top half to the isoc bottom + half. */ +struct crisv10_isoc_complete_data { + struct usb_hcd *hcd; + struct urb *urb; + struct work_struct usb_bh; +}; + +/* Entry item for URB lists for each endpint */ +typedef struct urb_entry +{ + struct urb *urb; + struct list_head list; +} urb_entry_t; + +/* --------------------------------------------------------------------------- + Virtual Root HUB + ------------------------------------------------------------------------- */ +/* destination of request */ +#define RH_INTERFACE 0x01 +#define RH_ENDPOINT 0x02 +#define RH_OTHER 0x03 + +#define RH_CLASS 0x20 +#define RH_VENDOR 0x40 + +/* Requests: bRequest << 8 | bmRequestType */ +#define RH_GET_STATUS 0x0080 +#define RH_CLEAR_FEATURE 0x0100 +#define RH_SET_FEATURE 0x0300 +#define RH_SET_ADDRESS 0x0500 +#define RH_GET_DESCRIPTOR 0x0680 +#define RH_SET_DESCRIPTOR 0x0700 +#define RH_GET_CONFIGURATION 0x0880 +#define RH_SET_CONFIGURATION 0x0900 +#define RH_GET_STATE 0x0280 +#define RH_GET_INTERFACE 0x0A80 +#define RH_SET_INTERFACE 0x0B00 +#define RH_SYNC_FRAME 0x0C80 +/* Our Vendor Specific Request */ +#define RH_SET_EP 0x2000 + + +/* Hub port features */ +#define RH_PORT_CONNECTION 0x00 +#define RH_PORT_ENABLE 0x01 +#define RH_PORT_SUSPEND 0x02 +#define RH_PORT_OVER_CURRENT 0x03 +#define RH_PORT_RESET 0x04 +#define RH_PORT_POWER 0x08 +#define RH_PORT_LOW_SPEED 0x09 +#define RH_C_PORT_CONNECTION 0x10 +#define RH_C_PORT_ENABLE 0x11 +#define RH_C_PORT_SUSPEND 0x12 +#define RH_C_PORT_OVER_CURRENT 0x13 +#define RH_C_PORT_RESET 0x14 + +/* Hub features */ +#define RH_C_HUB_LOCAL_POWER 0x00 +#define RH_C_HUB_OVER_CURRENT 0x01 + +#define RH_DEVICE_REMOTE_WAKEUP 0x00 +#define RH_ENDPOINT_STALL 0x01 + +/* Our Vendor Specific feature */ +#define RH_REMOVE_EP 0x00 + + +#define RH_ACK 0x01 +#define RH_REQ_ERR -1 +#define RH_NACK 0x00 + +/* Field definitions for */ + +#define USB_IN_command__eol__BITNR 0 /* command macros */ +#define USB_IN_command__eol__WIDTH 1 +#define USB_IN_command__eol__no 0 +#define USB_IN_command__eol__yes 1 + +#define USB_IN_command__intr__BITNR 3 +#define USB_IN_command__intr__WIDTH 1 +#define USB_IN_command__intr__no 0 +#define USB_IN_command__intr__yes 1 + +#define USB_IN_status__eop__BITNR 1 /* status macros. */ +#define USB_IN_status__eop__WIDTH 1 +#define USB_IN_status__eop__no 0 +#define USB_IN_status__eop__yes 1 + +#define USB_IN_status__eot__BITNR 5 +#define USB_IN_status__eot__WIDTH 1 +#define USB_IN_status__eot__no 0 +#define USB_IN_status__eot__yes 1 + +#define USB_IN_status__error__BITNR 6 +#define USB_IN_status__error__WIDTH 1 +#define USB_IN_status__error__no 0 +#define USB_IN_status__error__yes 1 + +#define USB_IN_status__nodata__BITNR 7 +#define USB_IN_status__nodata__WIDTH 1 +#define USB_IN_status__nodata__no 0 +#define USB_IN_status__nodata__yes 1 + +#define USB_IN_status__epid__BITNR 8 +#define USB_IN_status__epid__WIDTH 5 + +#define USB_EP_command__eol__BITNR 0 +#define USB_EP_command__eol__WIDTH 1 +#define USB_EP_command__eol__no 0 +#define USB_EP_command__eol__yes 1 + +#define USB_EP_command__eof__BITNR 1 +#define USB_EP_command__eof__WIDTH 1 +#define USB_EP_command__eof__no 0 +#define USB_EP_command__eof__yes 1 + +#define USB_EP_command__intr__BITNR 3 +#define USB_EP_command__intr__WIDTH 1 +#define USB_EP_command__intr__no 0 +#define USB_EP_command__intr__yes 1 + +#define USB_EP_command__enable__BITNR 4 +#define USB_EP_command__enable__WIDTH 1 +#define USB_EP_command__enable__no 0 +#define USB_EP_command__enable__yes 1 + +#define USB_EP_command__hw_valid__BITNR 5 +#define USB_EP_command__hw_valid__WIDTH 1 +#define USB_EP_command__hw_valid__no 0 +#define USB_EP_command__hw_valid__yes 1 + +#define USB_EP_command__epid__BITNR 8 +#define USB_EP_command__epid__WIDTH 5 + +#define USB_SB_command__eol__BITNR 0 /* command macros. */ +#define USB_SB_command__eol__WIDTH 1 +#define USB_SB_command__eol__no 0 +#define USB_SB_command__eol__yes 1 + +#define USB_SB_command__eot__BITNR 1 +#define USB_SB_command__eot__WIDTH 1 +#define USB_SB_command__eot__no 0 +#define USB_SB_command__eot__yes 1 + +#define USB_SB_command__intr__BITNR 3 +#define USB_SB_command__intr__WIDTH 1 +#define USB_SB_command__intr__no 0 +#define USB_SB_command__intr__yes 1 + +#define USB_SB_command__tt__BITNR 4 +#define USB_SB_command__tt__WIDTH 2 +#define USB_SB_command__tt__zout 0 +#define USB_SB_command__tt__in 1 +#define USB_SB_command__tt__out 2 +#define USB_SB_command__tt__setup 3 + + +#define USB_SB_command__rem__BITNR 8 +#define USB_SB_command__rem__WIDTH 6 + +#define USB_SB_command__full__BITNR 6 +#define USB_SB_command__full__WIDTH 1 +#define USB_SB_command__full__no 0 +#define USB_SB_command__full__yes 1 + +#endif diff --git a/target/linux/etrax/patches-2.6.30/300-usb_support.patch b/target/linux/etrax/patches-2.6.30/300-usb_support.patch new file mode 100644 index 0000000000..9ea39ec596 --- /dev/null +++ b/target/linux/etrax/patches-2.6.30/300-usb_support.patch @@ -0,0 +1,24 @@ +Index: linux-2.6.30-rc5/drivers/usb/host/Makefile +=================================================================== +--- linux-2.6.30-rc5.orig/drivers/usb/host/Makefile 2009-05-09 02:14:14.000000000 +0200 ++++ linux-2.6.30-rc5/drivers/usb/host/Makefile 2009-05-13 14:07:10.000000000 +0200 +@@ -25,6 +25,7 @@ + obj-$(CONFIG_USB_FHCI_HCD) += fhci.o + obj-$(CONFIG_USB_SL811_HCD) += sl811-hcd.o + obj-$(CONFIG_USB_SL811_CS) += sl811_cs.o ++obj-$(CONFIG_ETRAX_ARCH_V10) += hc-crisv10.o + obj-$(CONFIG_USB_U132_HCD) += u132-hcd.o + obj-$(CONFIG_USB_R8A66597_HCD) += r8a66597-hcd.o + obj-$(CONFIG_USB_ISP1760_HCD) += isp1760.o +Index: linux-2.6.30-rc5/drivers/usb/Makefile +=================================================================== +--- linux-2.6.30-rc5.orig/drivers/usb/Makefile 2009-05-13 14:06:45.000000000 +0200 ++++ linux-2.6.30-rc5/drivers/usb/Makefile 2009-05-13 14:07:10.000000000 +0200 +@@ -20,6 +20,7 @@ + obj-$(CONFIG_USB_R8A66597_HCD) += host/ + obj-$(CONFIG_USB_HWA_HCD) += host/ + obj-$(CONFIG_USB_ISP1760_HCD) += host/ ++obj-$(CONFIG_ETRAX_USB_HOST) += host/ + + obj-$(CONFIG_USB_C67X00_HCD) += c67x00/ + -- 2.25.1