mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-16 00:34:20 +08:00
c8c38de9d8
This patch adds support for ehci and ohci controller in the SPEAr platform. Changes since V2: added clear_tt_buffer_complete in ehci_spear_hc_driver Signed-off-by: Deepak Sikri <deepak.sikri@st.com> Signed-off-by: Viresh Kumar <viresh.kumar@st.com> Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
1347 lines
37 KiB
C
1347 lines
37 KiB
C
/*
|
|
* Copyright (c) 2000-2004 by David Brownell
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; either version 2 of the License, or (at your
|
|
* option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
|
|
* or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software Foundation,
|
|
* Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/dmapool.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/init.h>
|
|
#include <linux/timer.h>
|
|
#include <linux/ktime.h>
|
|
#include <linux/list.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/usb.h>
|
|
#include <linux/usb/hcd.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/debugfs.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/uaccess.h>
|
|
|
|
#include <asm/byteorder.h>
|
|
#include <asm/io.h>
|
|
#include <asm/irq.h>
|
|
#include <asm/system.h>
|
|
#include <asm/unaligned.h>
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
/*
|
|
* EHCI hc_driver implementation ... experimental, incomplete.
|
|
* Based on the final 1.0 register interface specification.
|
|
*
|
|
* USB 2.0 shows up in upcoming www.pcmcia.org technology.
|
|
* First was PCMCIA, like ISA; then CardBus, which is PCI.
|
|
* Next comes "CardBay", using USB 2.0 signals.
|
|
*
|
|
* Contains additional contributions by Brad Hards, Rory Bolt, and others.
|
|
* Special thanks to Intel and VIA for providing host controllers to
|
|
* test this driver on, and Cypress (including In-System Design) for
|
|
* providing early devices for those host controllers to talk to!
|
|
*/
|
|
|
|
#define DRIVER_AUTHOR "David Brownell"
|
|
#define DRIVER_DESC "USB 2.0 'Enhanced' Host Controller (EHCI) Driver"
|
|
|
|
static const char hcd_name [] = "ehci_hcd";
|
|
|
|
|
|
#undef VERBOSE_DEBUG
|
|
#undef EHCI_URB_TRACE
|
|
|
|
#ifdef DEBUG
|
|
#define EHCI_STATS
|
|
#endif
|
|
|
|
/* magic numbers that can affect system performance */
|
|
#define EHCI_TUNE_CERR 3 /* 0-3 qtd retries; 0 == don't stop */
|
|
#define EHCI_TUNE_RL_HS 4 /* nak throttle; see 4.9 */
|
|
#define EHCI_TUNE_RL_TT 0
|
|
#define EHCI_TUNE_MULT_HS 1 /* 1-3 transactions/uframe; 4.10.3 */
|
|
#define EHCI_TUNE_MULT_TT 1
|
|
/*
|
|
* Some drivers think it's safe to schedule isochronous transfers more than
|
|
* 256 ms into the future (partly as a result of an old bug in the scheduling
|
|
* code). In an attempt to avoid trouble, we will use a minimum scheduling
|
|
* length of 512 frames instead of 256.
|
|
*/
|
|
#define EHCI_TUNE_FLS 1 /* (medium) 512-frame schedule */
|
|
|
|
#define EHCI_IAA_MSECS 10 /* arbitrary */
|
|
#define EHCI_IO_JIFFIES (HZ/10) /* io watchdog > irq_thresh */
|
|
#define EHCI_ASYNC_JIFFIES (HZ/20) /* async idle timeout */
|
|
#define EHCI_SHRINK_FRAMES 5 /* async qh unlink delay */
|
|
|
|
/* Initial IRQ latency: faster than hw default */
|
|
static int log2_irq_thresh = 0; // 0 to 6
|
|
module_param (log2_irq_thresh, int, S_IRUGO);
|
|
MODULE_PARM_DESC (log2_irq_thresh, "log2 IRQ latency, 1-64 microframes");
|
|
|
|
/* initial park setting: slower than hw default */
|
|
static unsigned park = 0;
|
|
module_param (park, uint, S_IRUGO);
|
|
MODULE_PARM_DESC (park, "park setting; 1-3 back-to-back async packets");
|
|
|
|
/* for flakey hardware, ignore overcurrent indicators */
|
|
static int ignore_oc = 0;
|
|
module_param (ignore_oc, bool, S_IRUGO);
|
|
MODULE_PARM_DESC (ignore_oc, "ignore bogus hardware overcurrent indications");
|
|
|
|
/* for link power management(LPM) feature */
|
|
static unsigned int hird;
|
|
module_param(hird, int, S_IRUGO);
|
|
MODULE_PARM_DESC(hird, "host initiated resume duration, +1 for each 75us\n");
|
|
|
|
#define INTR_MASK (STS_IAA | STS_FATAL | STS_PCD | STS_ERR | STS_INT)
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
#include "ehci.h"
|
|
#include "ehci-dbg.c"
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
static void
|
|
timer_action(struct ehci_hcd *ehci, enum ehci_timer_action action)
|
|
{
|
|
/* Don't override timeouts which shrink or (later) disable
|
|
* the async ring; just the I/O watchdog. Note that if a
|
|
* SHRINK were pending, OFF would never be requested.
|
|
*/
|
|
if (timer_pending(&ehci->watchdog)
|
|
&& ((BIT(TIMER_ASYNC_SHRINK) | BIT(TIMER_ASYNC_OFF))
|
|
& ehci->actions))
|
|
return;
|
|
|
|
if (!test_and_set_bit(action, &ehci->actions)) {
|
|
unsigned long t;
|
|
|
|
switch (action) {
|
|
case TIMER_IO_WATCHDOG:
|
|
if (!ehci->need_io_watchdog)
|
|
return;
|
|
t = EHCI_IO_JIFFIES;
|
|
break;
|
|
case TIMER_ASYNC_OFF:
|
|
t = EHCI_ASYNC_JIFFIES;
|
|
break;
|
|
/* case TIMER_ASYNC_SHRINK: */
|
|
default:
|
|
/* add a jiffie since we synch against the
|
|
* 8 KHz uframe counter.
|
|
*/
|
|
t = DIV_ROUND_UP(EHCI_SHRINK_FRAMES * HZ, 1000) + 1;
|
|
break;
|
|
}
|
|
mod_timer(&ehci->watchdog, t + jiffies);
|
|
}
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
/*
|
|
* handshake - spin reading hc until handshake completes or fails
|
|
* @ptr: address of hc register to be read
|
|
* @mask: bits to look at in result of read
|
|
* @done: value of those bits when handshake succeeds
|
|
* @usec: timeout in microseconds
|
|
*
|
|
* Returns negative errno, or zero on success
|
|
*
|
|
* Success happens when the "mask" bits have the specified value (hardware
|
|
* handshake done). There are two failure modes: "usec" have passed (major
|
|
* hardware flakeout), or the register reads as all-ones (hardware removed).
|
|
*
|
|
* That last failure should_only happen in cases like physical cardbus eject
|
|
* before driver shutdown. But it also seems to be caused by bugs in cardbus
|
|
* bridge shutdown: shutting down the bridge before the devices using it.
|
|
*/
|
|
static int handshake (struct ehci_hcd *ehci, void __iomem *ptr,
|
|
u32 mask, u32 done, int usec)
|
|
{
|
|
u32 result;
|
|
|
|
do {
|
|
result = ehci_readl(ehci, ptr);
|
|
if (result == ~(u32)0) /* card removed */
|
|
return -ENODEV;
|
|
result &= mask;
|
|
if (result == done)
|
|
return 0;
|
|
udelay (1);
|
|
usec--;
|
|
} while (usec > 0);
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
/* check TDI/ARC silicon is in host mode */
|
|
static int tdi_in_host_mode (struct ehci_hcd *ehci)
|
|
{
|
|
u32 __iomem *reg_ptr;
|
|
u32 tmp;
|
|
|
|
reg_ptr = (u32 __iomem *)(((u8 __iomem *)ehci->regs) + USBMODE);
|
|
tmp = ehci_readl(ehci, reg_ptr);
|
|
return (tmp & 3) == USBMODE_CM_HC;
|
|
}
|
|
|
|
/* force HC to halt state from unknown (EHCI spec section 2.3) */
|
|
static int ehci_halt (struct ehci_hcd *ehci)
|
|
{
|
|
u32 temp = ehci_readl(ehci, &ehci->regs->status);
|
|
|
|
/* disable any irqs left enabled by previous code */
|
|
ehci_writel(ehci, 0, &ehci->regs->intr_enable);
|
|
|
|
if (ehci_is_TDI(ehci) && tdi_in_host_mode(ehci) == 0) {
|
|
return 0;
|
|
}
|
|
|
|
if ((temp & STS_HALT) != 0)
|
|
return 0;
|
|
|
|
temp = ehci_readl(ehci, &ehci->regs->command);
|
|
temp &= ~CMD_RUN;
|
|
ehci_writel(ehci, temp, &ehci->regs->command);
|
|
return handshake (ehci, &ehci->regs->status,
|
|
STS_HALT, STS_HALT, 16 * 125);
|
|
}
|
|
|
|
static int handshake_on_error_set_halt(struct ehci_hcd *ehci, void __iomem *ptr,
|
|
u32 mask, u32 done, int usec)
|
|
{
|
|
int error;
|
|
|
|
error = handshake(ehci, ptr, mask, done, usec);
|
|
if (error) {
|
|
ehci_halt(ehci);
|
|
ehci_to_hcd(ehci)->state = HC_STATE_HALT;
|
|
ehci_err(ehci, "force halt; handshake %p %08x %08x -> %d\n",
|
|
ptr, mask, done, error);
|
|
}
|
|
|
|
return error;
|
|
}
|
|
|
|
/* put TDI/ARC silicon into EHCI mode */
|
|
static void tdi_reset (struct ehci_hcd *ehci)
|
|
{
|
|
u32 __iomem *reg_ptr;
|
|
u32 tmp;
|
|
|
|
reg_ptr = (u32 __iomem *)(((u8 __iomem *)ehci->regs) + USBMODE);
|
|
tmp = ehci_readl(ehci, reg_ptr);
|
|
tmp |= USBMODE_CM_HC;
|
|
/* The default byte access to MMR space is LE after
|
|
* controller reset. Set the required endian mode
|
|
* for transfer buffers to match the host microprocessor
|
|
*/
|
|
if (ehci_big_endian_mmio(ehci))
|
|
tmp |= USBMODE_BE;
|
|
ehci_writel(ehci, tmp, reg_ptr);
|
|
}
|
|
|
|
/* reset a non-running (STS_HALT == 1) controller */
|
|
static int ehci_reset (struct ehci_hcd *ehci)
|
|
{
|
|
int retval;
|
|
u32 command = ehci_readl(ehci, &ehci->regs->command);
|
|
|
|
/* If the EHCI debug controller is active, special care must be
|
|
* taken before and after a host controller reset */
|
|
if (ehci->debug && !dbgp_reset_prep())
|
|
ehci->debug = NULL;
|
|
|
|
command |= CMD_RESET;
|
|
dbg_cmd (ehci, "reset", command);
|
|
ehci_writel(ehci, command, &ehci->regs->command);
|
|
ehci_to_hcd(ehci)->state = HC_STATE_HALT;
|
|
ehci->next_statechange = jiffies;
|
|
retval = handshake (ehci, &ehci->regs->command,
|
|
CMD_RESET, 0, 250 * 1000);
|
|
|
|
if (ehci->has_hostpc) {
|
|
ehci_writel(ehci, USBMODE_EX_HC | USBMODE_EX_VBPS,
|
|
(u32 __iomem *)(((u8 *)ehci->regs) + USBMODE_EX));
|
|
ehci_writel(ehci, TXFIFO_DEFAULT,
|
|
(u32 __iomem *)(((u8 *)ehci->regs) + TXFILLTUNING));
|
|
}
|
|
if (retval)
|
|
return retval;
|
|
|
|
if (ehci_is_TDI(ehci))
|
|
tdi_reset (ehci);
|
|
|
|
if (ehci->debug)
|
|
dbgp_external_startup();
|
|
|
|
return retval;
|
|
}
|
|
|
|
/* idle the controller (from running) */
|
|
static void ehci_quiesce (struct ehci_hcd *ehci)
|
|
{
|
|
u32 temp;
|
|
|
|
#ifdef DEBUG
|
|
if (!HC_IS_RUNNING (ehci_to_hcd(ehci)->state))
|
|
BUG ();
|
|
#endif
|
|
|
|
/* wait for any schedule enables/disables to take effect */
|
|
temp = ehci_readl(ehci, &ehci->regs->command) << 10;
|
|
temp &= STS_ASS | STS_PSS;
|
|
if (handshake_on_error_set_halt(ehci, &ehci->regs->status,
|
|
STS_ASS | STS_PSS, temp, 16 * 125))
|
|
return;
|
|
|
|
/* then disable anything that's still active */
|
|
temp = ehci_readl(ehci, &ehci->regs->command);
|
|
temp &= ~(CMD_ASE | CMD_IAAD | CMD_PSE);
|
|
ehci_writel(ehci, temp, &ehci->regs->command);
|
|
|
|
/* hardware can take 16 microframes to turn off ... */
|
|
handshake_on_error_set_halt(ehci, &ehci->regs->status,
|
|
STS_ASS | STS_PSS, 0, 16 * 125);
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
static void end_unlink_async(struct ehci_hcd *ehci);
|
|
static void ehci_work(struct ehci_hcd *ehci);
|
|
|
|
#include "ehci-hub.c"
|
|
#include "ehci-lpm.c"
|
|
#include "ehci-mem.c"
|
|
#include "ehci-q.c"
|
|
#include "ehci-sched.c"
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
static void ehci_iaa_watchdog(unsigned long param)
|
|
{
|
|
struct ehci_hcd *ehci = (struct ehci_hcd *) param;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave (&ehci->lock, flags);
|
|
|
|
/* Lost IAA irqs wedge things badly; seen first with a vt8235.
|
|
* So we need this watchdog, but must protect it against both
|
|
* (a) SMP races against real IAA firing and retriggering, and
|
|
* (b) clean HC shutdown, when IAA watchdog was pending.
|
|
*/
|
|
if (ehci->reclaim
|
|
&& !timer_pending(&ehci->iaa_watchdog)
|
|
&& HC_IS_RUNNING(ehci_to_hcd(ehci)->state)) {
|
|
u32 cmd, status;
|
|
|
|
/* If we get here, IAA is *REALLY* late. It's barely
|
|
* conceivable that the system is so busy that CMD_IAAD
|
|
* is still legitimately set, so let's be sure it's
|
|
* clear before we read STS_IAA. (The HC should clear
|
|
* CMD_IAAD when it sets STS_IAA.)
|
|
*/
|
|
cmd = ehci_readl(ehci, &ehci->regs->command);
|
|
if (cmd & CMD_IAAD)
|
|
ehci_writel(ehci, cmd & ~CMD_IAAD,
|
|
&ehci->regs->command);
|
|
|
|
/* If IAA is set here it either legitimately triggered
|
|
* before we cleared IAAD above (but _way_ late, so we'll
|
|
* still count it as lost) ... or a silicon erratum:
|
|
* - VIA seems to set IAA without triggering the IRQ;
|
|
* - IAAD potentially cleared without setting IAA.
|
|
*/
|
|
status = ehci_readl(ehci, &ehci->regs->status);
|
|
if ((status & STS_IAA) || !(cmd & CMD_IAAD)) {
|
|
COUNT (ehci->stats.lost_iaa);
|
|
ehci_writel(ehci, STS_IAA, &ehci->regs->status);
|
|
}
|
|
|
|
ehci_vdbg(ehci, "IAA watchdog: status %x cmd %x\n",
|
|
status, cmd);
|
|
end_unlink_async(ehci);
|
|
}
|
|
|
|
spin_unlock_irqrestore(&ehci->lock, flags);
|
|
}
|
|
|
|
static void ehci_watchdog(unsigned long param)
|
|
{
|
|
struct ehci_hcd *ehci = (struct ehci_hcd *) param;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&ehci->lock, flags);
|
|
|
|
/* stop async processing after it's idled a bit */
|
|
if (test_bit (TIMER_ASYNC_OFF, &ehci->actions))
|
|
start_unlink_async (ehci, ehci->async);
|
|
|
|
/* ehci could run by timer, without IRQs ... */
|
|
ehci_work (ehci);
|
|
|
|
spin_unlock_irqrestore (&ehci->lock, flags);
|
|
}
|
|
|
|
/* On some systems, leaving remote wakeup enabled prevents system shutdown.
|
|
* The firmware seems to think that powering off is a wakeup event!
|
|
* This routine turns off remote wakeup and everything else, on all ports.
|
|
*/
|
|
static void ehci_turn_off_all_ports(struct ehci_hcd *ehci)
|
|
{
|
|
int port = HCS_N_PORTS(ehci->hcs_params);
|
|
|
|
while (port--)
|
|
ehci_writel(ehci, PORT_RWC_BITS,
|
|
&ehci->regs->port_status[port]);
|
|
}
|
|
|
|
/*
|
|
* Halt HC, turn off all ports, and let the BIOS use the companion controllers.
|
|
* Should be called with ehci->lock held.
|
|
*/
|
|
static void ehci_silence_controller(struct ehci_hcd *ehci)
|
|
{
|
|
ehci_halt(ehci);
|
|
ehci_turn_off_all_ports(ehci);
|
|
|
|
/* make BIOS/etc use companion controller during reboot */
|
|
ehci_writel(ehci, 0, &ehci->regs->configured_flag);
|
|
|
|
/* unblock posted writes */
|
|
ehci_readl(ehci, &ehci->regs->configured_flag);
|
|
}
|
|
|
|
/* ehci_shutdown kick in for silicon on any bus (not just pci, etc).
|
|
* This forcibly disables dma and IRQs, helping kexec and other cases
|
|
* where the next system software may expect clean state.
|
|
*/
|
|
static void ehci_shutdown(struct usb_hcd *hcd)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci(hcd);
|
|
|
|
del_timer_sync(&ehci->watchdog);
|
|
del_timer_sync(&ehci->iaa_watchdog);
|
|
|
|
spin_lock_irq(&ehci->lock);
|
|
ehci_silence_controller(ehci);
|
|
spin_unlock_irq(&ehci->lock);
|
|
}
|
|
|
|
static void ehci_port_power (struct ehci_hcd *ehci, int is_on)
|
|
{
|
|
unsigned port;
|
|
|
|
if (!HCS_PPC (ehci->hcs_params))
|
|
return;
|
|
|
|
ehci_dbg (ehci, "...power%s ports...\n", is_on ? "up" : "down");
|
|
for (port = HCS_N_PORTS (ehci->hcs_params); port > 0; )
|
|
(void) ehci_hub_control(ehci_to_hcd(ehci),
|
|
is_on ? SetPortFeature : ClearPortFeature,
|
|
USB_PORT_FEAT_POWER,
|
|
port--, NULL, 0);
|
|
/* Flush those writes */
|
|
ehci_readl(ehci, &ehci->regs->command);
|
|
msleep(20);
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
/*
|
|
* ehci_work is called from some interrupts, timers, and so on.
|
|
* it calls driver completion functions, after dropping ehci->lock.
|
|
*/
|
|
static void ehci_work (struct ehci_hcd *ehci)
|
|
{
|
|
timer_action_done (ehci, TIMER_IO_WATCHDOG);
|
|
|
|
/* another CPU may drop ehci->lock during a schedule scan while
|
|
* it reports urb completions. this flag guards against bogus
|
|
* attempts at re-entrant schedule scanning.
|
|
*/
|
|
if (ehci->scanning)
|
|
return;
|
|
ehci->scanning = 1;
|
|
scan_async (ehci);
|
|
if (ehci->next_uframe != -1)
|
|
scan_periodic (ehci);
|
|
ehci->scanning = 0;
|
|
|
|
/* the IO watchdog guards against hardware or driver bugs that
|
|
* misplace IRQs, and should let us run completely without IRQs.
|
|
* such lossage has been observed on both VT6202 and VT8235.
|
|
*/
|
|
if (HC_IS_RUNNING (ehci_to_hcd(ehci)->state) &&
|
|
(ehci->async->qh_next.ptr != NULL ||
|
|
ehci->periodic_sched != 0))
|
|
timer_action (ehci, TIMER_IO_WATCHDOG);
|
|
}
|
|
|
|
/*
|
|
* Called when the ehci_hcd module is removed.
|
|
*/
|
|
static void ehci_stop (struct usb_hcd *hcd)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
|
|
ehci_dbg (ehci, "stop\n");
|
|
|
|
/* no more interrupts ... */
|
|
del_timer_sync (&ehci->watchdog);
|
|
del_timer_sync(&ehci->iaa_watchdog);
|
|
|
|
spin_lock_irq(&ehci->lock);
|
|
if (HC_IS_RUNNING (hcd->state))
|
|
ehci_quiesce (ehci);
|
|
|
|
ehci_silence_controller(ehci);
|
|
ehci_reset (ehci);
|
|
spin_unlock_irq(&ehci->lock);
|
|
|
|
remove_companion_file(ehci);
|
|
remove_debug_files (ehci);
|
|
|
|
/* root hub is shut down separately (first, when possible) */
|
|
spin_lock_irq (&ehci->lock);
|
|
if (ehci->async)
|
|
ehci_work (ehci);
|
|
spin_unlock_irq (&ehci->lock);
|
|
ehci_mem_cleanup (ehci);
|
|
|
|
#ifdef EHCI_STATS
|
|
ehci_dbg (ehci, "irq normal %ld err %ld reclaim %ld (lost %ld)\n",
|
|
ehci->stats.normal, ehci->stats.error, ehci->stats.reclaim,
|
|
ehci->stats.lost_iaa);
|
|
ehci_dbg (ehci, "complete %ld unlink %ld\n",
|
|
ehci->stats.complete, ehci->stats.unlink);
|
|
#endif
|
|
|
|
dbg_status (ehci, "ehci_stop completed",
|
|
ehci_readl(ehci, &ehci->regs->status));
|
|
}
|
|
|
|
/* one-time init, only for memory state */
|
|
static int ehci_init(struct usb_hcd *hcd)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci(hcd);
|
|
u32 temp;
|
|
int retval;
|
|
u32 hcc_params;
|
|
struct ehci_qh_hw *hw;
|
|
|
|
spin_lock_init(&ehci->lock);
|
|
|
|
/*
|
|
* keep io watchdog by default, those good HCDs could turn off it later
|
|
*/
|
|
ehci->need_io_watchdog = 1;
|
|
init_timer(&ehci->watchdog);
|
|
ehci->watchdog.function = ehci_watchdog;
|
|
ehci->watchdog.data = (unsigned long) ehci;
|
|
|
|
init_timer(&ehci->iaa_watchdog);
|
|
ehci->iaa_watchdog.function = ehci_iaa_watchdog;
|
|
ehci->iaa_watchdog.data = (unsigned long) ehci;
|
|
|
|
/*
|
|
* hw default: 1K periodic list heads, one per frame.
|
|
* periodic_size can shrink by USBCMD update if hcc_params allows.
|
|
*/
|
|
ehci->periodic_size = DEFAULT_I_TDPS;
|
|
INIT_LIST_HEAD(&ehci->cached_itd_list);
|
|
INIT_LIST_HEAD(&ehci->cached_sitd_list);
|
|
if ((retval = ehci_mem_init(ehci, GFP_KERNEL)) < 0)
|
|
return retval;
|
|
|
|
/* controllers may cache some of the periodic schedule ... */
|
|
hcc_params = ehci_readl(ehci, &ehci->caps->hcc_params);
|
|
if (HCC_ISOC_CACHE(hcc_params)) // full frame cache
|
|
ehci->i_thresh = 2 + 8;
|
|
else // N microframes cached
|
|
ehci->i_thresh = 2 + HCC_ISOC_THRES(hcc_params);
|
|
|
|
ehci->reclaim = NULL;
|
|
ehci->next_uframe = -1;
|
|
ehci->clock_frame = -1;
|
|
|
|
/*
|
|
* dedicate a qh for the async ring head, since we couldn't unlink
|
|
* a 'real' qh without stopping the async schedule [4.8]. use it
|
|
* as the 'reclamation list head' too.
|
|
* its dummy is used in hw_alt_next of many tds, to prevent the qh
|
|
* from automatically advancing to the next td after short reads.
|
|
*/
|
|
ehci->async->qh_next.qh = NULL;
|
|
hw = ehci->async->hw;
|
|
hw->hw_next = QH_NEXT(ehci, ehci->async->qh_dma);
|
|
hw->hw_info1 = cpu_to_hc32(ehci, QH_HEAD);
|
|
hw->hw_token = cpu_to_hc32(ehci, QTD_STS_HALT);
|
|
hw->hw_qtd_next = EHCI_LIST_END(ehci);
|
|
ehci->async->qh_state = QH_STATE_LINKED;
|
|
hw->hw_alt_next = QTD_NEXT(ehci, ehci->async->dummy->qtd_dma);
|
|
|
|
/* clear interrupt enables, set irq latency */
|
|
if (log2_irq_thresh < 0 || log2_irq_thresh > 6)
|
|
log2_irq_thresh = 0;
|
|
temp = 1 << (16 + log2_irq_thresh);
|
|
if (HCC_PER_PORT_CHANGE_EVENT(hcc_params)) {
|
|
ehci->has_ppcd = 1;
|
|
ehci_dbg(ehci, "enable per-port change event\n");
|
|
temp |= CMD_PPCEE;
|
|
}
|
|
if (HCC_CANPARK(hcc_params)) {
|
|
/* HW default park == 3, on hardware that supports it (like
|
|
* NVidia and ALI silicon), maximizes throughput on the async
|
|
* schedule by avoiding QH fetches between transfers.
|
|
*
|
|
* With fast usb storage devices and NForce2, "park" seems to
|
|
* make problems: throughput reduction (!), data errors...
|
|
*/
|
|
if (park) {
|
|
park = min(park, (unsigned) 3);
|
|
temp |= CMD_PARK;
|
|
temp |= park << 8;
|
|
}
|
|
ehci_dbg(ehci, "park %d\n", park);
|
|
}
|
|
if (HCC_PGM_FRAMELISTLEN(hcc_params)) {
|
|
/* periodic schedule size can be smaller than default */
|
|
temp &= ~(3 << 2);
|
|
temp |= (EHCI_TUNE_FLS << 2);
|
|
switch (EHCI_TUNE_FLS) {
|
|
case 0: ehci->periodic_size = 1024; break;
|
|
case 1: ehci->periodic_size = 512; break;
|
|
case 2: ehci->periodic_size = 256; break;
|
|
default: BUG();
|
|
}
|
|
}
|
|
if (HCC_LPM(hcc_params)) {
|
|
/* support link power management EHCI 1.1 addendum */
|
|
ehci_dbg(ehci, "support lpm\n");
|
|
ehci->has_lpm = 1;
|
|
if (hird > 0xf) {
|
|
ehci_dbg(ehci, "hird %d invalid, use default 0",
|
|
hird);
|
|
hird = 0;
|
|
}
|
|
temp |= hird << 24;
|
|
}
|
|
ehci->command = temp;
|
|
|
|
/* Accept arbitrarily long scatter-gather lists */
|
|
if (!(hcd->driver->flags & HCD_LOCAL_MEM))
|
|
hcd->self.sg_tablesize = ~0;
|
|
return 0;
|
|
}
|
|
|
|
/* start HC running; it's halted, ehci_init() has been run (once) */
|
|
static int ehci_run (struct usb_hcd *hcd)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
int retval;
|
|
u32 temp;
|
|
u32 hcc_params;
|
|
|
|
hcd->uses_new_polling = 1;
|
|
|
|
/* EHCI spec section 4.1 */
|
|
if ((retval = ehci_reset(ehci)) != 0) {
|
|
ehci_mem_cleanup(ehci);
|
|
return retval;
|
|
}
|
|
ehci_writel(ehci, ehci->periodic_dma, &ehci->regs->frame_list);
|
|
ehci_writel(ehci, (u32)ehci->async->qh_dma, &ehci->regs->async_next);
|
|
|
|
/*
|
|
* hcc_params controls whether ehci->regs->segment must (!!!)
|
|
* be used; it constrains QH/ITD/SITD and QTD locations.
|
|
* pci_pool consistent memory always uses segment zero.
|
|
* streaming mappings for I/O buffers, like pci_map_single(),
|
|
* can return segments above 4GB, if the device allows.
|
|
*
|
|
* NOTE: the dma mask is visible through dma_supported(), so
|
|
* drivers can pass this info along ... like NETIF_F_HIGHDMA,
|
|
* Scsi_Host.highmem_io, and so forth. It's readonly to all
|
|
* host side drivers though.
|
|
*/
|
|
hcc_params = ehci_readl(ehci, &ehci->caps->hcc_params);
|
|
if (HCC_64BIT_ADDR(hcc_params)) {
|
|
ehci_writel(ehci, 0, &ehci->regs->segment);
|
|
#if 0
|
|
// this is deeply broken on almost all architectures
|
|
if (!dma_set_mask(hcd->self.controller, DMA_BIT_MASK(64)))
|
|
ehci_info(ehci, "enabled 64bit DMA\n");
|
|
#endif
|
|
}
|
|
|
|
|
|
// Philips, Intel, and maybe others need CMD_RUN before the
|
|
// root hub will detect new devices (why?); NEC doesn't
|
|
ehci->command &= ~(CMD_LRESET|CMD_IAAD|CMD_PSE|CMD_ASE|CMD_RESET);
|
|
ehci->command |= CMD_RUN;
|
|
ehci_writel(ehci, ehci->command, &ehci->regs->command);
|
|
dbg_cmd (ehci, "init", ehci->command);
|
|
|
|
/*
|
|
* Start, enabling full USB 2.0 functionality ... usb 1.1 devices
|
|
* are explicitly handed to companion controller(s), so no TT is
|
|
* involved with the root hub. (Except where one is integrated,
|
|
* and there's no companion controller unless maybe for USB OTG.)
|
|
*
|
|
* Turning on the CF flag will transfer ownership of all ports
|
|
* from the companions to the EHCI controller. If any of the
|
|
* companions are in the middle of a port reset at the time, it
|
|
* could cause trouble. Write-locking ehci_cf_port_reset_rwsem
|
|
* guarantees that no resets are in progress. After we set CF,
|
|
* a short delay lets the hardware catch up; new resets shouldn't
|
|
* be started before the port switching actions could complete.
|
|
*/
|
|
down_write(&ehci_cf_port_reset_rwsem);
|
|
hcd->state = HC_STATE_RUNNING;
|
|
ehci_writel(ehci, FLAG_CF, &ehci->regs->configured_flag);
|
|
ehci_readl(ehci, &ehci->regs->command); /* unblock posted writes */
|
|
msleep(5);
|
|
up_write(&ehci_cf_port_reset_rwsem);
|
|
ehci->last_periodic_enable = ktime_get_real();
|
|
|
|
temp = HC_VERSION(ehci_readl(ehci, &ehci->caps->hc_capbase));
|
|
ehci_info (ehci,
|
|
"USB %x.%x started, EHCI %x.%02x%s\n",
|
|
((ehci->sbrn & 0xf0)>>4), (ehci->sbrn & 0x0f),
|
|
temp >> 8, temp & 0xff,
|
|
ignore_oc ? ", overcurrent ignored" : "");
|
|
|
|
ehci_writel(ehci, INTR_MASK,
|
|
&ehci->regs->intr_enable); /* Turn On Interrupts */
|
|
|
|
/* GRR this is run-once init(), being done every time the HC starts.
|
|
* So long as they're part of class devices, we can't do it init()
|
|
* since the class device isn't created that early.
|
|
*/
|
|
create_debug_files(ehci);
|
|
create_companion_file(ehci);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
static irqreturn_t ehci_irq (struct usb_hcd *hcd)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
u32 status, masked_status, pcd_status = 0, cmd;
|
|
int bh;
|
|
|
|
spin_lock (&ehci->lock);
|
|
|
|
status = ehci_readl(ehci, &ehci->regs->status);
|
|
|
|
/* e.g. cardbus physical eject */
|
|
if (status == ~(u32) 0) {
|
|
ehci_dbg (ehci, "device removed\n");
|
|
goto dead;
|
|
}
|
|
|
|
masked_status = status & INTR_MASK;
|
|
if (!masked_status) { /* irq sharing? */
|
|
spin_unlock(&ehci->lock);
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
/* clear (just) interrupts */
|
|
ehci_writel(ehci, masked_status, &ehci->regs->status);
|
|
cmd = ehci_readl(ehci, &ehci->regs->command);
|
|
bh = 0;
|
|
|
|
#ifdef VERBOSE_DEBUG
|
|
/* unrequested/ignored: Frame List Rollover */
|
|
dbg_status (ehci, "irq", status);
|
|
#endif
|
|
|
|
/* INT, ERR, and IAA interrupt rates can be throttled */
|
|
|
|
/* normal [4.15.1.2] or error [4.15.1.1] completion */
|
|
if (likely ((status & (STS_INT|STS_ERR)) != 0)) {
|
|
if (likely ((status & STS_ERR) == 0))
|
|
COUNT (ehci->stats.normal);
|
|
else
|
|
COUNT (ehci->stats.error);
|
|
bh = 1;
|
|
}
|
|
|
|
/* complete the unlinking of some qh [4.15.2.3] */
|
|
if (status & STS_IAA) {
|
|
/* guard against (alleged) silicon errata */
|
|
if (cmd & CMD_IAAD) {
|
|
ehci_writel(ehci, cmd & ~CMD_IAAD,
|
|
&ehci->regs->command);
|
|
ehci_dbg(ehci, "IAA with IAAD still set?\n");
|
|
}
|
|
if (ehci->reclaim) {
|
|
COUNT(ehci->stats.reclaim);
|
|
end_unlink_async(ehci);
|
|
} else
|
|
ehci_dbg(ehci, "IAA with nothing to reclaim?\n");
|
|
}
|
|
|
|
/* remote wakeup [4.3.1] */
|
|
if (status & STS_PCD) {
|
|
unsigned i = HCS_N_PORTS (ehci->hcs_params);
|
|
u32 ppcd = 0;
|
|
|
|
/* kick root hub later */
|
|
pcd_status = status;
|
|
|
|
/* resume root hub? */
|
|
if (!(cmd & CMD_RUN))
|
|
usb_hcd_resume_root_hub(hcd);
|
|
|
|
/* get per-port change detect bits */
|
|
if (ehci->has_ppcd)
|
|
ppcd = status >> 16;
|
|
|
|
while (i--) {
|
|
int pstatus;
|
|
|
|
/* leverage per-port change bits feature */
|
|
if (ehci->has_ppcd && !(ppcd & (1 << i)))
|
|
continue;
|
|
pstatus = ehci_readl(ehci,
|
|
&ehci->regs->port_status[i]);
|
|
|
|
if (pstatus & PORT_OWNER)
|
|
continue;
|
|
if (!(test_bit(i, &ehci->suspended_ports) &&
|
|
((pstatus & PORT_RESUME) ||
|
|
!(pstatus & PORT_SUSPEND)) &&
|
|
(pstatus & PORT_PE) &&
|
|
ehci->reset_done[i] == 0))
|
|
continue;
|
|
|
|
/* start 20 msec resume signaling from this port,
|
|
* and make khubd collect PORT_STAT_C_SUSPEND to
|
|
* stop that signaling. Use 5 ms extra for safety,
|
|
* like usb_port_resume() does.
|
|
*/
|
|
ehci->reset_done[i] = jiffies + msecs_to_jiffies(25);
|
|
ehci_dbg (ehci, "port %d remote wakeup\n", i + 1);
|
|
mod_timer(&hcd->rh_timer, ehci->reset_done[i]);
|
|
}
|
|
}
|
|
|
|
/* PCI errors [4.15.2.4] */
|
|
if (unlikely ((status & STS_FATAL) != 0)) {
|
|
ehci_err(ehci, "fatal error\n");
|
|
dbg_cmd(ehci, "fatal", cmd);
|
|
dbg_status(ehci, "fatal", status);
|
|
ehci_halt(ehci);
|
|
dead:
|
|
ehci_reset(ehci);
|
|
ehci_writel(ehci, 0, &ehci->regs->configured_flag);
|
|
/* generic layer kills/unlinks all urbs, then
|
|
* uses ehci_stop to clean up the rest
|
|
*/
|
|
bh = 1;
|
|
}
|
|
|
|
if (bh)
|
|
ehci_work (ehci);
|
|
spin_unlock (&ehci->lock);
|
|
if (pcd_status)
|
|
usb_hcd_poll_rh_status(hcd);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
/*
|
|
* non-error returns are a promise to giveback() the urb later
|
|
* we drop ownership so next owner (or urb unlink) can get it
|
|
*
|
|
* urb + dev is in hcd.self.controller.urb_list
|
|
* we're queueing TDs onto software and hardware lists
|
|
*
|
|
* hcd-specific init for hcpriv hasn't been done yet
|
|
*
|
|
* NOTE: control, bulk, and interrupt share the same code to append TDs
|
|
* to a (possibly active) QH, and the same QH scanning code.
|
|
*/
|
|
static int ehci_urb_enqueue (
|
|
struct usb_hcd *hcd,
|
|
struct urb *urb,
|
|
gfp_t mem_flags
|
|
) {
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
struct list_head qtd_list;
|
|
|
|
INIT_LIST_HEAD (&qtd_list);
|
|
|
|
switch (usb_pipetype (urb->pipe)) {
|
|
case PIPE_CONTROL:
|
|
/* qh_completions() code doesn't handle all the fault cases
|
|
* in multi-TD control transfers. Even 1KB is rare anyway.
|
|
*/
|
|
if (urb->transfer_buffer_length > (16 * 1024))
|
|
return -EMSGSIZE;
|
|
/* FALLTHROUGH */
|
|
/* case PIPE_BULK: */
|
|
default:
|
|
if (!qh_urb_transaction (ehci, urb, &qtd_list, mem_flags))
|
|
return -ENOMEM;
|
|
return submit_async(ehci, urb, &qtd_list, mem_flags);
|
|
|
|
case PIPE_INTERRUPT:
|
|
if (!qh_urb_transaction (ehci, urb, &qtd_list, mem_flags))
|
|
return -ENOMEM;
|
|
return intr_submit(ehci, urb, &qtd_list, mem_flags);
|
|
|
|
case PIPE_ISOCHRONOUS:
|
|
if (urb->dev->speed == USB_SPEED_HIGH)
|
|
return itd_submit (ehci, urb, mem_flags);
|
|
else
|
|
return sitd_submit (ehci, urb, mem_flags);
|
|
}
|
|
}
|
|
|
|
static void unlink_async (struct ehci_hcd *ehci, struct ehci_qh *qh)
|
|
{
|
|
/* failfast */
|
|
if (!HC_IS_RUNNING(ehci_to_hcd(ehci)->state) && ehci->reclaim)
|
|
end_unlink_async(ehci);
|
|
|
|
/* If the QH isn't linked then there's nothing we can do
|
|
* unless we were called during a giveback, in which case
|
|
* qh_completions() has to deal with it.
|
|
*/
|
|
if (qh->qh_state != QH_STATE_LINKED) {
|
|
if (qh->qh_state == QH_STATE_COMPLETING)
|
|
qh->needs_rescan = 1;
|
|
return;
|
|
}
|
|
|
|
/* defer till later if busy */
|
|
if (ehci->reclaim) {
|
|
struct ehci_qh *last;
|
|
|
|
for (last = ehci->reclaim;
|
|
last->reclaim;
|
|
last = last->reclaim)
|
|
continue;
|
|
qh->qh_state = QH_STATE_UNLINK_WAIT;
|
|
last->reclaim = qh;
|
|
|
|
/* start IAA cycle */
|
|
} else
|
|
start_unlink_async (ehci, qh);
|
|
}
|
|
|
|
/* remove from hardware lists
|
|
* completions normally happen asynchronously
|
|
*/
|
|
|
|
static int ehci_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
struct ehci_qh *qh;
|
|
unsigned long flags;
|
|
int rc;
|
|
|
|
spin_lock_irqsave (&ehci->lock, flags);
|
|
rc = usb_hcd_check_unlink_urb(hcd, urb, status);
|
|
if (rc)
|
|
goto done;
|
|
|
|
switch (usb_pipetype (urb->pipe)) {
|
|
// case PIPE_CONTROL:
|
|
// case PIPE_BULK:
|
|
default:
|
|
qh = (struct ehci_qh *) urb->hcpriv;
|
|
if (!qh)
|
|
break;
|
|
switch (qh->qh_state) {
|
|
case QH_STATE_LINKED:
|
|
case QH_STATE_COMPLETING:
|
|
unlink_async(ehci, qh);
|
|
break;
|
|
case QH_STATE_UNLINK:
|
|
case QH_STATE_UNLINK_WAIT:
|
|
/* already started */
|
|
break;
|
|
case QH_STATE_IDLE:
|
|
/* QH might be waiting for a Clear-TT-Buffer */
|
|
qh_completions(ehci, qh);
|
|
break;
|
|
}
|
|
break;
|
|
|
|
case PIPE_INTERRUPT:
|
|
qh = (struct ehci_qh *) urb->hcpriv;
|
|
if (!qh)
|
|
break;
|
|
switch (qh->qh_state) {
|
|
case QH_STATE_LINKED:
|
|
case QH_STATE_COMPLETING:
|
|
intr_deschedule (ehci, qh);
|
|
break;
|
|
case QH_STATE_IDLE:
|
|
qh_completions (ehci, qh);
|
|
break;
|
|
default:
|
|
ehci_dbg (ehci, "bogus qh %p state %d\n",
|
|
qh, qh->qh_state);
|
|
goto done;
|
|
}
|
|
break;
|
|
|
|
case PIPE_ISOCHRONOUS:
|
|
// itd or sitd ...
|
|
|
|
// wait till next completion, do it then.
|
|
// completion irqs can wait up to 1024 msec,
|
|
break;
|
|
}
|
|
done:
|
|
spin_unlock_irqrestore (&ehci->lock, flags);
|
|
return rc;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
// bulk qh holds the data toggle
|
|
|
|
static void
|
|
ehci_endpoint_disable (struct usb_hcd *hcd, struct usb_host_endpoint *ep)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
unsigned long flags;
|
|
struct ehci_qh *qh, *tmp;
|
|
|
|
/* ASSERT: any requests/urbs are being unlinked */
|
|
/* ASSERT: nobody can be submitting urbs for this any more */
|
|
|
|
rescan:
|
|
spin_lock_irqsave (&ehci->lock, flags);
|
|
qh = ep->hcpriv;
|
|
if (!qh)
|
|
goto done;
|
|
|
|
/* endpoints can be iso streams. for now, we don't
|
|
* accelerate iso completions ... so spin a while.
|
|
*/
|
|
if (qh->hw == NULL) {
|
|
ehci_vdbg (ehci, "iso delay\n");
|
|
goto idle_timeout;
|
|
}
|
|
|
|
if (!HC_IS_RUNNING (hcd->state))
|
|
qh->qh_state = QH_STATE_IDLE;
|
|
switch (qh->qh_state) {
|
|
case QH_STATE_LINKED:
|
|
case QH_STATE_COMPLETING:
|
|
for (tmp = ehci->async->qh_next.qh;
|
|
tmp && tmp != qh;
|
|
tmp = tmp->qh_next.qh)
|
|
continue;
|
|
/* periodic qh self-unlinks on empty */
|
|
if (!tmp)
|
|
goto nogood;
|
|
unlink_async (ehci, qh);
|
|
/* FALL THROUGH */
|
|
case QH_STATE_UNLINK: /* wait for hw to finish? */
|
|
case QH_STATE_UNLINK_WAIT:
|
|
idle_timeout:
|
|
spin_unlock_irqrestore (&ehci->lock, flags);
|
|
schedule_timeout_uninterruptible(1);
|
|
goto rescan;
|
|
case QH_STATE_IDLE: /* fully unlinked */
|
|
if (qh->clearing_tt)
|
|
goto idle_timeout;
|
|
if (list_empty (&qh->qtd_list)) {
|
|
qh_put (qh);
|
|
break;
|
|
}
|
|
/* else FALL THROUGH */
|
|
default:
|
|
nogood:
|
|
/* caller was supposed to have unlinked any requests;
|
|
* that's not our job. just leak this memory.
|
|
*/
|
|
ehci_err (ehci, "qh %p (#%02x) state %d%s\n",
|
|
qh, ep->desc.bEndpointAddress, qh->qh_state,
|
|
list_empty (&qh->qtd_list) ? "" : "(has tds)");
|
|
break;
|
|
}
|
|
ep->hcpriv = NULL;
|
|
done:
|
|
spin_unlock_irqrestore (&ehci->lock, flags);
|
|
}
|
|
|
|
static void
|
|
ehci_endpoint_reset(struct usb_hcd *hcd, struct usb_host_endpoint *ep)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci(hcd);
|
|
struct ehci_qh *qh;
|
|
int eptype = usb_endpoint_type(&ep->desc);
|
|
int epnum = usb_endpoint_num(&ep->desc);
|
|
int is_out = usb_endpoint_dir_out(&ep->desc);
|
|
unsigned long flags;
|
|
|
|
if (eptype != USB_ENDPOINT_XFER_BULK && eptype != USB_ENDPOINT_XFER_INT)
|
|
return;
|
|
|
|
spin_lock_irqsave(&ehci->lock, flags);
|
|
qh = ep->hcpriv;
|
|
|
|
/* For Bulk and Interrupt endpoints we maintain the toggle state
|
|
* in the hardware; the toggle bits in udev aren't used at all.
|
|
* When an endpoint is reset by usb_clear_halt() we must reset
|
|
* the toggle bit in the QH.
|
|
*/
|
|
if (qh) {
|
|
usb_settoggle(qh->dev, epnum, is_out, 0);
|
|
if (!list_empty(&qh->qtd_list)) {
|
|
WARN_ONCE(1, "clear_halt for a busy endpoint\n");
|
|
} else if (qh->qh_state == QH_STATE_LINKED ||
|
|
qh->qh_state == QH_STATE_COMPLETING) {
|
|
|
|
/* The toggle value in the QH can't be updated
|
|
* while the QH is active. Unlink it now;
|
|
* re-linking will call qh_refresh().
|
|
*/
|
|
if (eptype == USB_ENDPOINT_XFER_BULK)
|
|
unlink_async(ehci, qh);
|
|
else
|
|
intr_deschedule(ehci, qh);
|
|
}
|
|
}
|
|
spin_unlock_irqrestore(&ehci->lock, flags);
|
|
}
|
|
|
|
static int ehci_get_frame (struct usb_hcd *hcd)
|
|
{
|
|
struct ehci_hcd *ehci = hcd_to_ehci (hcd);
|
|
return (ehci_readl(ehci, &ehci->regs->frame_index) >> 3) %
|
|
ehci->periodic_size;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------*/
|
|
|
|
MODULE_DESCRIPTION(DRIVER_DESC);
|
|
MODULE_AUTHOR (DRIVER_AUTHOR);
|
|
MODULE_LICENSE ("GPL");
|
|
|
|
#ifdef CONFIG_PCI
|
|
#include "ehci-pci.c"
|
|
#define PCI_DRIVER ehci_pci_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_USB_EHCI_FSL
|
|
#include "ehci-fsl.c"
|
|
#define PLATFORM_DRIVER ehci_fsl_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_USB_EHCI_MXC
|
|
#include "ehci-mxc.c"
|
|
#define PLATFORM_DRIVER ehci_mxc_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_SOC_AU1200
|
|
#include "ehci-au1xxx.c"
|
|
#define PLATFORM_DRIVER ehci_hcd_au1xxx_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARCH_OMAP3
|
|
#include "ehci-omap.c"
|
|
#define PLATFORM_DRIVER ehci_hcd_omap_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC_PS3
|
|
#include "ehci-ps3.c"
|
|
#define PS3_SYSTEM_BUS_DRIVER ps3_ehci_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_USB_EHCI_HCD_PPC_OF
|
|
#include "ehci-ppc-of.c"
|
|
#define OF_PLATFORM_DRIVER ehci_hcd_ppc_of_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_XPS_USB_HCD_XILINX
|
|
#include "ehci-xilinx-of.c"
|
|
#define XILINX_OF_PLATFORM_DRIVER ehci_hcd_xilinx_of_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_PLAT_ORION
|
|
#include "ehci-orion.c"
|
|
#define PLATFORM_DRIVER ehci_orion_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARCH_IXP4XX
|
|
#include "ehci-ixp4xx.c"
|
|
#define PLATFORM_DRIVER ixp4xx_ehci_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_USB_W90X900_EHCI
|
|
#include "ehci-w90x900.c"
|
|
#define PLATFORM_DRIVER ehci_hcd_w90x900_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARCH_AT91
|
|
#include "ehci-atmel.c"
|
|
#define PLATFORM_DRIVER ehci_atmel_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_USB_OCTEON_EHCI
|
|
#include "ehci-octeon.c"
|
|
#define PLATFORM_DRIVER ehci_octeon_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARCH_VT8500
|
|
#include "ehci-vt8500.c"
|
|
#define PLATFORM_DRIVER vt8500_ehci_driver
|
|
#endif
|
|
|
|
#ifdef CONFIG_PLAT_SPEAR
|
|
#include "ehci-spear.c"
|
|
#define PLATFORM_DRIVER spear_ehci_hcd_driver
|
|
#endif
|
|
|
|
#if !defined(PCI_DRIVER) && !defined(PLATFORM_DRIVER) && \
|
|
!defined(PS3_SYSTEM_BUS_DRIVER) && !defined(OF_PLATFORM_DRIVER) && \
|
|
!defined(XILINX_OF_PLATFORM_DRIVER)
|
|
#error "missing bus glue for ehci-hcd"
|
|
#endif
|
|
|
|
static int __init ehci_hcd_init(void)
|
|
{
|
|
int retval = 0;
|
|
|
|
if (usb_disabled())
|
|
return -ENODEV;
|
|
|
|
printk(KERN_INFO "%s: " DRIVER_DESC "\n", hcd_name);
|
|
set_bit(USB_EHCI_LOADED, &usb_hcds_loaded);
|
|
if (test_bit(USB_UHCI_LOADED, &usb_hcds_loaded) ||
|
|
test_bit(USB_OHCI_LOADED, &usb_hcds_loaded))
|
|
printk(KERN_WARNING "Warning! ehci_hcd should always be loaded"
|
|
" before uhci_hcd and ohci_hcd, not after\n");
|
|
|
|
pr_debug("%s: block sizes: qh %Zd qtd %Zd itd %Zd sitd %Zd\n",
|
|
hcd_name,
|
|
sizeof(struct ehci_qh), sizeof(struct ehci_qtd),
|
|
sizeof(struct ehci_itd), sizeof(struct ehci_sitd));
|
|
|
|
#ifdef DEBUG
|
|
ehci_debug_root = debugfs_create_dir("ehci", usb_debug_root);
|
|
if (!ehci_debug_root) {
|
|
retval = -ENOENT;
|
|
goto err_debug;
|
|
}
|
|
#endif
|
|
|
|
#ifdef PLATFORM_DRIVER
|
|
retval = platform_driver_register(&PLATFORM_DRIVER);
|
|
if (retval < 0)
|
|
goto clean0;
|
|
#endif
|
|
|
|
#ifdef PCI_DRIVER
|
|
retval = pci_register_driver(&PCI_DRIVER);
|
|
if (retval < 0)
|
|
goto clean1;
|
|
#endif
|
|
|
|
#ifdef PS3_SYSTEM_BUS_DRIVER
|
|
retval = ps3_ehci_driver_register(&PS3_SYSTEM_BUS_DRIVER);
|
|
if (retval < 0)
|
|
goto clean2;
|
|
#endif
|
|
|
|
#ifdef OF_PLATFORM_DRIVER
|
|
retval = of_register_platform_driver(&OF_PLATFORM_DRIVER);
|
|
if (retval < 0)
|
|
goto clean3;
|
|
#endif
|
|
|
|
#ifdef XILINX_OF_PLATFORM_DRIVER
|
|
retval = of_register_platform_driver(&XILINX_OF_PLATFORM_DRIVER);
|
|
if (retval < 0)
|
|
goto clean4;
|
|
#endif
|
|
return retval;
|
|
|
|
#ifdef XILINX_OF_PLATFORM_DRIVER
|
|
/* of_unregister_platform_driver(&XILINX_OF_PLATFORM_DRIVER); */
|
|
clean4:
|
|
#endif
|
|
#ifdef OF_PLATFORM_DRIVER
|
|
of_unregister_platform_driver(&OF_PLATFORM_DRIVER);
|
|
clean3:
|
|
#endif
|
|
#ifdef PS3_SYSTEM_BUS_DRIVER
|
|
ps3_ehci_driver_unregister(&PS3_SYSTEM_BUS_DRIVER);
|
|
clean2:
|
|
#endif
|
|
#ifdef PCI_DRIVER
|
|
pci_unregister_driver(&PCI_DRIVER);
|
|
clean1:
|
|
#endif
|
|
#ifdef PLATFORM_DRIVER
|
|
platform_driver_unregister(&PLATFORM_DRIVER);
|
|
clean0:
|
|
#endif
|
|
#ifdef DEBUG
|
|
debugfs_remove(ehci_debug_root);
|
|
ehci_debug_root = NULL;
|
|
err_debug:
|
|
#endif
|
|
clear_bit(USB_EHCI_LOADED, &usb_hcds_loaded);
|
|
return retval;
|
|
}
|
|
module_init(ehci_hcd_init);
|
|
|
|
static void __exit ehci_hcd_cleanup(void)
|
|
{
|
|
#ifdef XILINX_OF_PLATFORM_DRIVER
|
|
of_unregister_platform_driver(&XILINX_OF_PLATFORM_DRIVER);
|
|
#endif
|
|
#ifdef OF_PLATFORM_DRIVER
|
|
of_unregister_platform_driver(&OF_PLATFORM_DRIVER);
|
|
#endif
|
|
#ifdef PLATFORM_DRIVER
|
|
platform_driver_unregister(&PLATFORM_DRIVER);
|
|
#endif
|
|
#ifdef PCI_DRIVER
|
|
pci_unregister_driver(&PCI_DRIVER);
|
|
#endif
|
|
#ifdef PS3_SYSTEM_BUS_DRIVER
|
|
ps3_ehci_driver_unregister(&PS3_SYSTEM_BUS_DRIVER);
|
|
#endif
|
|
#ifdef DEBUG
|
|
debugfs_remove(ehci_debug_root);
|
|
#endif
|
|
clear_bit(USB_EHCI_LOADED, &usb_hcds_loaded);
|
|
}
|
|
module_exit(ehci_hcd_cleanup);
|
|
|