[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1446538452-23099-4-git-send-email-baolu.lu@linux.intel.com>
Date: Tue, 3 Nov 2015 16:14:03 +0800
From: Lu Baolu <baolu.lu@...ux.intel.com>
To: Mathias Nyman <mathias.nyman@...el.com>,
Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
Alan Stern <stern@...land.harvard.edu>
Cc: linux-usb@...r.kernel.org, linux-kernel@...r.kernel.org,
Lu Baolu <baolu.lu@...ux.intel.com>
Subject: [PATCH v2 03/12] usb: xhci: dbc: probe and setup xhci debug capability
xHCI debug capability (DbC) is an optional functionality provided
by an xHCI host controller. Software learns this capability by
walking through the extended capability list in mmio of the host.
This patch introduces the code to probe and initialize the debug
capability hardware during early boot. With hardware initialization
done, the debug target (system under debug which has DbC enabled)
will present a debug device through the debug port. The debug device
is fully compliant with the USB framework and provides the equivalent
of a very high performance (USB3) full-duplex serial link between the
debug host and target.
Signed-off-by: Lu Baolu <baolu.lu@...ux.intel.com>
---
MAINTAINERS | 7 +
arch/x86/Kconfig.debug | 12 +
drivers/usb/early/Makefile | 1 +
drivers/usb/early/xhci-dbc.c | 787 +++++++++++++++++++++++++++++++++++++++++++
include/linux/usb/xhci-dbc.h | 187 ++++++++++
5 files changed, 994 insertions(+)
create mode 100644 drivers/usb/early/xhci-dbc.c
create mode 100644 include/linux/usb/xhci-dbc.h
diff --git a/MAINTAINERS b/MAINTAINERS
index 0425167..585a369 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -11040,6 +11040,13 @@ S: Supported
F: drivers/usb/host/xhci*
F: drivers/usb/host/pci-quirks*
+USB XHCI DEBUG PORT
+M: Lu Baolu <baolu.lu@...ux.intel.com>
+L: linux-usb@...r.kernel.org
+S: Supported
+F: drivers/usb/early/xhci-dbc.c
+F: include/linux/usb/xhci-dbc.h
+
USB ZD1201 DRIVER
L: linux-wireless@...r.kernel.org
W: http://linux-lc100020.sourceforge.net
diff --git a/arch/x86/Kconfig.debug b/arch/x86/Kconfig.debug
index d8c0d32..8d95abd 100644
--- a/arch/x86/Kconfig.debug
+++ b/arch/x86/Kconfig.debug
@@ -65,6 +65,18 @@ config EARLY_PRINTK_EFI
This is useful for kernel debugging when your machine crashes very
early before the console code is initialized.
+config EARLY_PRINTK_XDBC
+ bool "Early printk via xHCI debug port"
+ depends on EARLY_PRINTK && PCI
+ ---help---
+ Write kernel log output directly into the xHCI debug port.
+
+ This is useful for kernel debugging when your machine crashes very
+ early before the console code is initialized. For normal operation
+ it is not recommended because it looks ugly and doesn't cooperate
+ with klogd/syslogd or the X server. You should normally N here,
+ unless you want to debug such a crash.
+
config X86_PTDUMP
bool "Export kernel pagetable layout to userspace via debugfs"
depends on DEBUG_KERNEL
diff --git a/drivers/usb/early/Makefile b/drivers/usb/early/Makefile
index 24bbe51..2db5906 100644
--- a/drivers/usb/early/Makefile
+++ b/drivers/usb/early/Makefile
@@ -3,3 +3,4 @@
#
obj-$(CONFIG_EARLY_PRINTK_DBGP) += ehci-dbgp.o
+obj-$(CONFIG_EARLY_PRINTK_XDBC) += xhci-dbc.o
diff --git a/drivers/usb/early/xhci-dbc.c b/drivers/usb/early/xhci-dbc.c
new file mode 100644
index 0000000..22a1de9
--- /dev/null
+++ b/drivers/usb/early/xhci-dbc.c
@@ -0,0 +1,787 @@
+/**
+ * xhci-dbc.c - xHCI debug capability driver
+ *
+ * Copyright (C) 2015 Intel Corporation
+ *
+ * Author: Lu Baolu <baolu.lu@...ux.intel.com>
+ * Some code shared with EHCI debug port and xHCI driver.
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ */
+#include <linux/pci_regs.h>
+#include <linux/pci_ids.h>
+#include <linux/bootmem.h>
+#include <linux/io.h>
+#include <asm/pci-direct.h>
+#include <asm/fixmap.h>
+#include <linux/bcd.h>
+#include <linux/export.h>
+#include <linux/version.h>
+#include <linux/usb/xhci-dbc.h>
+
+#include "../host/xhci.h"
+
+#define XDBC_PROTOCOL 1 /* GNU Remote Debug Command Set */
+#define XDBC_VENDOR_ID 0x1d6b /* Linux Foundation 0x1d6b */
+#define XDBC_PRODUCT_ID 0x0004 /* __le16 idProduct; device 0004 */
+#define XDBC_DEVICE_REV 0x0010 /* 0.10 */
+
+static struct xdbc_state xdbc_stat;
+static struct xdbc_state *xdbcp = &xdbc_stat;
+
+#ifdef DBC_DEBUG
+/* place holder */
+#define xdbc_trace printk
+static void xdbc_dbg_dump_regs(char *str)
+{
+ if (!xdbcp->xdbc_reg) {
+ xdbc_trace("register not mapped\n");
+ return;
+ }
+
+ xdbc_trace("XDBC registers: %s\n", str);
+ xdbc_trace(" Capability: %08x\n",
+ readl(&xdbcp->xdbc_reg->capability));
+ xdbc_trace(" Door bell: %08x\n",
+ readl(&xdbcp->xdbc_reg->doorbell));
+ xdbc_trace(" Event Ring Segment Table Size: %08x\n",
+ readl(&xdbcp->xdbc_reg->ersts));
+ xdbc_trace(" Event Ring Segment Table Base Address: %16llx\n",
+ xdbc_read64(&xdbcp->xdbc_reg->erstba));
+ xdbc_trace(" Event Ring Dequeue Pointer: %16llx\n",
+ xdbc_read64(&xdbcp->xdbc_reg->erdp));
+ xdbc_trace(" Port status and control: %08x\n",
+ readl(&xdbcp->xdbc_reg->portsc));
+ xdbc_trace(" Debug Capability Context Pointer: %16llx\n",
+ xdbc_read64(&xdbcp->xdbc_reg->dccp));
+ xdbc_trace(" Device Descriptor Info Register 1: %08x\n",
+ readl(&xdbcp->xdbc_reg->devinfo1));
+ xdbc_trace(" Device Descriptor Info Register 2: %08x\n",
+ readl(&xdbcp->xdbc_reg->devinfo2));
+}
+
+static void xdbc_dbg_dump_info_context(char *str)
+{
+ int i;
+ u64 addr;
+ u32 *field;
+
+ if (!xdbcp->dbcc_base)
+ return;
+
+ xdbc_trace("%s:\n", str);
+
+ field = (u32 *)xdbcp->dbcc_base;
+ addr = xdbcp->dbcc_dma;
+ for (i = 0; i < XDBC_INFO_CONTEXT_SIZE;) {
+ xdbc_trace("@%016llx %08x %08x %08x %08x\n",
+ addr,
+ le32_to_cpu(field[i]),
+ le32_to_cpu(field[i + 1]),
+ le32_to_cpu(field[i + 2]),
+ le32_to_cpu(field[i + 3]));
+ addr += 16;
+ i += 4;
+ }
+}
+
+static void xdbc_dbg_dump_erst(char *str)
+{
+ int i;
+ u64 addr = xdbcp->erst_dma;
+ struct xdbc_erst_entry *entry;
+
+ if (!xdbcp->erst_base)
+ return;
+
+ xdbc_trace("%s\n", str);
+
+ for (i = 0; i < xdbcp->erst_size / sizeof(*entry); i++) {
+ entry = (struct xdbc_erst_entry *)xdbcp->erst_base + i;
+ xdbc_trace("@%016llx %08x %08x %08x %08x\n",
+ addr,
+ lower_32_bits(le64_to_cpu(entry->seg_addr)),
+ upper_32_bits(le64_to_cpu(entry->seg_addr)),
+ le32_to_cpu(entry->seg_size),
+ le32_to_cpu(entry->rsvd));
+ addr += sizeof(*entry);
+ }
+}
+
+static void xdbc_dbg_dump_segment(struct xdbc_segment *seg, char *str)
+{
+ int i;
+ u64 addr = seg->dma;
+ struct xdbc_trb *trb;
+
+ if (!seg->trbs)
+ return;
+
+ xdbc_trace("%s\n", str);
+
+ for (i = 0; i < XDBC_TRBS_PER_SEGMENT; i++) {
+ trb = &seg->trbs[i];
+ xdbc_trace("@%016llx %08x %08x %08x %08x\n", addr,
+ le32_to_cpu(trb->field[0]),
+ le32_to_cpu(trb->field[1]),
+ le32_to_cpu(trb->field[2]),
+ le32_to_cpu(trb->field[3]));
+ addr += sizeof(*trb);
+ }
+}
+
+static void xdbc_dbg_dump_string(char *str)
+{
+ u32 *string = (u32 *)xdbcp->string_base;
+ int i, max;
+
+ max = xdbcp->string_size / sizeof(u32);
+
+ xdbc_trace("%s\n", str);
+
+ for (i = 0; i < max; i += 4) {
+ xdbc_trace("@%016llx %08x %08x %08x %08x\n",
+ xdbcp->string_dma + i * 16,
+ le32_to_cpu(string[0]),
+ le32_to_cpu(string[1]),
+ le32_to_cpu(string[2]),
+ le32_to_cpu(string[3]));
+ string += 4;
+ }
+}
+
+static void xdbc_dbg_dump_data(char *str)
+{
+ xdbc_trace("XDBC data structure: %s\n", str);
+ xdbc_dbg_dump_erst("ERST:");
+ xdbc_dbg_dump_segment(&xdbcp->evt_seg, "Event Ring Segment:");
+ xdbc_dbg_dump_segment(&xdbcp->out_seg, "TXout Ring Segment:");
+ xdbc_dbg_dump_segment(&xdbcp->in_seg, "TXin Ring Segment:");
+ xdbc_dbg_dump_info_context("DBCC:");
+ xdbc_dbg_dump_string("String Descriptor:");
+}
+
+#else
+static inline void xdbc_trace(const char *fmt, ...) { }
+static inline void xdbc_dbg_dump_regs(char *str) { }
+static inline void xdbc_dbg_dump_data(char *str) { }
+#endif /* DBC_DEBUG */
+
+/*
+ * FIXME: kernel provided delay interfaces, like usleep, isn't ready yet
+ * at the time DbC gets initialized. Below implementation is only
+ * for x86 platform. Need to reconsider this when porting it onto
+ * other architectures.
+ */
+static inline void xdbc_udelay(int us)
+{
+ while (us-- > 0)
+ outb(0x1, 0x80);
+}
+
+static void __iomem *xdbc_map_pci_mmio(u32 bus,
+ u32 dev, u32 func, u8 bar, size_t *length)
+{
+ u32 val, sz;
+ u64 val64, sz64, mask64;
+ u8 byte;
+ unsigned long idx, max_idx;
+ void __iomem *base;
+
+ val = read_pci_config(bus, dev, func, bar);
+ write_pci_config(bus, dev, func, bar, ~0);
+ sz = read_pci_config(bus, dev, func, bar);
+ write_pci_config(bus, dev, func, bar, val);
+ if (val == 0xffffffff || sz == 0xffffffff) {
+ xdbc_trace("invalid mmio bar\n");
+ return NULL;
+ }
+
+ val64 = val & PCI_BASE_ADDRESS_MEM_MASK;
+ sz64 = sz & PCI_BASE_ADDRESS_MEM_MASK;
+ mask64 = (u32)PCI_BASE_ADDRESS_MEM_MASK;
+
+ if ((val & PCI_BASE_ADDRESS_MEM_TYPE_MASK) ==
+ PCI_BASE_ADDRESS_MEM_TYPE_64) {
+ val = read_pci_config(bus, dev, func, bar + 4);
+ write_pci_config(bus, dev, func, bar + 4, ~0);
+ sz = read_pci_config(bus, dev, func, bar + 4);
+ write_pci_config(bus, dev, func, bar + 4, val);
+
+ val64 |= ((u64)val << 32);
+ sz64 |= ((u64)sz << 32);
+ mask64 |= ((u64)~0 << 32);
+ }
+
+ sz64 &= mask64;
+
+ if (sizeof(dma_addr_t) < 8 || !sz64) {
+ xdbc_trace("can't handle 64bit BAR\n");
+ return NULL;
+ }
+
+ sz64 = 1ULL << __ffs64(sz64);
+
+ if (sz64 > (FIX_XDBC_END - FIX_XDBC_BASE + 1) * PAGE_SIZE) {
+ xdbc_trace("mmio size beyond 64k not supported\n");
+ return NULL;
+ }
+
+ xdbc_trace("bar: base 0x%llx size 0x%llx offset %03x\n",
+ val64, sz64, bar);
+
+ /* check if the mem space is enabled */
+ byte = read_pci_config_byte(bus, dev, func, PCI_COMMAND);
+ if (!(byte & PCI_COMMAND_MEMORY)) {
+ byte |= PCI_COMMAND_MEMORY;
+ write_pci_config_byte(bus, dev, func, PCI_COMMAND, byte);
+ xdbc_trace("mmio for xhci enabled\n");
+ }
+
+ /* 64k mmio will be fix-mapped */
+ max_idx = FIX_XDBC_END - FIX_XDBC_BASE;
+ for (idx = 0; idx <= max_idx; idx++)
+ set_fixmap_nocache(FIX_XDBC_BASE + idx,
+ (val64 & PAGE_MASK) + (max_idx - idx) * PAGE_SIZE);
+ base = (void __iomem *)__fix_to_virt(FIX_XDBC_END);
+ base += val64 & ~PAGE_MASK;
+
+ /* save in the state block */
+ xdbcp->bus = bus;
+ xdbcp->dev = dev;
+ xdbcp->func = func;
+ xdbcp->bar = bar;
+ xdbcp->xhci_base = base;
+ xdbcp->xhci_length = sz64;
+
+ if (length)
+ *length = sz64;
+
+ return base;
+}
+
+/*
+ * FIXME: The bootmem allocator isn't ready at the time when DbC gets
+ * initialized. Below implementation reserves DMA memory blocks
+ * in the kernel static data segment.
+ */
+static void *xdbc_get_page(dma_addr_t *dma_addr,
+ enum xdbc_page_type type)
+{
+ void *virt;
+ static char event_page[PAGE_SIZE] __aligned(PAGE_SIZE);
+ static char in_ring_page[PAGE_SIZE] __aligned(PAGE_SIZE);
+ static char out_ring_page[PAGE_SIZE] __aligned(PAGE_SIZE);
+ static char table_page[PAGE_SIZE] __aligned(PAGE_SIZE);
+
+ switch (type) {
+ case XDBC_PAGE_EVENT:
+ virt = (void *)event_page;
+ break;
+ case XDBC_PAGE_TXIN:
+ virt = (void *)in_ring_page;
+ break;
+ case XDBC_PAGE_TXOUT:
+ virt = (void *)out_ring_page;
+ break;
+ case XDBC_PAGE_TABLE:
+ virt = (void *)table_page;
+ break;
+ default:
+ return NULL;
+ }
+
+ memset(virt, 0, PAGE_SIZE);
+
+ if (dma_addr)
+ *dma_addr = (dma_addr_t)__pa(virt);
+
+ return virt;
+}
+
+typedef void (*xdbc_walk_excap_cb)(int cap_offset, void *data);
+
+/*
+ * xdbc_walk_excap:
+ *
+ * xHCI extended capability list walker.
+ *
+ * @bus - xHC PCI bus#
+ * @dev - xHC PCI dev#
+ * @func - xHC PCI function#
+ * @cap - capability ID
+ * @oneshot - return immediately once hit match
+ * @cb - call back
+ * @data - callback private data
+ *
+ * Return the last cap offset, otherwize 0.
+ */
+static u32 xdbc_walk_excap(u32 bus, u32 dev, u32 func, int cap,
+ bool oneshot, xdbc_walk_excap_cb cb, void *data)
+{
+ void __iomem *base;
+ int ext_cap_offset, retval = 0;
+ size_t len = 0;
+ u32 val;
+
+ if (xdbcp->xhci_base && xdbcp->xhci_length) {
+ if (xdbcp->bus != bus ||
+ xdbcp->dev != dev ||
+ xdbcp->func != func) {
+ xdbc_trace("only one DbC can be used\n");
+ return 0;
+ }
+
+ len = xdbcp->xhci_length;
+ base = xdbcp->xhci_base;
+ } else {
+ base = xdbc_map_pci_mmio(bus, dev, func,
+ PCI_BASE_ADDRESS_0, &len);
+ if (base == NULL)
+ return 0;
+ }
+
+ ext_cap_offset = xhci_find_next_cap_offset(base,
+ XHCI_HCC_PARAMS_OFFSET);
+ do {
+ if ((ext_cap_offset + sizeof(val)) >= len) {
+ xdbc_trace("malformed extended capability list\n");
+ break;
+ }
+
+ if (!ext_cap_offset)
+ break;
+
+ val = readl(base + ext_cap_offset);
+ if (XHCI_EXT_CAPS_ID(val) == cap) {
+ retval = ext_cap_offset;
+ if (cb)
+ cb(ext_cap_offset, data);
+ if (oneshot)
+ break;
+ }
+ ext_cap_offset = xhci_find_next_cap_offset(base,
+ ext_cap_offset);
+ } while (1);
+
+ return retval;
+}
+
+static u32 __init xdbc_find_dbgp(int xdbc_num,
+ u32 *rbus, u32 *rdev, u32 *rfunc)
+{
+ u32 bus, dev, func, class;
+ unsigned cap;
+
+ for (bus = 0; bus < XDBC_PCI_MAX_BUSES; bus++) {
+ for (dev = 0; dev < XDBC_PCI_MAX_DEVICES; dev++) {
+ for (func = 0; func < XDBC_PCI_MAX_FUNCTION; func++) {
+ class = read_pci_config(bus, dev, func,
+ PCI_CLASS_REVISION);
+ if ((class >> 8) != PCI_CLASS_SERIAL_USB_XHCI)
+ continue;
+
+ if (xdbc_num-- != 0)
+ continue;
+
+ cap = xdbc_walk_excap(bus, dev, func,
+ XHCI_EXT_CAPS_DEBUG,
+ true, NULL, NULL);
+ *rbus = bus;
+ *rdev = dev;
+ *rfunc = func;
+ return cap;
+ }
+ }
+ }
+
+ return 0;
+}
+
+static int handshake(void __iomem *ptr, u32 mask, u32 done,
+ int wait_usec, int delay_usec)
+{
+ u32 result;
+
+ do {
+ result = readl(ptr);
+ result &= mask;
+ if (result == done)
+ return 0;
+ xdbc_udelay(delay_usec);
+ wait_usec -= delay_usec;
+ } while (wait_usec > 0);
+
+ return -ETIMEDOUT;
+}
+
+static void __init xdbc_bios_handoff(void)
+{
+ int ext_cap_offset;
+ int timeout;
+ u32 val;
+
+ ext_cap_offset = xdbc_walk_excap(xdbcp->bus,
+ xdbcp->dev,
+ xdbcp->func,
+ XHCI_EXT_CAPS_LEGACY,
+ true, NULL, NULL);
+ val = readl(xdbcp->xhci_base + ext_cap_offset);
+
+ /* If the BIOS owns the HC, signal that the OS wants it, and wait */
+ if (val & XHCI_HC_BIOS_OWNED) {
+ writel(val | XHCI_HC_OS_OWNED,
+ xdbcp->xhci_base + ext_cap_offset);
+ timeout = handshake(xdbcp->xhci_base + ext_cap_offset,
+ XHCI_HC_BIOS_OWNED, 0, 5000, 10);
+
+ /* Assume a buggy BIOS and take HC ownership anyway */
+ if (timeout) {
+ xdbc_trace("xHCI BIOS handoff failed (BIOS bug ?)\n");
+ writel(val & ~XHCI_HC_BIOS_OWNED,
+ xdbcp->xhci_base + ext_cap_offset);
+ }
+ }
+
+ /* Disable any BIOS SMIs and clear all SMI events*/
+ val = readl(xdbcp->xhci_base + ext_cap_offset +
+ XHCI_LEGACY_CONTROL_OFFSET);
+ val &= XHCI_LEGACY_DISABLE_SMI;
+ val |= XHCI_LEGACY_SMI_EVENTS;
+ writel(val, xdbcp->xhci_base + ext_cap_offset +
+ XHCI_LEGACY_CONTROL_OFFSET);
+}
+
+/*
+ * xdbc_alloc_ring: allocate physical memory for a ring
+ */
+static int xdbc_alloc_ring(struct xdbc_segment *seg,
+ struct xdbc_ring *ring,
+ enum xdbc_page_type type)
+{
+ struct xdbc_trb *link_trb;
+
+ seg->trbs = xdbc_get_page(&seg->dma, type);
+ if (!seg->trbs)
+ return -ENOMEM;
+
+ ring->segment = seg;
+ ring->enqueue = seg->trbs;
+ ring->dequeue = seg->trbs;
+ ring->cycle_state = 1;
+
+ if (type == XDBC_PAGE_TXIN || type == XDBC_PAGE_TXOUT) {
+ link_trb = &seg->trbs[XDBC_TRBS_PER_SEGMENT - 1];
+ link_trb->field[0] = cpu_to_le32(lower_32_bits(seg->dma));
+ link_trb->field[1] = cpu_to_le32(upper_32_bits(seg->dma));
+ link_trb->field[3] = cpu_to_le32(TRB_TYPE(TRB_LINK)) |
+ cpu_to_le32(LINK_TOGGLE);
+ }
+
+ return 0;
+}
+
+static inline void xdbc_put_utf16(u16 *s, const char *c, size_t size)
+{
+ int i;
+
+ for (i = 0; i < size; i++)
+ s[i] = cpu_to_le16(c[i]);
+}
+
+static int xdbc_mem_init(void)
+{
+ struct xdbc_erst_entry *entry;
+ struct xdbc_strings *strings;
+ struct xdbc_context *context;
+ struct xdbc_ep_context *ep_in, *ep_out;
+ struct usb_string_descriptor *s_desc;
+ unsigned int max_burst;
+ u32 string_length;
+ int ret, index = 0;
+ u32 dev_info;
+
+ /* allocate table page */
+ xdbcp->table_base = xdbc_get_page(&xdbcp->table_dma,
+ XDBC_PAGE_TABLE);
+ if (!xdbcp->table_base) {
+ xdbc_trace("falied to alloc table page\n");
+ return -ENOMEM;
+ }
+
+ /* allocate and initialize event ring */
+ ret = xdbc_alloc_ring(&xdbcp->evt_seg, &xdbcp->evt_ring,
+ XDBC_PAGE_EVENT);
+ if (ret < 0) {
+ xdbc_trace("failed to alloc event ring\n");
+ return ret;
+ }
+
+ /* allocate event ring segment table */
+ xdbcp->erst_size = 16;
+ xdbcp->erst_base = xdbcp->table_base +
+ index * XDBC_TABLE_ENTRY_SIZE;
+ xdbcp->erst_dma = xdbcp->table_dma +
+ index * XDBC_TABLE_ENTRY_SIZE;
+ index += XDBC_ERST_ENTRY_NUM;
+
+ /* Initialize Event Ring Segment Table */
+ entry = (struct xdbc_erst_entry *)xdbcp->erst_base;
+ entry->seg_addr = cpu_to_le64(xdbcp->evt_seg.dma);
+ entry->seg_size = cpu_to_le32(XDBC_TRBS_PER_SEGMENT);
+ entry->rsvd = 0;
+
+ /* Initialize ERST registers */
+ writel(1, &xdbcp->xdbc_reg->ersts);
+ xdbc_write64(xdbcp->erst_dma, &xdbcp->xdbc_reg->erstba);
+ xdbc_write64(xdbcp->evt_seg.dma, &xdbcp->xdbc_reg->erdp);
+
+ /* debug capability contexts */
+ BUILD_BUG_ON(sizeof(struct xdbc_info_context) != 64);
+ BUILD_BUG_ON(sizeof(struct xdbc_ep_context) != 64);
+ BUILD_BUG_ON(sizeof(struct xdbc_context) != 64 * 3);
+
+ xdbcp->dbcc_size = 64 * 3;
+ xdbcp->dbcc_base = xdbcp->table_base +
+ index * XDBC_TABLE_ENTRY_SIZE;
+ xdbcp->dbcc_dma = xdbcp->table_dma +
+ index * XDBC_TABLE_ENTRY_SIZE;
+ index += XDBC_DBCC_ENTRY_NUM;
+
+ /* IN/OUT endpoint transfer ring */
+ ret = xdbc_alloc_ring(&xdbcp->in_seg, &xdbcp->in_ring,
+ XDBC_PAGE_TXIN);
+ if (ret < 0) {
+ xdbc_trace("failed to alloc IN transfer ring\n");
+ return ret;
+ }
+
+ ret = xdbc_alloc_ring(&xdbcp->out_seg, &xdbcp->out_ring,
+ XDBC_PAGE_TXOUT);
+ if (ret < 0) {
+ xdbc_trace("failed to alloc OUT transfer ring\n");
+ return ret;
+ }
+
+ /* strings */
+ xdbcp->string_size = sizeof(struct xdbc_strings);
+ xdbcp->string_base = xdbcp->table_base +
+ index * XDBC_TABLE_ENTRY_SIZE;
+ xdbcp->string_dma = xdbcp->table_dma +
+ index * XDBC_TABLE_ENTRY_SIZE;
+ index += XDBC_STRING_ENTRY_NUM;
+
+ strings = (struct xdbc_strings *)xdbcp->string_base;
+
+ /* serial string */
+ s_desc = (struct usb_string_descriptor *)strings->serial;
+ s_desc->bLength = (strlen(XDBC_STRING_SERIAL) + 1) * 2;
+ s_desc->bDescriptorType = USB_DT_STRING;
+ xdbc_put_utf16(s_desc->wData, XDBC_STRING_SERIAL,
+ strlen(XDBC_STRING_SERIAL));
+
+ string_length = s_desc->bLength;
+ string_length <<= 8;
+
+ /* product string */
+ s_desc = (struct usb_string_descriptor *)strings->product;
+ s_desc->bLength = (strlen(XDBC_STRING_PRODUCT) + 1) * 2;
+ s_desc->bDescriptorType = USB_DT_STRING;
+ xdbc_put_utf16(s_desc->wData, XDBC_STRING_PRODUCT,
+ strlen(XDBC_STRING_PRODUCT));
+
+ string_length += s_desc->bLength;
+ string_length <<= 8;
+
+ /* manufacture string */
+ s_desc = (struct usb_string_descriptor *)strings->manufacture;
+ s_desc->bLength = (strlen(XDBC_STRING_MANUFACTURE) + 1) * 2;
+ s_desc->bDescriptorType = USB_DT_STRING;
+ xdbc_put_utf16(s_desc->wData, XDBC_STRING_MANUFACTURE,
+ strlen(XDBC_STRING_MANUFACTURE));
+
+ string_length += s_desc->bLength;
+ string_length <<= 8;
+
+ /* string 0 */
+ strings->string0[0] = 4;
+ strings->string0[1] = USB_DT_STRING;
+ strings->string0[2] = 0x09;
+ strings->string0[3] = 0x04;
+
+ string_length += 4;
+
+ /* populate the contexts */
+ context = (struct xdbc_context *)xdbcp->dbcc_base;
+ context->info.string0 = cpu_to_le64(xdbcp->string_dma);
+ context->info.manufacture = cpu_to_le64(xdbcp->string_dma +
+ XDBC_MAX_STRING_LENGTH);
+ context->info.product = cpu_to_le64(xdbcp->string_dma +
+ XDBC_MAX_STRING_LENGTH * 2);
+ context->info.serial = cpu_to_le64(xdbcp->string_dma +
+ XDBC_MAX_STRING_LENGTH * 3);
+ context->info.length = cpu_to_le32(string_length);
+
+ max_burst = DEBUG_MAX_BURST(readl(&xdbcp->xdbc_reg->control));
+ ep_out = (struct xdbc_ep_context *)&context->out;
+ ep_out->ep_info1 = 0;
+ ep_out->ep_info2 = cpu_to_le32(EP_TYPE(BULK_OUT_EP) |
+ MAX_PACKET(1024) | MAX_BURST(max_burst));
+ ep_out->deq = cpu_to_le64(xdbcp->out_seg.dma |
+ xdbcp->out_ring.cycle_state);
+
+ ep_in = (struct xdbc_ep_context *)&context->in;
+ ep_in->ep_info1 = 0;
+ ep_in->ep_info2 = cpu_to_le32(EP_TYPE(BULK_OUT_EP) |
+ MAX_PACKET(1024) | MAX_BURST(max_burst));
+ ep_in->deq = cpu_to_le64(xdbcp->in_seg.dma |
+ xdbcp->in_ring.cycle_state);
+
+ /* write DbC context pointer register */
+ xdbc_write64(xdbcp->dbcc_dma, &xdbcp->xdbc_reg->dccp);
+
+ /* device descriptor info registers */
+ dev_info = cpu_to_le32((XDBC_VENDOR_ID << 16) | XDBC_PROTOCOL);
+ writel(dev_info, &xdbcp->xdbc_reg->devinfo1);
+ dev_info = cpu_to_le32((XDBC_DEVICE_REV << 16) | XDBC_PRODUCT_ID);
+ writel(dev_info, &xdbcp->xdbc_reg->devinfo2);
+
+ return 0;
+}
+
+/*
+ * xdbc_start: start DbC
+ *
+ * Set DbC enable bit and wait until DbC run bit being set or timed out.
+ */
+static int xdbc_start(void)
+{
+ u32 ctrl, status;
+
+ ctrl = readl(&xdbcp->xdbc_reg->control);
+ writel(ctrl | CTRL_DCE | CTRL_LSE, &xdbcp->xdbc_reg->control);
+
+ if (handshake(&xdbcp->xdbc_reg->control, CTRL_DCE,
+ CTRL_DCE, 100000, 100) < 0) {
+ xdbc_trace("falied to initialize hardware\n");
+ return -ENODEV;
+ }
+
+ /* wait for port connection */
+ if (handshake(&xdbcp->xdbc_reg->portsc, PORTSC_CCS,
+ PORTSC_CCS, 5000000, 100) < 0) {
+ xdbc_trace("waiting for connection timed out\n");
+ return -ETIMEDOUT;
+ }
+ xdbc_trace("port connection detected\n");
+
+ /* wait for debug device to be configured */
+ if (handshake(&xdbcp->xdbc_reg->control, CTRL_DCR,
+ CTRL_DCR, 5000000, 100) < 0) {
+ xdbc_trace("waiting for device configuration timed out\n");
+ return -ETIMEDOUT;
+ }
+
+ /* port should have a valid port# */
+ status = readl(&xdbcp->xdbc_reg->status);
+ if (!DCST_DPN(status)) {
+ xdbc_trace("invalid root hub port number\n");
+ return -ENODEV;
+ }
+
+ xdbc_trace("root hub port number %d\n", DCST_DPN(status));
+
+ xdbc_trace("DbC is running now, control 0x%08x\n",
+ readl(&xdbcp->xdbc_reg->control));
+
+ return 0;
+}
+
+static int xdbc_setup(void)
+{
+ int ret;
+
+ writel(0, &xdbcp->xdbc_reg->control);
+ if (handshake(&xdbcp->xdbc_reg->control, CTRL_DCE,
+ 0, 100000, 100) < 0) {
+ xdbc_trace("falied to initialize hardware\n");
+ return -ETIMEDOUT;
+ }
+
+ /* allocate and initialize all memory data structures */
+ ret = xdbc_mem_init();
+ if (ret < 0) {
+ xdbc_trace("failed to initialize memory\n");
+ return ret;
+ }
+
+ /*
+ * Memory barrier to ensure hardware sees the bits
+ * setting above.
+ */
+ mmiowb();
+
+ /* dump registers and data structures */
+ xdbc_dbg_dump_regs("hardware setup completed");
+ xdbc_dbg_dump_data("hardware setup completed");
+
+ ret = xdbc_start();
+ if (ret < 0) {
+ xdbc_trace("failed to start DbC, cable connected?\n");
+ return ret;
+ }
+
+ return 0;
+}
+
+int __init early_xdbc_init(char *s)
+{
+ u32 bus = 0, dev = 0, func = 0;
+ unsigned long dbgp_num = 0;
+ u32 offset;
+ int ret;
+
+ if (!early_pci_allowed())
+ return -EPERM;
+
+ /* FIXME: early printk "keep" option will be supported later */
+ if (strstr(s, "keep"))
+ return -EPERM;
+
+ if (xdbcp->xdbc_reg)
+ return 0;
+
+ if (*s && kstrtoul(s, 0, &dbgp_num))
+ dbgp_num = 0;
+
+ xdbc_trace("dbgp_num: %lu\n", dbgp_num);
+
+ offset = xdbc_find_dbgp(dbgp_num, &bus, &dev, &func);
+ if (!offset)
+ return -ENODEV;
+
+ xdbc_trace("Found xHCI debug capability on %02x:%02x.%1x\n",
+ bus, dev, func);
+
+ if (!xdbcp->xhci_base)
+ return -EINVAL;
+
+ xdbcp->xdbc_reg = (struct xdbc_regs __iomem *)
+ (xdbcp->xhci_base + offset);
+ xdbc_dbg_dump_regs("debug capability located");
+
+ /* hand over the owner of host from BIOS */
+ xdbc_bios_handoff();
+
+ ret = xdbc_setup();
+ if (ret < 0) {
+ pr_notice("failed to setup xHCI DbC connection\n");
+ xdbcp->xhci_base = NULL;
+ xdbcp->xdbc_reg = NULL;
+ return ret;
+ }
+
+ return 0;
+}
diff --git a/include/linux/usb/xhci-dbc.h b/include/linux/usb/xhci-dbc.h
new file mode 100644
index 0000000..153fb87
--- /dev/null
+++ b/include/linux/usb/xhci-dbc.h
@@ -0,0 +1,187 @@
+/*
+ * xHCI debug capability driver
+ *
+ * Copyright (C) 2015 Intel Corporation
+ *
+ * Author: Lu Baolu <baolu.lu@...ux.intel.com>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ */
+
+#ifndef __LINUX_XHCI_DBC_H
+#define __LINUX_XHCI_DBC_H
+
+#include <linux/types.h>
+#include <linux/usb/ch9.h>
+
+/**
+ * struct xdbc_regs - xHCI Debug Capability Register interface.
+ */
+struct xdbc_regs {
+ __le32 capability;
+ __le32 doorbell;
+ __le32 ersts; /* Event Ring Segment Table Size*/
+ __le32 rvd0; /* 0c~0f reserved bits */
+ __le64 erstba; /* Event Ring Segment Table Base Address */
+ __le64 erdp; /* Event Ring Dequeue Pointer */
+ __le32 control;
+#define DEBUG_MAX_BURST(p) (((p) >> 16) & 0xff)
+#define CTRL_DCR (1 << 0) /* DbC Run */
+#define CTRL_PED (1 << 1) /* Port Enable/Disable */
+#define CTRL_HOT (1 << 2) /* Halt Out TR */
+#define CTRL_HIT (1 << 3) /* Halt In TR */
+#define CTRL_DRC (1 << 4) /* DbC run change */
+#define CTRL_DCE (1 << 31) /* DbC enable */
+#define CTRL_LSE (1 << 1)
+ __le32 status;
+#define DCST_DPN(p) (((p) >> 24) & 0xff)
+ __le32 portsc; /* Port status and control */
+#define PORTSC_CCS (1 << 0)
+#define PORTSC_CSC (1 << 17)
+#define PORTSC_PRC (1 << 21)
+#define PORTSC_PLC (1 << 22)
+#define PORTSC_CEC (1 << 23)
+ __le32 rvd1; /* 2b~28 reserved bits */
+ __le64 dccp; /* Debug Capability Context Pointer */
+ __le32 devinfo1; /* Device Descriptor Info Register 1 */
+ __le32 devinfo2; /* Device Descriptor Info Register 2 */
+};
+
+/*
+ * xHCI Debug Capability data structures
+ */
+struct xdbc_trb {
+ __le32 field[4];
+};
+
+struct xdbc_erst_entry {
+ __le64 seg_addr;
+ __le32 seg_size;
+ __le32 rsvd;
+};
+
+struct xdbc_info_context {
+ __le64 string0;
+ __le64 manufacture;
+ __le64 product;
+ __le64 serial;
+ __le32 length;
+ __le32 rsvdz[7];
+};
+
+struct xdbc_ep_context {
+ __le32 ep_info1;
+ __le32 ep_info2;
+ __le64 deq;
+ __le32 tx_info;
+ __le32 rsvd0[11];
+};
+
+struct xdbc_context {
+ struct xdbc_info_context info;
+ struct xdbc_ep_context out;
+ struct xdbc_ep_context in;
+};
+
+#define XDBC_INFO_CONTEXT_SIZE 48
+
+#define XDBC_MAX_STRING_LENGTH 64
+#define XDBC_STRING_MANUFACTURE "Linux"
+#define XDBC_STRING_PRODUCT "Remote GDB"
+#define XDBC_STRING_SERIAL "0001"
+struct xdbc_strings {
+ char string0[XDBC_MAX_STRING_LENGTH];
+ char manufacture[XDBC_MAX_STRING_LENGTH];
+ char product[XDBC_MAX_STRING_LENGTH];
+ char serial[XDBC_MAX_STRING_LENGTH];
+};
+
+/*
+ * software state structure
+ */
+struct xdbc_segment {
+ struct xdbc_trb *trbs;
+ dma_addr_t dma;
+};
+
+#define XDBC_TRBS_PER_SEGMENT 256
+
+struct xdbc_ring {
+ struct xdbc_segment *segment;
+ struct xdbc_trb *enqueue;
+ struct xdbc_trb *dequeue;
+ u32 cycle_state;
+};
+
+enum xdbc_page_type {
+ XDBC_PAGE_EVENT,
+ XDBC_PAGE_TXIN,
+ XDBC_PAGE_TXOUT,
+ XDBC_PAGE_TABLE,
+};
+
+struct xdbc_state {
+ /* pci device info*/
+ u32 bus;
+ u32 dev;
+ u32 func;
+ u8 bar;
+ void __iomem *xhci_base;
+ size_t xhci_length;
+#define XDBC_PCI_MAX_BUSES 256
+#define XDBC_PCI_MAX_DEVICES 32
+#define XDBC_PCI_MAX_FUNCTION 8
+
+ /* DbC register base */
+ struct xdbc_regs __iomem *xdbc_reg;
+
+ /* DbC table page */
+ dma_addr_t table_dma;
+ void *table_base;
+
+#define XDBC_TABLE_ENTRY_SIZE 64
+#define XDBC_ERST_ENTRY_NUM 1
+#define XDBC_DBCC_ENTRY_NUM 3
+#define XDBC_STRING_ENTRY_NUM 4
+
+ /* event ring segment table */
+ dma_addr_t erst_dma;
+ size_t erst_size;
+ void *erst_base;
+
+ /* event ring segments */
+ struct xdbc_ring evt_ring;
+ struct xdbc_segment evt_seg;
+
+ /* debug capability contexts */
+ dma_addr_t dbcc_dma;
+ size_t dbcc_size;
+ void *dbcc_base;
+
+ /* descriptor strings */
+ dma_addr_t string_dma;
+ size_t string_size;
+ void *string_base;
+
+ /* bulk OUT endpoint */
+ struct xdbc_ring out_ring;
+ struct xdbc_segment out_seg;
+
+ /* bulk IN endpoint */
+ struct xdbc_ring in_ring;
+ struct xdbc_segment in_seg;
+};
+
+#define XDBC_MAX_PACKET 1024
+
+/* door bell target */
+#define OUT_EP_DOORBELL 0
+#define IN_EP_DOORBELL 1
+#define DOOR_BELL_TARGET(p) (((p) & 0xff) << 8)
+
+#define xdbc_read64(regs) xhci_read_64(NULL, (regs))
+#define xdbc_write64(val, regs) xhci_write_64(NULL, (val), (regs))
+
+#endif /* __LINUX_XHCI_DBC_H */
--
2.1.4
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists