[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <82807C73E0BDBB498164AAE7B8F6A1481C61263E@MX105CL01.corp.emc.com>
Date: Tue, 19 Jan 2016 11:58:06 +0000
From: "Singhal, Maneesh" <Maneesh.Singhal@....com>
To: "linux-scsi@...r.kernel.org" <linux-scsi@...r.kernel.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"JBottomley@...n.com" <JBottomley@...n.com>,
"martin.petersen@...cle.com" <martin.petersen@...cle.com>,
"linux-api@...r.kernel.org" <linux-api@...r.kernel.org>
Subject: [PATCH] drivers/scsi/emcctd: drivers/scsi/emcctd: Client driver
implementation for EMC-Symmetrix GuestOS emulated Cut-Through Device
Hello,
Kindly review the following patch for the following driver to be added in SCSI subsystem -
Regards
Maneesh
----------------------------------------------------------------------------
>From f3c4b836d6f130b1d7ded618002c8164f8f4a06d Mon Sep 17 00:00:00 2001
From: "maneesh.singhal" <maneesh.singhal@....com>
Date: Tue, 19 Jan 2016 06:39:35 -0500
Subject: [PATCH] [PATCH] drivers/scsi/emcctd: Client driver implementation for
EMC-Symmetrix GuestOS emulated Cut-Through Device.
The patch is a driver implementation EMC-Symmetrix GuestOS emulated Cut-Through
Device. The Cut-Through Device PCI emulation is implemented for GuestOS
environments in the HyperMax OS. GuestOS environments allows loading of
any x86 compliant operating systems like Linux/FreeBSD etc.
The client driver is a SCSI HBA implementation which interfaces with SCSI
midlayer in the north-bound interfaces and connects with the emulated PCI device
on the south side.
The PCI vendor ID:product ID for emulated Cut-Through Device is 0x1120:0x1B00.
Signed-off-by: maneesh.singhal <maneesh.singhal@....com>
---
Documentation/scsi/emcctd.txt | 57 +
MAINTAINERS | 9 +
drivers/scsi/Kconfig | 1 +
drivers/scsi/Makefile | 1 +
drivers/scsi/emcctd/Kconfig | 7 +
drivers/scsi/emcctd/Makefile | 1 +
drivers/scsi/emcctd/README | 10 +
drivers/scsi/emcctd/emc_ctd_interface.h | 386 +++++
drivers/scsi/emcctd/emcctd.c | 2840 +++++++++++++++++++++++++++++++
drivers/scsi/emcctd/emcctd.h | 232 +++
10 files changed, 3544 insertions(+)
create mode 100644 Documentation/scsi/emcctd.txt
create mode 100644 drivers/scsi/emcctd/Kconfig
create mode 100644 drivers/scsi/emcctd/Makefile
create mode 100644 drivers/scsi/emcctd/README
create mode 100644 drivers/scsi/emcctd/emc_ctd_interface.h
create mode 100644 drivers/scsi/emcctd/emcctd.c
create mode 100644 drivers/scsi/emcctd/emcctd.h
diff --git a/Documentation/scsi/emcctd.txt b/Documentation/scsi/emcctd.txt
new file mode 100644
index 0000000..bcafc87
--- /dev/null
+++ b/Documentation/scsi/emcctd.txt
@@ -0,0 +1,56 @@
+This file contains brief information about the EMC Cut-Through Driver (emcctd).
+The driver is currently maintained by Singhal, Maneesh (maneesh.singhal@....com)
+
+Last modified: Mon Jan 18 2016 by Maneesh Singhal
+
+BASICS
+
+Its a client driver implementation for EMC-Symmetrix GuestOS emulated
+Cut-Through Device. The Cut-Through Device PCI emulation is implemented for
+GuestOS environments in the HyperMax OS. GuestOS environments allows loading of
+any x86 compliant operating systems like Linux/FreeBSD etc.
+
+The client driver is a SCSI HBA implementation which interfaces with SCSI
+midlayer in the north-bound interfaces and connects with the emulated PCI device
+on the south side.
+
+The PCI vendor ID:product ID for emulated Cut-Through Device is 0x1120:0x1B00.
+
+VERSIONING
+
+The Version of the driver is maintained as 2.0.0.X, where 2 refers to the CTD
+protocol in use, and X refers to the ongoing version of the driver.
+
+
+SYSFS SUPPORT
+
+The driver creates the directory /sys/module/emcctd and populates it with
+version file and a directory for various parameters as described in MODULE
+PARAMETERS section.
+
+PROCFS SUPPORT
+
+The driver creates the directory /proc/emc and creates files emcctd_stats_x
+where 'x' refers to the PCI emulation number this client driver connected to.
+These files cotains WWN information and IO statistics for the particular PCI
+emulation.
+
+MODULE PARAMETERS
+
+The supported parameters which could add debuggability or change the runtime
+behavior of the driver are as following:
+
+ctd_debug=0 | 1 Enable driver debug messages(0=off, 1=on)
+
+max_luns=xx Specify the maximum number of LUN's per
+ host(default=16384)
+
+cmd_per_lun=xx Specify the maximum commands per lun(default=16)
+
+DEBUGGING HINTS
+
+Debugging code is now compiled in by default but debugging is turned off
+with the kernel module parameter debug_flag defaulting to 0. To enable debug at
+module load time add ctd_debug=1 to the module load options.
+Debugging can also be enabled and disabled by writing a '0' (disable) or '1'
+(enable) to the sysfs file /sys/module/emcctd/parameters/ctd_debug
diff --git a/MAINTAINERS b/MAINTAINERS
index 1d23f70..671173a 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -6601,6 +6601,15 @@ F: drivers/message/fusion/
F: drivers/scsi/mpt2sas/
F: drivers/scsi/mpt3sas/
+EMC Cut-Through Driver
+M: fredette, matt <matt.fredette@....com>
+M: Pirotte, Serge <serge.pirotte@....com>
+M: Singh Animesh <Animesh.Singh@....com>
+M: Singhal, Maneesh <Maneesh.Singhal@....com>
+L: linux-scsi@...r.kernel.org
+S: Maintained
+F: drivers/scsis/emcctd/
+
LSILOGIC/SYMBIOS/NCR 53C8XX and 53C1010 PCI-SCSI drivers
M: Matthew Wilcox <matthew@....cx>
L: linux-scsi@...r.kernel.org
diff --git a/drivers/scsi/Kconfig b/drivers/scsi/Kconfig
index c1fe0d2..cbc03d7 100644
--- a/drivers/scsi/Kconfig
+++ b/drivers/scsi/Kconfig
@@ -477,6 +477,7 @@ source "drivers/scsi/aic7xxx/Kconfig.aic79xx"
source "drivers/scsi/aic94xx/Kconfig"
source "drivers/scsi/hisi_sas/Kconfig"
source "drivers/scsi/mvsas/Kconfig"
+source "drivers/scsi/emcctd/Kconfig"
config SCSI_MVUMI
tristate "Marvell UMI driver"
diff --git a/drivers/scsi/Makefile b/drivers/scsi/Makefile
index 862ab4e..55bea65 100644
--- a/drivers/scsi/Makefile
+++ b/drivers/scsi/Makefile
@@ -132,6 +132,7 @@ obj-$(CONFIG_SCSI_IBMVFC) += ibmvscsi/
obj-$(CONFIG_SCSI_HPTIOP) += hptiop.o
obj-$(CONFIG_SCSI_STEX) += stex.o
obj-$(CONFIG_SCSI_MVSAS) += mvsas/
+obj-$(CONFIG_SCSI_EMCCTD) += emcctd/
obj-$(CONFIG_SCSI_MVUMI) += mvumi.o
obj-$(CONFIG_PS3_ROM) += ps3rom.o
obj-$(CONFIG_SCSI_CXGB3_ISCSI) += libiscsi.o libiscsi_tcp.o cxgbi/
diff --git a/drivers/scsi/emcctd/Kconfig b/drivers/scsi/emcctd/Kconfig
new file mode 100644
index 0000000..d995b53
--- /dev/null
+++ b/drivers/scsi/emcctd/Kconfig
@@ -0,0 +1,7 @@
+config SCSI_EMCCTD
+ tristate "EMC Cut-Through client driver"
+ depends on SCSI && PCI && X86
+ help
+ This driver supports EMC virtual PCI Cut-Through-Device.
+ To compile this driver as a module, choose M here: the
+ module will be called emcctd.
diff --git a/drivers/scsi/emcctd/Makefile b/drivers/scsi/emcctd/Makefile
new file mode 100644
index 0000000..74a070e
--- /dev/null
+++ b/drivers/scsi/emcctd/Makefile
@@ -0,0 +1 @@
+obj-$(CONFIG_SCSI_EMCCTD) += emcctd.o
diff --git a/drivers/scsi/emcctd/README b/drivers/scsi/emcctd/README
new file mode 100644
index 0000000..496ce7f
--- /dev/null
+++ b/drivers/scsi/emcctd/README
@@ -0,0 +1,10 @@
+EMCCTD : Client driver implementation for EMC-Symmetrix GuestOS emulated
+Cut-Through Device. The Cut-Through Device PCI emulation is implemented for
+GuestOS environments in the HyperMax OS. GuestOS environments allows loading of
+any x86 compliant operating systems like Linux/FreeBSD etc.
+
+The client driver is a SCSI HBA implementation which interfaces with SCSI
+midlayer in the north-bound interfaces and connects with the emulated PCI device
+on the south side.
+
+The PCI vendor ID:product ID for emulated Cut-Through Device is 0x1120:0x1B00.
diff --git a/drivers/scsi/emcctd/emc_ctd_interface.h b/drivers/scsi/emcctd/emc_ctd_interface.h
new file mode 100644
index 0000000..58a0276
--- /dev/null
+++ b/drivers/scsi/emcctd/emc_ctd_interface.h
@@ -0,0 +1,386 @@
+/*
+ * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem.
+ *
+ * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA.
+ *
+ * Authors:
+ * fredette, matt <matt.fredette@....com>
+ * Pirotte, Serge <serge.pirotte@....com>
+ * Singh Animesh <Animesh.Singh@....com>
+ * Singhal, Maneesh <Maneesh.Singhal@....com>
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ */
+
+#ifndef _EMC_CTD_INTERFACE_H
+#define _EMC_CTD_INTERFACE_H
+/*
+ *
+ * DESCRIPTION: Data structures used by the Cut-through subsystem.
+ *
+ * NOTES: Changes to any of these structures will mean that any Clients that
+ * depend on them will also need to be modified. Since many of those
+ * Clients are not part of the Enginuity build process, this will almost
+ * certainly require new versions of Guest OS CTD Client code to be
+ * released.
+ *
+ * Before modifying any data structures in this file please discuss the
+ * change with the maintainers.
+ *
+ */
+
+
+/* macros: */
+
+/* the PCI vendor ID for all devices: */
+#define EMC_CTD_PCI_VENDOR (0x1120)
+
+/* the PCI product ID for all version 1.x devices: */
+#define EMC_CTD_V010_PCI_PRODUCT (0x1b00)
+
+/* the PCI revision ID for the first version 1.0 device: */
+#define EMC_CTD_V010_PCI_REVISION (1)
+
+/* the 64-bit BAR pair for the transmit and receive rings: */
+#define EMC_CTD_V010_BAR_RINGS (0)
+
+/* the 64-bit BAR pair for the fast registers: */
+#define EMC_CTD_V010_BAR_FREGS (2)
+
+/* the 32-bit BAR for the slow registers: */
+#define EMC_CTD_V010_BAR_SREGS (4)
+
+/* the maximum number of immediate SGL entries: */
+#define EMC_CTD_V010_SGL_IMMEDIATE_MAX (7)
+
+/* the CTD v010 whats: */
+#define EMC_CTD_V010_WHAT_DETECT (0)
+#define EMC_CTD_V010_WHAT_SCSI_COMMAND (1)
+#define EMC_CTD_V010_WHAT_SCSI_PHASE (2)
+#define EMC_CTD_V010_WHAT_SCSI_RESPONSE (3)
+
+/* the CTD v010 detect flags. all undefined flags must be zero: */
+
+/* if this is set, the name is a SCSI target: */
+#define EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET (1 << 0)
+
+/* if this is set, the name is a SCSI initiator: */
+#define EMC_CTD_V010_DETECT_FLAG_SCSI_INITIATOR (1 << 1)
+
+/* the CTD v010 SCSI command flags. all undefined flags must be zero: */
+
+/* when the guest receives a SCSI command message, this flag is undefined.
+ *
+ * If this is set, at the beginning of any data phase the target is the data
+ * source. if this is clear, at the beginning of any data phase the target is
+ * the data sink.
+ */
+#define EMC_CTD_V010_SCSI_COMMAND_FLAG_SOURCE (1 << 0)
+
+/* when the guest receives a SCSI command message, this flag is undefined.
+ *
+ * if this is set, the first SGL entry in the message points to an extended SGL,
+ * and the remaining SGL entries in the message are undefined. if this is
+ * clear, the SGL entries in the message are used:
+ */
+#define EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL (1 << 1)
+
+/* the CTD v010 SCSI response flags. all undefined flags must be zero: */
+
+/* if this is set, the SCSI command failed. if this is clear, the
+ * command succeeded:
+ */
+#define EMC_CTD_V010_SCSI_RESPONSE_FLAG_FAILED (1 << 0)
+
+/* if this is set, any extra information is sense data. if this is clear, any
+ * extra information is 64-bit timestamps:
+ */
+#define EMC_CTD_V010_SCSI_RESPONSE_FLAG_SENSE (1 << 1)
+
+/* the CTD v010 SCSI phase flags. all undefined flags must be zero: */
+
+/* when the guest receives a SCSI phase message, this flag is undefined.
+ *
+ * if this is set, at this point in the data phase the message receiver is the
+ * data source. if this is clear, at this point in the data phase the message
+ * receiver is the data sink:
+ */
+#define EMC_CTD_V010_SCSI_PHASE_FLAG_SOURCE (1 << 0)
+
+/* when the guest receives a SCSI phase message, this flag is undefined.
+ *
+ * if this is set, the first SGL entry in the message points to an extended SGL,
+ * and the remaining SGL entries in the message are undefined. if this is
+ * clear, the SGL entries in the message are used:
+ */
+#define EMC_CTD_V010_SCSI_PHASE_FLAG_ESGL (1 << 1)
+
+/* if this is set, the message receiver is the target. if this is clear, the
+ * message receiver is the initiator:
+ */
+#define EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET (1 << 2)
+
+/* if this is set, the SCSI command is aborted: */
+#define EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT (1 << 3)
+
+/* the size of the log of errored transmit messages: */
+#define EMC_CTD_V010_LOG_ERROR_TX_SIZE (4)
+
+/* errors: */
+
+/* no error: */
+#define EMC_CTD_V010_ERROR_NULL (0)
+
+/* the guest tried to transmit a message on a disconnected channel: */
+#define EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED (1)
+
+/* the guest tried to transmit a message with a bad what: */
+#define EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT (2)
+
+/* the guest tried to transmit a message with a reserved field set to
+ * the wrong value:
+ */
+#define EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED (3)
+
+/* the guest tried to transmit an out-of-order message: */
+#define EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER (4)
+
+/* the guest tried to transmit a message to an endpoint whose type
+ * doesn't support it:
+ */
+#define EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE (5)
+
+/* the guest tried to transmit a message with an unknown message
+ * receiver's opaque value:
+ */
+#define EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN (6)
+
+/* types: */
+
+/* a CTD v010 scatter/gather list entry: */
+struct emc_ctd_v010_sgl {
+
+ /* the physical address of the buffer: */
+ emc_ctd_uint32_t emc_ctd_v010_sgl_paddr_0_31;
+ emc_ctd_uint32_t emc_ctd_v010_sgl_paddr_32_63;
+
+ /* the size of the buffer: */
+ emc_ctd_uint32_t emc_ctd_v010_sgl_size;
+};
+
+/* a CTD v010 header: */
+struct emc_ctd_v010_header {
+
+ /* the other address: */
+ emc_ctd_uint16_t emc_ctd_v010_header_address;
+
+ /* the minor version: */
+ emc_ctd_uint8_t emc_ctd_v010_header_minor;
+
+ /* the what: */
+ emc_ctd_uint8_t emc_ctd_v010_header_what;
+};
+
+/* a CTD v010 name: */
+struct emc_ctd_v010_name {
+
+ /* the name: */
+ emc_ctd_uint8_t emc_ctd_v010_name_bytes[8];
+};
+
+/* a CTD v010 detect message: */
+struct emc_ctd_v010_detect {
+
+ /* the header: */
+ struct emc_ctd_v010_header emc_ctd_v010_detect_header;
+
+ /* the flags: */
+ emc_ctd_uint32_t emc_ctd_v010_detect_flags;
+
+ /* the name: */
+ struct emc_ctd_v010_name emc_ctd_v010_detect_name;
+
+ /* the key: */
+ emc_ctd_uint64_t emc_ctd_v010_detect_key;
+};
+
+/* a CTD v010 SCSI command message: */
+struct emc_ctd_v010_scsi_command {
+
+ /* the header: */
+ struct emc_ctd_v010_header emc_ctd_v010_scsi_command_header;
+
+ /* the flags: */
+ emc_ctd_uint32_t emc_ctd_v010_scsi_command_flags;
+
+ /* the initiator's opaque value: */
+ emc_ctd_uint64_t emc_ctd_v010_scsi_command_opaque;
+
+ /* the SCSI LUN: */
+ emc_ctd_uint8_t emc_ctd_v010_scsi_command_lun[8];
+
+ /* the SCSI CDB: */
+ emc_ctd_uint8_t emc_ctd_v010_scsi_command_cdb[16];
+
+ /* the data size: */
+ emc_ctd_uint32_t emc_ctd_v010_scsi_command_data_size;
+
+ union {
+
+ /* any SGL entries: */
+ /* when received by the guest, these are undefined: */
+ struct emc_ctd_v010_sgl
+ emc_ctd_v010_scsi_command_sgl[EMC_CTD_V010_SGL_IMMEDIATE_MAX];
+
+ } emc_ctd_v010_scsi_command_u;
+};
+
+/* a CTD v010 SCSI response message: */
+struct emc_ctd_v010_scsi_response {
+
+ /* the header: */
+ struct emc_ctd_v010_header emc_ctd_v010_scsi_response_header;
+
+ /* the flags: */
+ emc_ctd_uint16_t emc_ctd_v010_scsi_response_flags;
+
+ /* the extra information size: */
+ emc_ctd_uint8_t emc_ctd_v010_scsi_response_extra_size;
+
+ /* the SCSI status: */
+ emc_ctd_uint8_t emc_ctd_v010_scsi_response_status;
+
+ /* the initiator's opaque value: */
+ emc_ctd_uint64_t emc_ctd_v010_scsi_response_opaque;
+
+ /* the data size: */
+ emc_ctd_uint32_t emc_ctd_v010_scsi_response_data_size;
+
+ union {
+
+ /* any extra information: */
+ emc_ctd_uint8_t emc_ctd_v010_scsi_response_extra[108];
+
+ } emc_ctd_v010_scsi_response_u;
+};
+
+/* a CTD v010 SCSI phase message: */
+struct emc_ctd_v010_scsi_phase {
+
+ /* the header: */
+ struct emc_ctd_v010_header emc_ctd_v010_scsi_phase_header;
+
+ /* the flags: */
+ emc_ctd_uint32_t emc_ctd_v010_scsi_phase_flags;
+
+ /* the message receiver's opaque value: */
+ emc_ctd_uint64_t emc_ctd_v010_scsi_phase_opaque_rx;
+
+ /* the message transmitter's opaque value: */
+ emc_ctd_uint64_t emc_ctd_v010_scsi_phase_opaque_tx;
+
+ union {
+
+ /* any SGL entries: */
+ /* when received by the guest, these are undefined: */
+ struct emc_ctd_v010_sgl
+ emc_ctd_v010_scsi_phase_sgl[EMC_CTD_V010_SGL_IMMEDIATE_MAX];
+
+ } emc_ctd_v010_scsi_phase_u;
+};
+
+/* a CTD v010 message: */
+union emc_ctd_v010_message {
+
+ /* the header: */
+ struct emc_ctd_v010_header emc_ctd_v010_message_header;
+
+ /* a detect message: */
+ struct emc_ctd_v010_detect emc_ctd_v010_message_detect;
+
+ /* a SCSI command message: */
+ struct emc_ctd_v010_scsi_command emc_ctd_v010_message_scsi_command;
+
+ /* a SCSI response message: */
+ struct emc_ctd_v010_scsi_response emc_ctd_v010_message_scsi_response;
+
+ /* a SCSI phase message: */
+ struct emc_ctd_v010_scsi_phase emc_ctd_v010_message_scsi_phase;
+
+ /* padding: */
+ emc_ctd_uint8_t emc_ctd_v010_message_padding[128];
+};
+
+/* the fast registers: */
+struct emc_ctd_v010_fregs {
+
+ /* the transmit ring producer index (TPI): */
+ volatile emc_ctd_uint32_t emc_ctd_v010_fregs_tx_index_producer;
+
+ /* the error flag: */
+ volatile emc_ctd_uint32_t emc_ctd_v010_fregs_error_flag;
+
+ /* errors 1..14: */
+ volatile emc_ctd_uint32_t emc_ctd_v010_fregs_errors_1_14[14];
+
+ /* the transmit ring consumer index (TCI): */
+ volatile emc_ctd_uint32_t emc_ctd_v010_fregs_tx_index_consumer;
+
+ /* the device name: */
+ struct emc_ctd_v010_name emc_ctd_v010_fregs_device_name;
+
+ /* padding: */
+ emc_ctd_uint8_t
+ emc_ctd_v010_fregs_pad_07f[64 -
+ (sizeof(emc_ctd_uint32_t) +
+ sizeof(struct emc_ctd_v010_name))];
+
+ /* the receive ring producer index (RPI): */
+ volatile emc_ctd_uint32_t emc_ctd_v010_fregs_rx_index_producer;
+
+ /* the interrupt throttle, in units of nanoseconds. zero disables
+ * the throttle:
+ */
+ emc_ctd_uint32_t emc_ctd_v010_fregs_interrupt_throttle_nsecs;
+
+ /* padding: */
+ emc_ctd_uint8_t
+ emc_ctd_v010_fregs_pad_0bf[64 -
+ (sizeof(emc_ctd_uint32_t) + sizeof(emc_ctd_uint32_t))];
+
+ /* the receive ring consumer index (RCI): */
+ volatile emc_ctd_uint32_t emc_ctd_v010_fregs_rx_index_consumer;
+
+ /* padding: */
+ emc_ctd_uint8_t
+ emc_ctd_v010_fregs_pad_0ff[64 -
+ (sizeof(emc_ctd_uint32_t) +
+ (sizeof(emc_ctd_uint32_t) *
+ EMC_CTD_V010_LOG_ERROR_TX_SIZE))];
+
+ /* the errors for the log of errored transmit messages: */
+ volatile emc_ctd_uint32_t
+ emc_ctd_v010_fregs_log_error_tx_error[EMC_CTD_V010_LOG_ERROR_TX_SIZE];
+
+ /* the log of errored transmit messages: */
+ union emc_ctd_v010_message
+ emc_ctd_v010_fregs_log_error_tx_message[EMC_CTD_V010_LOG_ERROR_TX_SIZE];
+};
+
+/* the slow registers: */
+struct emc_ctd_v010_sregs {
+
+ /* the reset register: */
+ emc_ctd_uint32_t emc_ctd_v010_sregs_reset;
+};
+
+#endif /* _EMC_CTD_INTERFACE_H */
diff --git a/drivers/scsi/emcctd/emcctd.c b/drivers/scsi/emcctd/emcctd.c
new file mode 100644
index 0000000..aa6dd0e
--- /dev/null
+++ b/drivers/scsi/emcctd/emcctd.c
@@ -0,0 +1,2840 @@
+/*
+ * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem.
+ *
+ * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA.
+ *
+ * Authors:
+ * fredette, matt <matt.fredette@....com>
+ * Pirotte, Serge <serge.pirotte@....com>
+ * Singh Animesh <Animesh.Singh@....com>
+ * Singhal, Maneesh <Maneesh.Singhal@....com>
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ */
+
+#include <linux/init.h>
+#include <linux/kernel.h>
+#include <linux/kthread.h>
+#include <linux/slab.h>
+#include <linux/list.h>
+#include <linux/module.h>
+#include <linux/spinlock.h>
+#include <linux/pci.h>
+#include <linux/interrupt.h>
+#include <linux/proc_fs.h>
+#include <linux/version.h>
+#include <linux/delay.h>
+#include <linux/blkdev.h>
+#include <linux/atomic.h>
+
+#include <linux/cache.h>
+
+#include <linux/seq_file.h>
+
+#include <scsi/scsi.h>
+#include <scsi/scsi_cmnd.h>
+#include <scsi/scsi_device.h>
+#include <scsi/scsi_host.h>
+#include <scsi/scsi_transport.h>
+#include <scsi/scsi_dbg.h>
+
+#define emc_ctd_uint8_t u8
+#define emc_ctd_uint16_t u16
+#define emc_ctd_uint32_t u32
+#define emc_ctd_uint64_t u64
+
+#include "emc_ctd_interface.h"
+
+#include "emcctd.h"
+
+#include <scsi/scsi_dbg.h>
+
+/* nomenclature for versioning
+ * MAJOR:MINOR:SUBVERSION:PATCH
+ */
+
+#define EMCCTD_MODULE_VERSION "2.0.0.24"
+
+MODULE_LICENSE("GPL");
+MODULE_AUTHOR("EMC");
+MODULE_DESCRIPTION("EMC CTD V1 - Build 18-Jan-2016");
+MODULE_VERSION(EMCCTD_MODULE_VERSION);
+
+int ctd_debug;
+module_param_named(ctd_debug, ctd_debug, int, S_IRUGO | S_IWUSR);
+MODULE_PARM_DESC(ctd_debug, "Enable driver debug messages(0=off, 1=on)");
+
+static int emcctd_max_luns = EMCCTD_MAX_LUN;
+module_param_named(max_luns, emcctd_max_luns, int, S_IRUGO | S_IWUSR);
+MODULE_PARM_DESC(max_luns,
+ "Specify the maximum number of LUN's per host(default=16384)");
+
+static int emcctd_cmd_per_lun = EMCCTD_CMD_PER_LUN;
+module_param_named(cmd_per_lun, emcctd_cmd_per_lun, int, S_IRUGO | S_IWUSR);
+MODULE_PARM_DESC(cmd_per_lun,
+ "Specify the maximum commands per lun(default=64)");
+
+/*
+ * routines to handle scsi cmd buffer
+ */
+static int ctd_xmit_command(struct scsi_cmnd *cmnd,
+ struct ctd_pci_private *ctd_private);
+static void ctd_handle_scsi_response(
+ struct emc_ctd_v010_scsi_response *io_response,
+ struct ctd_pci_private *ctd_private);
+
+static int ctd_hw_enqueue_request(union emc_ctd_v010_message *ctd_request,
+ struct ctd_pci_private *ctd_private);
+static int ctd_hw_dequeue_response(union emc_ctd_v010_message *ctd_response,
+ struct ctd_pci_private *ctd_private);
+static int ctd_scsi_response_sanity_check(
+ struct ctd_request_private *request_private,
+ struct ctd_pci_private *ctd_private);
+
+/*
+ * routines to handle memory and queue management between client and server
+ */
+static int ctd_hw_execute_command(struct scsi_cmnd *cmnd,
+ struct ctd_pci_private *ctd_private);
+static int ctd_initiator_translate_sgl(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request,
+ struct ctd_pci_private *ctd_private);
+static int ctd_initiator_translate_request(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request,
+ struct ctd_pci_private *ctd_private);
+static void ctd_initiator_translate_lun(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request);
+
+static struct ctd_request_private *ctd_acquire_request(
+ struct ctd_pci_private *ctd_private);
+static void ctd_release_request(struct ctd_request_private *ctd_request,
+ struct ctd_pci_private *ctd_private);
+static void ctd_release_io_pool(struct ctd_pci_private *ctd_private);
+static int ctd_alloc_io_pool(struct ctd_pci_private *ctd_private,
+ unsigned int pool_size);
+static void ctd_check_error_condition(struct pci_dev *pci_dev);
+static int ctd_init_event_thread(struct ctd_pci_private *ctd_private);
+static void ctd_destroy_event_thread(struct ctd_pci_private *ctd_private);
+/*
+ * routines related to tasklet functionality
+ */
+static void ctd_check_response_queue(unsigned long instance_addr);
+
+static int ctd_ITnexus_handler(struct ctd_pci_private *ctd_private);
+/*
+ * routines related to internal representation of scsi targets
+ */
+static int ctd_target_alloc(struct scsi_target *starget);
+static void ctd_target_destroy(struct scsi_target *starget);
+/*
+ * routines registered with the linux scsi midlayer
+ */
+static int ctd_queuecommand_lck(struct scsi_cmnd *cmnd,
+ void (*done)(struct scsi_cmnd *));
+static DEF_SCSI_QCMD(ctd_queuecommand)
+static int ctd_slave_configure(struct scsi_device *sdp);
+static int ctd_slave_alloc(struct scsi_device *sdev);
+static void ctd_slave_destroy(struct scsi_device *sdev);
+static int ctd_abort_handler(struct scsi_cmnd *cmd);
+
+static enum blk_eh_timer_return ctd_timeout_handler(struct scsi_cmnd *scmd);
+
+
+/*
+ * routines related to initialization of the pseudo hardware
+ */
+static void ctd_init_scsi_host_private(struct Scsi_Host *shost,
+ struct pci_dev *pci_dev);
+
+static int ctd_scsi_layer_init(struct pci_dev *pci_dev);
+static int ctd_scsi_layer_cleanup(struct pci_dev *pci_dev);
+
+#ifdef CONFIG_PM
+static int ctd_pci_suspend(struct pci_dev *pci_dev, pm_message_t state);
+static int ctd_pci_resume(struct pci_dev *pci_dev);
+#endif
+
+static void ctd_pci_remove(struct pci_dev *pci_dev);
+static int ctd_request_msi(struct pci_dev *pci_dev);
+static int ctd_pci_probe(struct pci_dev *pci_dev,
+ const struct pci_device_id *id);
+static void ctd_clear_io_queue(struct ctd_pci_private *ctd_private);
+
+#if !defined(__VMKLNX__)
+static int ctd_proc_init(struct pci_dev *pci_dev);
+static void ctd_proc_remove(struct pci_dev *pci_dev);
+
+static int ctd_proc_open(struct inode *inode, struct file *file);
+#endif
+
+static int ctd_post_event(union emc_ctd_v010_message *io_msg,
+ struct ctd_pci_private *ctd_private);
+static int ctd_event_handler(void *ctd_thread_args);
+
+unsigned int lun_discovery_complete;
+wait_queue_head_t lun_discovery_event_barrier;
+
+static const struct pci_device_id ctd_pci_id_table[] = {
+ { EMC_CTD_PCI_VENDOR, EMC_CTD_V010_PCI_PRODUCT,
+ PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 },
+ { 0 },
+};
+
+MODULE_DEVICE_TABLE(pci, ctd_pci_id_table);
+
+static struct pci_driver ctd_pci_driver = {
+ .name = "emcctd",
+ .id_table = ctd_pci_id_table,
+ .probe = ctd_pci_probe,
+ .remove = ctd_pci_remove,
+#ifdef CONFIG_PM
+ .suspend = ctd_pci_suspend,
+ .resume = ctd_pci_resume,
+#endif
+};
+
+static struct scsi_host_template scsi_ctd_template = {
+ .name = DRV_NAME,
+ .proc_name = DRV_NAME,
+ .queuecommand = ctd_queuecommand,
+
+ .eh_timed_out = ctd_timeout_handler,
+
+ .slave_alloc = ctd_slave_alloc,
+ .slave_configure = ctd_slave_configure,
+ .slave_destroy = ctd_slave_destroy,
+ .eh_abort_handler = ctd_abort_handler,
+ .target_alloc = ctd_target_alloc,
+ .target_destroy = ctd_target_destroy,
+ .can_queue = EMCCTD_CMD_PER_LUN,
+ .this_id = EMCCTD_THIS_ID,
+ .sg_tablesize = SG_ALL,
+ .max_sectors = SCSI_DEFAULT_MAX_SECTORS,
+ .cmd_per_lun = EMCCTD_CMD_PER_LUN,
+ .use_clustering = DISABLE_CLUSTERING,
+ .module = THIS_MODULE,
+};
+
+
+#if !defined(__VMKLNX__)
+static struct proc_dir_entry *ctd_proc_directory;
+
+static int
+ctd_proc_show(struct seq_file *m, void *v)
+{
+ int i;
+ struct ctd_pci_private *ctd_private =
+ (struct ctd_pci_private *)m->private;
+
+ seq_printf(m,
+ "number interrupts: %ld\n"
+ "requests queued: %ld\n"
+ "responses received: %ld\n"
+ "pending IO count: %ld\n"
+ "Abort Sent: %ld\n"
+ "Abort received: %ld\n"
+ "What received: %ld\n"
+ "What sent: %ld\n"
+ "free IO entries : %ld\n"
+ "CTD WWN: %x.%x.%x.%x.%x.%x.%x.%x\n",
+ ctd_private->hw_stats.interrupts.counter,
+ ctd_private->hw_stats.requests_sent.counter,
+ ctd_private->hw_stats.responses_received.counter,
+ ctd_private->hw_stats.active_io_count.counter,
+ ctd_private->hw_stats.abort_sent.counter,
+ ctd_private->hw_stats.abort_received.counter,
+ ctd_private->hw_stats.what_in.counter,
+ ctd_private->hw_stats.what_out.counter,
+ ctd_private->hw_stats.free_io_entries.counter,
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[0],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[1],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[2],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[3],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[4],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[5],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[6],
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[7]);
+
+#define MAX_ENTRIES_IN_LINE 10
+ seq_puts(m, "\nIO Latency (in tsc) for last 200 IOs:\n");
+
+ for (i = 0; i < CTD_MAX_IO_STATS; i++) {
+ if (0 == (i % MAX_ENTRIES_IN_LINE))
+ seq_puts(m, "\n");
+
+ seq_printf(m, "%lld \t", ctd_private->hw_stats.io_stats[i]);
+ }
+ seq_puts(m, "\n");
+
+ return 0;
+
+}
+
+#endif
+
+static void
+scsi_translate_sam_code(struct scsi_cmnd *cmnd, unsigned char ScsiStatus)
+{
+ unsigned char host_status;
+ unsigned char driver_status;
+
+ host_status = DID_OK;
+ driver_status = DRIVER_OK;
+ cmnd->result |= ScsiStatus & 0xff;
+
+ switch (ScsiStatus) {
+
+ case SAM_STAT_GOOD:
+ case SAM_STAT_CONDITION_MET:
+ case SAM_STAT_INTERMEDIATE_CONDITION_MET:
+ case SAM_STAT_INTERMEDIATE:
+ break;
+ case SAM_STAT_CHECK_CONDITION:
+ case SAM_STAT_RESERVATION_CONFLICT:
+ case SAM_STAT_ACA_ACTIVE:
+ driver_status = DRIVER_SENSE;
+ break;
+ case SAM_STAT_TASK_SET_FULL:
+ case SAM_STAT_BUSY:
+ driver_status = DRIVER_BUSY;
+ host_status = DID_REQUEUE;
+ break;
+ case SAM_STAT_TASK_ABORTED:
+ driver_status = DRIVER_ERROR;
+ host_status = DID_ABORT;
+ break;
+ case SAM_STAT_COMMAND_TERMINATED:
+ default:
+ ctd_dprintk_crit(
+ "cmnd = %p [ channel:%d id:%d lun:%lld] INVALID SAM = %x\n",
+ cmnd, cmnd->device->channel, cmnd->device->id,
+ cmnd->device->lun, ScsiStatus);
+
+ driver_status = DRIVER_INVALID;
+ host_status = DID_ABORT;
+ break;
+ }
+ set_driver_byte(cmnd, driver_status);
+ set_host_byte(cmnd, host_status);
+}
+
+static void
+scsi_free_ctd_request_private(struct ctd_request_private *request_private,
+ struct ctd_pci_private *ctd_private)
+{
+
+ if (request_private->cdb_page) {
+ __free_pages(request_private->cdb_page,
+ request_private->cdb_page_order);
+ }
+ if (request_private->sgllist_page) {
+ __free_pages(request_private->sgllist_page,
+ request_private->sgllist_page_order);
+ }
+
+ ctd_release_request(request_private, ctd_private);
+}
+
+static int
+ctd_handle_disconnect(struct emc_ctd_v010_detect *io_detect,
+ struct ctd_pci_private *ctd_private)
+{
+ int i;
+ int error;
+ struct ctd_target_info *ctd_target;
+ struct ctd_host_info *ctd_host;
+
+ error = -ENODEV;
+
+ ctd_dprintk_crit("\n");
+
+ ctd_host = ctd_private->host_private;
+
+ /* Current implementation only handles the disconnect of
+ * the target and not initiators
+ */
+ for (i = 0; i < EMCCTD_MAX_ID; i++) {
+ ctd_target = &ctd_host->target[i];
+
+ /* detect header address is used to uniquely identify the target
+ * for which the disconnect event has been posted by the server
+ */
+
+ if (ctd_target->ctd_detect.emc_ctd_detect_header_address ==
+ io_detect->emc_ctd_detect_header_address) {
+ /* check the current link status of the target */
+ if (ctd_target->ctd_detect.emc_ctd_v010_detect_flags) {
+ ctd_dprintk_crit("\n");
+
+ ctd_target->ctd_detect.emc_ctd_v010_detect_flags =
+ io_detect->emc_ctd_v010_detect_flags;
+
+ ctd_check_error_condition(ctd_private->pci_dev);
+ } else {
+ ctd_dprintk_crit(
+ "target %llx already in disconnect state\n",
+ (emc_ctd_uint64_t)ctd_target->ctd_detect_name_bytes[0]);
+ }
+ error = 0;
+ break;
+ }
+ }
+
+ if (error)
+ ctd_dprintk_crit("Error\n");
+
+ return error;
+}
+
+static int
+ctd_handle_target_addition(struct emc_ctd_v010_detect *io_detect,
+ struct ctd_pci_private *ctd_private)
+{
+ int i;
+ int error;
+ struct ctd_target_info *ctd_target;
+ struct ctd_host_info *ctd_host;
+
+ error = -ENOMEM;
+ ctd_host = ctd_private->host_private;
+
+ ctd_dprintk("header addr -> %x key -> %llx\n",
+ io_detect->emc_ctd_detect_header_address,
+ io_detect->emc_ctd_v010_detect_key);
+
+ for (i = 0; i < EMCCTD_MAX_ID; i++) {
+ ctd_target = &ctd_host->target[i];
+
+ /* detect header address is used to uniquely identify the target
+ * for which the connect event has been posted by the server
+ * check if this particular target is already recorded with the
+ * client check if the recorded target is in the correct state
+ * and if not found record this particular target in the list of
+ * the detected targets
+ */
+
+ if (ctd_target->ctd_detect.emc_ctd_v010_detect_key ==
+ io_detect->emc_ctd_v010_detect_key) {
+ error = 0;
+ ctd_dprintk("\n");
+
+ scsi_target_unblock(&ctd_target->starget->dev,
+ SDEV_RUNNING);
+ ctd_target->ctd_detect.emc_ctd_v010_detect_flags =
+ io_detect->emc_ctd_v010_detect_flags;
+ break;
+ }
+
+ /* End of list for the recorded targets in the client, so the
+ * reported target is a new target being reported by the server
+ * thus needs to be added to the list
+ */
+ if (ctd_target->ctd_detect.emc_ctd_v010_detect_flags == 0x0) {
+ error = 0;
+ ctd_dprintk("\n");
+ ctd_target->ctd_detect = *io_detect;
+ break;
+ }
+ }
+ if (error)
+ ctd_dprintk_crit("Error\n");
+
+ return error;
+}
+
+static int
+ctd_handle_source_addition(struct emc_ctd_v010_detect *io_detect,
+ struct ctd_pci_private *ctd_private)
+{
+ ctd_dprintk("functionality not supported\n");
+ return -ENODEV;
+}
+
+static int
+ctd_handle_detect(struct emc_ctd_v010_detect *io_detect,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+
+ /*
+ * We post the detect event in the event queue and return. While
+ * ctd_event_thread actually consumes the requests in the event queue.
+ * This is done to serialize the consecutive detect requests (disconnect
+ * followed by connect). This mechanism handles the situation when
+ * multiple detect comes in a quick succession. Also, there is a
+ * seperate thread and its queue for each adapter, so detect requests
+ * for different adapters shall be handled seperately.
+ */
+ error = ctd_post_event((union emc_ctd_v010_message *)io_detect,
+ ctd_private);
+
+ if (!error)
+ atomic_long_inc(&ctd_private->hw_stats.what_in);
+
+ return error;
+
+}
+
+static void
+ctd_handle_scsi_command(struct emc_ctd_v010_scsi_command *io_command,
+ struct ctd_pci_private *ctd_private)
+{
+ ctd_dprintk_crit("unsupported\n");
+}
+
+static void
+ctd_handle_scsi_phase(struct emc_ctd_v010_scsi_phase *io_phase,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+ unsigned long flags;
+ struct scsi_cmnd *cmnd;
+ struct ctd_request_private *request_private;
+
+ /* check the phase flag to confirm we have received correct phase msg */
+ if (io_phase->emc_ctd_v010_scsi_phase_flags &
+ EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET) {
+ ctd_dprintk_crit("SCSI_PHASE_FLAG_TARGET not supported\n");
+ return;
+ }
+
+ if (!(io_phase->emc_ctd_v010_scsi_phase_flags &
+ EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT)) {
+ ctd_dprintk_crit("scsi_phase_flags %x invalid\n",
+ io_phase->emc_ctd_v010_scsi_phase_flags);
+ return;
+ }
+
+ request_private = (struct ctd_request_private *)
+ ((uintptr_t)(io_phase->emc_ctd_v010_scsi_phase_opaque_rx));
+
+ error = ctd_scsi_response_sanity_check(request_private, ctd_private);
+
+ if (error)
+ return;
+
+ cmnd = request_private->cmnd;
+
+ ctd_dprintk_crit("SCSI_PHASE_FLAG_ABORT cmnd-> %p request -> %p\n",
+ cmnd, request_private);
+
+ atomic_long_inc(&ctd_private->hw_stats.abort_received);
+
+ switch (request_private->io_state) {
+
+ case CTD_IO_REQUEST_QUEUED:
+ case CTD_IO_REQUEST_REQUEUED:
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ list_del(&request_private->list);
+ atomic_long_dec(&ctd_private->hw_stats.active_io_count);
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+ break;
+
+ case CTD_IO_REQUEST_ABORTED:
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ list_del(&request_private->list);
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+
+ case CTD_IO_REQUEST_REPLY_AWAITED:
+ scsi_free_ctd_request_private(request_private, ctd_private);
+ return;
+
+ case CTD_IO_REQUEST_FREE:
+ case CTD_IO_REQUEST_INVALID:
+ default:
+ ctd_dprintk_crit("opaque @ %p in unknown state %x\n",
+ request_private, request_private->io_state);
+ return;
+ }
+
+ cmnd->host_scribble = NULL;
+ request_private->cmnd = NULL;
+ request_private->io_state = CTD_IO_REQUEST_COMPLETED;
+ scsi_free_ctd_request_private(request_private, ctd_private);
+
+ /* error propagation to the SCSI midlayer */
+ scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED);
+ scsi_set_resid(cmnd, scsi_bufflen(cmnd));
+ cmnd->scsi_done(cmnd);
+}
+
+static void
+ctd_handle_response(union emc_ctd_v010_message *io_message,
+ struct ctd_pci_private *ctd_private)
+{
+ struct emc_ctd_v010_header *msg_header;
+
+ msg_header = &io_message->emc_ctd_v010_message_header;
+
+ switch (msg_header->emc_ctd_v010_header_what) {
+
+ case EMC_CTD_V010_WHAT_DETECT:
+ ctd_handle_detect((struct emc_ctd_v010_detect *)io_message,
+ ctd_private);
+ break;
+ case EMC_CTD_V010_WHAT_SCSI_COMMAND:
+ ctd_handle_scsi_command(
+ (struct emc_ctd_v010_scsi_command *)io_message,
+ ctd_private);
+ break;
+ case EMC_CTD_V010_WHAT_SCSI_PHASE:
+ ctd_handle_scsi_phase(
+ (struct emc_ctd_v010_scsi_phase *) io_message,
+ ctd_private);
+ break;
+ case EMC_CTD_V010_WHAT_SCSI_RESPONSE:
+ ctd_handle_scsi_response(
+ (struct emc_ctd_v010_scsi_response *)io_message,
+ ctd_private);
+ break;
+ default:
+ ctd_dprintk_crit("unknown what -> %x ctd_private -> %p",
+ msg_header->emc_ctd_v010_header_what, ctd_private);
+ }
+}
+
+
+static int
+ctd_scsi_response_sanity_check(struct ctd_request_private *request_private,
+ struct ctd_pci_private *ctd_private)
+{
+ int rc;
+ unsigned long flags;
+ struct scsi_cmnd *cmnd;
+ struct ctd_request_private *request, *request_next;
+ int outstanding_io_count = 0;
+
+ rc = -EFAULT;
+
+ /* check if the opaque is within the valid range */
+ if (!((request_private >= ctd_private->io_map) &&
+ (request_private < ctd_private->io_map_end))) {
+ ctd_dprintk_crit(
+ "ERROR request_private -> %p in invalid range\n",
+ request_private);
+ goto ctd_scsi_response_sanity_check_complete;
+ }
+
+ if (request_private) {
+ cmnd = request_private->cmnd;
+ if (cmnd) {
+ /* check if the back pointer is valid before we declare
+ * request sane
+ */
+ if ((request_private == (struct ctd_request_private
+ *)cmnd->host_scribble)) {
+ rc = 0;
+ goto ctd_scsi_response_sanity_check_complete;
+ }
+ }
+ /* check if the request has already been completed to the SCSI
+ * Midlayer
+ */
+ else if ((request_private->io_state ==
+ CTD_IO_REQUEST_ABORTED) ||
+ (request_private->io_state ==
+ CTD_IO_REQUEST_REPLY_AWAITED)) {
+ rc = 0;
+ goto ctd_scsi_response_sanity_check_complete;
+ }
+
+
+ ctd_dprintk_crit(
+ "ERROR cmnd -> %p mismatched request_private -> %p host_scribble -> %p requests send -> %ld responses received -> %ld state -> %s\n",
+ cmnd, request_private,
+ (cmnd ? cmnd->host_scribble : 0x0),
+ ctd_private->hw_stats.requests_sent.counter,
+ ctd_private->hw_stats.responses_received.counter,
+ (request_private->io_state ==
+ CTD_IO_REQUEST_QUEUED ?
+ "CTD_IO_REQUEST_QUEUED" :
+ (request_private->io_state ==
+ CTD_IO_REQUEST_REQUEUED ?
+ "CTD_IO_REQUEST_REQUEUED" :
+ (request_private->io_state ==
+ CTD_IO_REQUEST_ABORTED ?
+ "CTD_IO_REQUEST_ABORTED" :
+ (request_private->io_state ==
+ CTD_IO_REQUEST_FREE ?
+ "CTD_IO_REQUEST_FREE" :
+ (request_private->io_state ==
+ CTD_IO_REQUEST_INVALID ?
+ "CTD_IO_REQUEST_INVALID" :
+ (request_private->io_state ==
+ CTD_IO_REQUEST_COMPLETED ?
+ "CTD_IO_REQUEST_COMPLETED" :
+ (request_private->io_state ==
+ CTD_IO_REQUEST_REPLY_AWAITED ?
+ "CTD_IO_REQUEST_REPLY_AWAITED" :
+ "UNKNOWN"))))))));
+
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->aborted_io_list, list) {
+ if (request == request_private) {
+ ctd_dprintk_crit(
+ "request_private -> %p in aborted_io_list\n",
+ request_private);
+ break;
+ }
+ }
+
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->queued_io_list, list) {
+ if (request == request_private) {
+ ctd_dprintk_crit(
+ "request_private -> %p in queued_io_list\n",
+ request_private);
+ }
+ outstanding_io_count++;
+ }
+ ctd_dprintk_crit("outstanding_io_count = %d\n",
+ outstanding_io_count);
+
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->requeued_io_list, list) {
+ if (request == request_private) {
+ ctd_dprintk_crit(
+ "request_private -> %p in requeued_io_list\n",
+ request_private);
+ break;
+ }
+ }
+
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->io_pool, list) {
+ if (request == request_private) {
+ ctd_dprintk_crit(
+ "request_private -> %p in free io_pool\n",
+ request_private);
+ break;
+ }
+ }
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+ } else {
+ ctd_dprintk_crit("ERROR request_private -> NULL\n");
+ }
+
+ctd_scsi_response_sanity_check_complete:
+ return rc;
+}
+
+static void
+ctd_handle_scsi_response(struct emc_ctd_v010_scsi_response *io_response,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+ unsigned long flags;
+ struct scsi_cmnd *cmnd;
+ struct ctd_request_private *request_private;
+ unsigned long long io_stats;
+
+ request_private = (struct ctd_request_private *)
+ ((uintptr_t)(io_response->emc_ctd_v010_scsi_response_opaque));
+
+ error = ctd_scsi_response_sanity_check(request_private, ctd_private);
+
+ if (error)
+ return;
+
+ io_stats = ctd_read_tsc();
+
+ io_stats -= request_private->io_start_time;
+
+ ctd_private->hw_stats.io_stats[ctd_private->hw_stats.io_stats_index++] =
+ io_stats;
+
+ if (ctd_private->hw_stats.io_stats_index == CTD_MAX_IO_STATS)
+ ctd_private->hw_stats.io_stats_index = 0;
+ /*
+ * check if the IO context is in the aborted IO list
+ */
+
+ /* Increment the responses_received stats */
+ atomic_long_inc(&ctd_private->hw_stats.responses_received);
+
+ switch (request_private->io_state) {
+
+ /*
+ * The state of the request is important
+ * CTD_IO_REQUEST_QUEUED : cmnd is still alive and valid in kernel
+ * CTD_IO_REQUEST_ABORTED : cmnd has already been handled before
+ * the response from the device and
+ * only request needs to be cleaned
+ * up from the abort list
+ * CTD_IO_REQUEST_FREE : represents a state which is unhandled (unknown)
+ * CTD_IO_REQUEST_REPLY_AWAITED : represents a state where abort
+ * could not be sent by the timeout handler
+ * CTD_IO_REQUEST_INVALID : represents a state which is
+ * unhandled (unknown)
+ */
+
+ case CTD_IO_REQUEST_QUEUED:
+ case CTD_IO_REQUEST_REQUEUED:
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ list_del(&request_private->list);
+ request_private->io_state = CTD_IO_REQUEST_COMPLETED;
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+ break;
+
+ case CTD_IO_REQUEST_ABORTED:
+ /*
+ * cmnd is already disassociated from the request private and IO
+ * completed to the SCSI Midlayer by the timeout|abort handler
+ * delink the request private from the aborted list and cleanup
+ */
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+ list_del(&request_private->list);
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+
+ case CTD_IO_REQUEST_REPLY_AWAITED:
+ /*
+ * return the context back to the io_pool for its reuse
+ */
+ request_private->io_state = CTD_IO_REQUEST_COMPLETED;
+ scsi_free_ctd_request_private(request_private, ctd_private);
+ /* IO already completed to the Midlayer */
+ return;
+
+ case CTD_IO_REQUEST_FREE:
+ case CTD_IO_REQUEST_INVALID:
+ default:
+ ctd_dprintk_crit(
+ "opaque @ %p in unknown state %x\n",
+ request_private, request_private->io_state);
+ return;
+
+ }
+
+ /* Decrement active_io_count only when request is still queued. */
+ atomic_long_dec(&ctd_private->hw_stats.active_io_count);
+
+ cmnd = request_private->cmnd;
+
+ cmnd->result = 0;
+
+ scsi_translate_sam_code(cmnd,
+ io_response->emc_ctd_v010_scsi_response_status);
+
+ scsi_set_resid(cmnd, scsi_bufflen(cmnd) -
+ io_response->emc_ctd_v010_scsi_response_data_size);
+
+ if (io_response->emc_ctd_v010_scsi_response_flags &
+ EMC_CTD_V010_SCSI_RESPONSE_FLAG_FAILED) {
+ ctd_dprintk_crit(
+ "cmnd = %p CTDCM_FAILED channel:%d id:%d lun:%lld] status = %x\n",
+ cmnd, cmnd->device->channel, cmnd->device->id,
+ cmnd->device->lun,
+ io_response->emc_ctd_v010_scsi_response_status);
+
+ set_host_byte(cmnd, DID_ERROR);
+ if (ctd_debug) {
+ scsi_print_command(cmnd);
+ scsi_print_result(cmnd, NULL, FAILED);
+ scsi_print_sense(cmnd);
+ }
+ }
+ if (io_response->emc_ctd_v010_scsi_response_flags &
+ EMC_CTD_V010_SCSI_RESPONSE_FLAG_SENSE) {
+ unsigned char *sense_data;
+ unsigned char sense_data_length;
+
+ sense_data = io_response->emc_ctd_scsi_response_extra;
+ sense_data_length =
+ io_response->emc_ctd_v010_scsi_response_extra_size;
+
+ memcpy(cmnd->sense_buffer, sense_data, sense_data_length);
+
+ set_driver_byte(cmnd, DRIVER_SENSE);
+ if (ctd_debug) {
+ scsi_print_command(cmnd);
+ scsi_print_result(cmnd, "emcctd sense", SUCCESS);
+ scsi_print_sense(cmnd);
+ }
+ }
+
+ if ((io_response->emc_ctd_v010_scsi_response_status ==
+ SAM_STAT_TASK_SET_FULL) ||
+ (io_response->emc_ctd_v010_scsi_response_status ==
+ SAM_STAT_BUSY)) {
+
+ ctd_dprintk(
+ "QUEUE DEPTH change for channel:%d id:%d lun:%lld] active io count = %lx\n",
+ cmnd->device->channel, cmnd->device->id,
+ cmnd->device->lun,
+ ctd_private->hw_stats.active_io_count.counter);
+
+ scsi_track_queue_full(cmnd->device,
+ ctd_private->hw_stats.active_io_count.counter);
+ }
+
+
+ cmnd->host_scribble = NULL;
+
+ scsi_free_ctd_request_private(request_private, ctd_private);
+
+ cmnd->scsi_done(cmnd);
+}
+
+static void
+ctd_scsi_transfer_info(unsigned char *cmd, unsigned long long *lba,
+ unsigned int *num, unsigned int *ei_lba)
+{
+ *ei_lba = 0;
+
+ switch (*cmd) {
+
+ case VARIABLE_LENGTH_CMD:
+ *lba = (u64)cmd[19] | (u64)cmd[18] << 8 |
+ (u64)cmd[17] << 16 | (u64)cmd[16] << 24 |
+ (u64)cmd[15] << 32 | (u64)cmd[14] << 40 |
+ (u64)cmd[13] << 48 | (u64)cmd[12] << 56;
+
+ *ei_lba = (u32)cmd[23] | (u32)cmd[22] << 8 |
+ (u32)cmd[21] << 16 | (u32)cmd[20] << 24;
+
+ *num = (u32)cmd[31] | (u32)cmd[30] << 8 | (u32)cmd[29] << 16 |
+ (u32)cmd[28] << 24;
+ break;
+
+ case WRITE_SAME_16:
+ case WRITE_16:
+ case READ_16:
+ *lba = (u64)cmd[9] | (u64)cmd[8] << 8 |
+ (u64)cmd[7] << 16 | (u64)cmd[6] << 24 |
+ (u64)cmd[5] << 32 | (u64)cmd[4] << 40 |
+ (u64)cmd[3] << 48 | (u64)cmd[2] << 56;
+
+ *num = (u32)cmd[13] | (u32)cmd[12] << 8 | (u32)cmd[11] << 16 |
+ (u32)cmd[10] << 24;
+ break;
+ case WRITE_12:
+ case READ_12:
+ *lba = (u32)cmd[5] | (u32)cmd[4] << 8 | (u32)cmd[3] << 16 |
+ (u32)cmd[2] << 24;
+
+ *num = (u32)cmd[9] | (u32)cmd[8] << 8 | (u32)cmd[7] << 16 |
+ (u32)cmd[6] << 24;
+ break;
+ case WRITE_SAME:
+ case WRITE_10:
+ case READ_10:
+ case XDWRITEREAD_10:
+ *lba = (u32)cmd[5] | (u32)cmd[4] << 8 | (u32)cmd[3] << 16 |
+ (u32)cmd[2] << 24;
+
+ *num = (u32)cmd[8] | (u32)cmd[7] << 8;
+ break;
+ case WRITE_6:
+ case READ_6:
+ *lba = (u32)cmd[3] | (u32)cmd[2] << 8 |
+ (u32)(cmd[1] & 0x1f) << 16;
+ *num = (cmd[4] == 0) ? 256 : cmd[4];
+ break;
+ default:
+ break;
+ }
+}
+
+
+static int
+ctd_initiator_validate_sgl(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request,
+ struct ctd_pci_private *ctd_private)
+{
+ int i;
+ int error;
+ struct emc_ctd_v010_sgl *sgl_current;
+ struct emc_ctd_v010_sgl *sgl_extended;
+ emc_ctd_uint64_t sgl_count;
+ unsigned int sgl_buffersize;
+ unsigned long long lba;
+ unsigned int num;
+ unsigned int ei_lba;
+
+ lba = 0;
+ num = 0;
+ ei_lba = 0;
+ error = -EINVAL;
+ sgl_count = 0;
+ sgl_extended = &ctd_request->emc_ctd_scsi_command_sgl[0];
+
+ if (ctd_request->emc_ctd_v010_scsi_command_flags &
+ EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL) {
+ emc_ctd_uint64_t extented_sgl_size;
+
+ sgl_current = phys_to_virt(
+ ((emc_ctd_uint64_t)sgl_extended->emc_ctd_v010_sgl_paddr_32_63 << 32) |
+ ((emc_ctd_uint64_t)sgl_extended->emc_ctd_v010_sgl_paddr_0_31 & 0xFFFFFFFF));
+
+ extented_sgl_size = sgl_extended->emc_ctd_v010_sgl_size;
+
+ do_div(extented_sgl_size, sizeof(struct emc_ctd_v010_sgl));
+
+ sgl_count = extented_sgl_size;
+ } else {
+ sgl_current = sgl_extended;
+ while (sgl_current < sgl_extended +
+ EMC_CTD_V010_SGL_IMMEDIATE_MAX) {
+ if (sgl_current->emc_ctd_v010_sgl_size == 0)
+ break;
+
+ sgl_current++;
+ sgl_count++;
+ }
+ sgl_current = sgl_extended;
+ }
+
+ if (scsi_sg_count(cmnd) != sgl_count) {
+ scsi_print_command(cmnd);
+ ctd_dprintk_crit(
+ "Mismatched cmnd_sgl_count %d sgl_count = %lld sgl_size = %d\n",
+ scsi_sg_count(cmnd),
+ sgl_count, sgl_extended->emc_ctd_v010_sgl_size);
+ }
+
+ if (sgl_count) {
+ struct scatterlist *cmnd_sg;
+ int cmnd_sg_count = scsi_sg_count(cmnd);
+
+ if (cmnd_sg_count > EMC_CTD_V010_SGL_IMMEDIATE_MAX) {
+ if ((ctd_request->emc_ctd_v010_scsi_command_flags &
+ EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL) == 0) {
+ scsi_print_command(cmnd);
+ ctd_dprintk_crit(
+ "scsi_sg_count = %d but EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL not set\n",
+ scsi_sg_count(cmnd));
+
+ cmnd_sg_count = EMC_CTD_V010_SGL_IMMEDIATE_MAX;
+ }
+ }
+
+ scsi_for_each_sg(cmnd, cmnd_sg, cmnd_sg_count, i) {
+ int cmnd_sg_size;
+ emc_ctd_uint64_t cmnd_buffer_pfn;
+ struct page *cmnd_page;
+ emc_ctd_uint32_t cmnd_offset;
+ emc_ctd_uint32_t sgl_size;
+ emc_ctd_uint64_t buffer_pfn;
+
+ cmnd_page = sg_page(cmnd_sg);
+ cmnd_buffer_pfn = page_to_phys(cmnd_page);
+ cmnd_sg_size = cmnd_sg->length;
+ cmnd_offset = cmnd_sg->offset;
+
+
+ sgl_size = (sgl_current + i)->emc_ctd_v010_sgl_size;
+ buffer_pfn = (((emc_ctd_uint64_t)
+ ((sgl_current + i)->emc_ctd_v010_sgl_paddr_32_63) << 32) |
+ ((emc_ctd_uint64_t)((sgl_current + i)->emc_ctd_v010_sgl_paddr_0_31) & 0xFFFFFFFF));
+
+ if ((sgl_size != cmnd_sg_size) ||
+ (buffer_pfn != (cmnd_buffer_pfn +
+ cmnd_offset))) {
+ scsi_print_command(cmnd);
+ ctd_dprintk_crit(
+ "Mismatch @ i = %d cmnd_sg_size = %d cmnd_buffer_pfn = %llx sgl_size = %d buffer_pfn = %llx\n",
+ i, cmnd_sg_size, cmnd_buffer_pfn,
+ sgl_size, buffer_pfn);
+ }
+ }
+ goto ctd_initiator_validate_sgl_end;
+ }
+
+ for (i = 0, sgl_buffersize = 0; i < sgl_count; i++, sgl_current++)
+ sgl_buffersize += sgl_current->emc_ctd_v010_sgl_size;
+
+ if (scsi_bufflen(cmnd) && (sgl_buffersize != scsi_bufflen(cmnd))) {
+ scsi_print_command(cmnd);
+ ctd_dprintk_crit("Mismatched buffer size %d %d\n",
+ scsi_bufflen(cmnd), sgl_buffersize);
+ goto ctd_initiator_validate_sgl_end;
+ }
+
+ ctd_scsi_transfer_info((unsigned char *)cmnd->cmnd, &lba,
+ &num, &ei_lba);
+
+ if (num && (sgl_buffersize != (num * 512))) {
+ scsi_print_command(cmnd);
+ ctd_dprintk_crit("Mismatched buffer size %d %d\n",
+ (num * 512), sgl_buffersize);
+ goto ctd_initiator_validate_sgl_end;
+ }
+ error = 0;
+ctd_initiator_validate_sgl_end:
+ return error;
+}
+
+static int
+ctd_initiator_translate_sgl(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request,
+ struct ctd_pci_private *ctd_private)
+{
+ int i;
+ int size;
+ int error;
+ int sg_count;
+ int rq_count;
+ int embedded_sgl_count;
+ int embedded_sgl_index;
+ struct ctd_request_private *request_private;
+ emc_ctd_uint64_t buffer_pfn;
+
+ error = FAILED;
+ i = size = 0;
+ embedded_sgl_index = 0;
+ embedded_sgl_count = EMC_CTD_V010_SGL_IMMEDIATE_MAX;
+ rq_count = 0;
+
+ request_private = (struct ctd_request_private *)cmnd->host_scribble;
+ sg_count = scsi_sg_count(cmnd);
+
+ if (sg_count > embedded_sgl_count) {
+ struct scatterlist *sg;
+ struct emc_ctd_v010_sgl *sgl_current;
+ struct emc_ctd_v010_sgl *sgl_extended;
+
+ request_private->sgllist_page_order =
+ get_order((sizeof(struct emc_ctd_v010_sgl) * sg_count));
+ request_private->sgllist_page =
+ alloc_pages(GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN,
+ request_private->sgllist_page_order);
+
+ if (!request_private->sgllist_page) {
+ ctd_dprintk_crit("alloc_page failure\n");
+ goto scsi_initiator_translate_sgl_end;
+ }
+
+ sgl_current = (struct emc_ctd_v010_sgl *)
+ page_address(request_private->sgllist_page);
+
+ scsi_for_each_sg(cmnd, sg, sg_count, i) {
+#if defined(__VMKLNX__)
+ sgl_current->emc_ctd_v010_sgl_paddr_0_31 =
+ sg->cursgel->addr & 0xFFFFFFFF;
+ sgl_current->emc_ctd_v010_sgl_paddr_32_63 =
+ sg->cursgel->addr >> 32;
+ sgl_current->emc_ctd_v010_sgl_size = sg_dma_len(sg);
+#else
+ struct page *page;
+
+ page = sg_page(sg);
+ buffer_pfn = page_to_phys(page);
+ size += sg->length;
+
+ sgl_current->emc_ctd_v010_sgl_paddr_0_31 =
+ (buffer_pfn + sg->offset) & 0xFFFFFFFF;
+ sgl_current->emc_ctd_v010_sgl_paddr_32_63 =
+ (buffer_pfn + sg->offset) >> 32;
+ sgl_current->emc_ctd_v010_sgl_size = sg->length;
+#endif
+ sgl_current++;
+ rq_count++;
+ }
+ sgl_extended = &ctd_request->emc_ctd_scsi_command_sgl[0];
+ buffer_pfn = page_to_phys(request_private->sgllist_page);
+
+ sgl_extended->emc_ctd_v010_sgl_paddr_0_31 =
+ buffer_pfn & 0xFFFFFFFF;
+ sgl_extended->emc_ctd_v010_sgl_paddr_32_63 = buffer_pfn >> 32;
+ sgl_extended->emc_ctd_v010_sgl_size =
+ rq_count * sizeof(struct emc_ctd_v010_sgl);
+ ctd_request->emc_ctd_v010_scsi_command_flags |=
+ EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL;
+ } else {
+ struct scatterlist *sg;
+ struct emc_ctd_v010_sgl *sgl_current;
+
+ sgl_current = &ctd_request->emc_ctd_scsi_command_sgl[0];
+
+ scsi_for_each_sg(cmnd, sg, sg_count, i) {
+#if defined(__VMKLNX__)
+ sgl_current->emc_ctd_v010_sgl_paddr_0_31 =
+ sg->cursgel->addr & 0xFFFFFFFF;
+ sgl_current->emc_ctd_v010_sgl_paddr_32_63 =
+ sg->cursgel->addr >> 32;
+ sgl_current->emc_ctd_v010_sgl_size = sg_dma_len(sg);
+#else
+ struct page *page;
+
+ size += sg->length;
+ page = sg_page(sg);
+ buffer_pfn = page_to_phys(page);
+
+ sgl_current->emc_ctd_v010_sgl_paddr_0_31 =
+ (buffer_pfn + sg->offset) & 0xFFFFFFFF;
+ sgl_current->emc_ctd_v010_sgl_paddr_32_63 =
+ (buffer_pfn + sg->offset) >> 32;
+ sgl_current->emc_ctd_v010_sgl_size = sg->length;
+#endif
+
+ sgl_current++;
+ rq_count++;
+ }
+ }
+ error = SUCCESS;
+
+ if (ctd_debug)
+ ctd_initiator_validate_sgl(cmnd, ctd_request, ctd_private);
+
+scsi_initiator_translate_sgl_end:
+ return error;
+}
+
+
+void
+ctd_initiator_translate_lun(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request)
+{
+
+ int i;
+ union ctd_scsi_lun {
+ emc_ctd_uint64_t scsi_lun_64;
+ emc_ctd_uint8_t scsi_lun_8[8];
+ } ctd_scsi_lun_conversion;
+
+ ctd_scsi_lun_conversion.scsi_lun_64 = cpu_to_be64(cmnd->device->lun);
+ for (i = 0; i < sizeof(ctd_scsi_lun_conversion); i++) {
+ ctd_request->emc_ctd_v010_scsi_command_lun[i] =
+ ctd_scsi_lun_conversion.scsi_lun_8[i];
+ }
+}
+
+static int
+ctd_initiator_translate_request(struct scsi_cmnd *cmnd,
+ struct emc_ctd_v010_scsi_command *ctd_request,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+ struct ctd_dev_info *ctd_device;
+ struct ctd_request_private *request_private;
+ int scsi_cdb_size;
+
+ error = FAILED;
+ request_private = NULL;
+ ctd_device = cmnd->device->hostdata;
+
+ if (!(ctd_device->ctd_target_detect->emc_ctd_v010_detect_flags &
+ EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) {
+ goto scsi_initiator_translate_request_end;
+ }
+ memset(ctd_request, 0x0, sizeof(struct emc_ctd_v010_scsi_command));
+
+ ctd_request->emc_ctd_scsi_command_header_address =
+ ctd_device->ctd_target_detect->emc_ctd_detect_header_address;
+ ctd_request->emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_minor =
+ EMCCTD_V010_PROTOCOL_MINOR_VERSION;
+ ctd_request->emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_what =
+ EMC_CTD_V010_WHAT_SCSI_COMMAND;
+ ctd_request->emc_ctd_v010_scsi_command_flags |=
+ ((cmnd->sc_data_direction == DMA_FROM_DEVICE) ?
+ EMC_CTD_V010_SCSI_COMMAND_FLAG_SOURCE :
+ ((cmnd->sc_data_direction ==
+ DMA_TO_DEVICE) ? 0 : 0));
+
+ request_private = ctd_acquire_request(ctd_private);
+
+ if (request_private == NULL) {
+ ctd_dprintk_crit("ctd_acquire_request failure\n");
+ goto scsi_initiator_translate_request_end;
+ }
+
+ request_private->cmnd = cmnd;
+
+ cmnd->host_scribble = (unsigned char *)request_private;
+
+ ctd_request->emc_ctd_v010_scsi_command_data_size = scsi_bufflen(cmnd);
+
+ ctd_request->emc_ctd_v010_scsi_command_opaque =
+ (uintptr_t)request_private;
+
+ ctd_initiator_translate_lun(cmnd, ctd_request);
+
+ scsi_cdb_size = cmnd->cmd_len;
+
+ if (scsi_cdb_size <=
+ sizeof(ctd_request->emc_ctd_v010_scsi_command_cdb)) {
+ memcpy(ctd_request->emc_ctd_v010_scsi_command_cdb,
+ cmnd->cmnd, scsi_cdb_size);
+ } else {
+ ctd_dprintk_crit("unsupported scsi cdb size = %d\n",
+ scsi_cdb_size);
+ goto scsi_initiator_translate_request_end;
+ }
+
+ error = ctd_initiator_translate_sgl(cmnd, ctd_request, ctd_private);
+
+scsi_initiator_translate_request_end:
+ if (error == FAILED) {
+ cmnd->host_scribble = NULL;
+ if (request_private) {
+ scsi_free_ctd_request_private(request_private,
+ ctd_private);
+ }
+ }
+ return error;
+}
+
+
+static int
+ctd_hw_execute_command(struct scsi_cmnd *cmnd,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+ unsigned long flags;
+ struct emc_ctd_v010_scsi_command ctd_request;
+
+ error = ctd_initiator_translate_request(cmnd, &ctd_request,
+ ctd_private);
+
+ if (error == SUCCESS) {
+ struct ctd_request_private *request_private;
+
+ request_private =
+ (struct ctd_request_private *)cmnd->host_scribble;
+
+ /* lock ordering ... io_mgmt_lock followed by isr_lock
+ * ensures that request placed in legitimate queue ...
+ * so that response finds in the correct queue
+ */
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ request_private->io_start_time = ctd_read_tsc();
+
+ error = ctd_hw_enqueue_request(
+ (union emc_ctd_v010_message *)&ctd_request,
+ ctd_private);
+ if (error == SUCCESS) {
+ list_add_tail(&request_private->list,
+ &ctd_private->queued_io_list);
+ atomic_long_inc(&ctd_private->hw_stats.requests_sent);
+ atomic_long_inc(&ctd_private->hw_stats.active_io_count);
+ }
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+
+ if (error != SUCCESS) {
+ ctd_dprintk_crit("ctd_hw_enqueue_request error\n");
+ scsi_free_ctd_request_private(
+ (struct ctd_request_private *)
+ (uintptr_t)
+ (ctd_request.emc_ctd_v010_scsi_command_opaque),
+ ctd_private);
+ }
+ }
+
+ return error;
+}
+
+static int
+ctd_hw_enqueue_request(union emc_ctd_v010_message *ctd_request,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+ unsigned long flags;
+ union emc_ctd_v010_message *ctd_request_block;
+
+ spin_lock_irqsave(&ctd_private->isr_lock, flags);
+
+ /*
+ * check if any space is available in the array
+ */
+ ctd_request_block =
+ ((((ctd_private->request_producer_index + 1) %
+ ctd_private->pci_request_queue_size) ==
+ ctd_private->request_consumer_index) ?
+ NULL :
+ (ctd_private->pci_request_array +
+ ctd_private->request_producer_index));
+
+
+ error = (ctd_request_block ? SUCCESS : FAILED);
+
+ if (error == SUCCESS) {
+ *ctd_request_block = *ctd_request;
+
+ ctd_private->request_producer_index =
+ ((ctd_private->request_producer_index + 1) %
+ ctd_private->pci_request_queue_size);
+
+ }
+
+ spin_unlock_irqrestore(&ctd_private->isr_lock, flags);
+
+ return error;
+}
+
+static int
+ctd_hw_dequeue_response(union emc_ctd_v010_message *ctd_response,
+ struct ctd_pci_private *ctd_private)
+{
+ int rc;
+ unsigned long flags;
+ union emc_ctd_v010_message *ctd_response_block;
+
+ /* protect ourselves from another instance */
+ spin_lock_irqsave(&ctd_private->isr_lock, flags);
+
+ ctd_response_block =
+ ((ctd_private->response_consumer_index ==
+ ctd_private->response_producer_index) ? NULL :
+ (ctd_private->pci_response_array +
+ ctd_private->response_consumer_index));
+
+ rc = ctd_response_block ? SUCCESS : FAILED;
+
+ if (rc == SUCCESS) {
+ *ctd_response = *ctd_response_block;
+ ctd_private->response_consumer_index =
+ ((ctd_private->response_consumer_index + 1) %
+ ctd_private->pci_response_queue_size);
+ } else {
+ ctd_check_error_condition(ctd_private->pci_dev);
+ }
+
+ spin_unlock_irqrestore(&ctd_private->isr_lock, flags);
+
+ return rc;
+
+}
+
+static int
+ctd_xmit_command(struct scsi_cmnd *cmnd,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+
+ cmnd->result = DID_OK;
+ error = ctd_hw_execute_command(cmnd, ctd_private);
+ return error;
+}
+
+
+static int
+ctd_queuecommand_lck(struct scsi_cmnd *cmnd, void (*done)(struct scsi_cmnd *))
+{
+ int error;
+ struct ctd_host_info *ctd_host;
+ struct ctd_pci_private *ctd_private;
+
+ error = 0;
+ ctd_host = shost_priv(cmnd->device->host);
+ ctd_private = pci_get_drvdata(ctd_host->pci_dev);
+
+ switch (ctd_private->hw_state) {
+
+ case CTD_HW_STATE_INITIALIZED:
+ cmnd->scsi_done = done;
+ if (ctd_xmit_command(cmnd, ctd_private) == SUCCESS)
+ break;
+
+ case CTD_HW_STATE_DISABLED:
+ cmnd->scsi_done = done;
+ scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED);
+ scsi_set_resid(cmnd, scsi_bufflen(cmnd));
+ cmnd->scsi_done(cmnd);
+ break;
+ default:
+ error = SCSI_MLQUEUE_HOST_BUSY;
+ }
+
+ return error;
+}
+
+static int
+ctd_abort_handler(struct scsi_cmnd *cmnd)
+{
+ ctd_dprintk_crit("SCSI cmnd -> %p ERROR\n",
+ cmnd);
+ return SUCCESS;
+}
+
+
+static int
+ctd_target_alloc(struct scsi_target *starget)
+{
+ int error;
+ struct ctd_target_info *ctd_target;
+ struct ctd_host_info *ctd_host;
+
+ error = -ENODEV;
+ ctd_host = shost_priv(dev_to_shost(&starget->dev));
+
+ ctd_dprintk("starget -> %p id -> %x\n",
+ starget, starget->id);
+
+ ctd_target = &ctd_host->target[starget->id];
+
+ /* check for the connection status in the detect flag
+ * also check if the target already registered with the SCSI midlayer
+ */
+ if (ctd_target->starget == NULL &&
+ (ctd_target->ctd_detect.emc_ctd_v010_detect_flags &
+ EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) {
+ error = 0;
+ ctd_target->starget = starget;
+ starget->hostdata = ctd_target;
+ } else {
+ if (ctd_target->starget != starget) {
+ ctd_dprintk_crit(
+ "failure ctd_target->starget %p and starget %p dissimilar\n",
+ ctd_target->starget, starget);
+ } else {
+ ctd_dprintk_crit("failure starget %p unexpected\n",
+ starget);
+ }
+ }
+
+ return error;
+}
+
+void
+ctd_target_destroy(struct scsi_target *starget)
+{
+ int i;
+ int error;
+ struct ctd_target_info *ctd_target;
+ struct ctd_host_info *ctd_host;
+
+ error = -ENODEV;
+ ctd_host = shost_priv(dev_to_shost(&starget->dev));
+
+ ctd_dprintk_crit("starget @ id = %x\n", starget->id);
+
+ for (i = 0; i < EMCCTD_MAX_ID; i++) {
+ ctd_target = &ctd_host->target[i];
+ if (ctd_target->starget == starget) {
+ ctd_target->starget = NULL;
+ error = 0;
+ break;
+ }
+ }
+
+ if (error) {
+ ctd_dprintk_crit("failure for starget @ id = %x\n",
+ starget->id);
+ }
+}
+
+static int
+ctd_slave_configure(struct scsi_device *sdevice)
+{
+ int error;
+ struct ctd_dev_info *ctd_device;
+ struct ctd_host_info *ctd_host;
+
+ error = 0;
+ ctd_host = shost_priv(sdevice->host);
+ ctd_device = sdevice->hostdata;
+
+ /* tune the block layer to generate timout
+ * for the requests queued and reply is awaited
+ */
+ blk_queue_rq_timeout(sdevice->request_queue, EMCCTD_REQUEST_TIMEOUT);
+
+ return error;
+}
+
+
+
+static int
+ctd_slave_alloc(struct scsi_device *sdev)
+{
+ int error;
+ struct ctd_host_info *ctd_host;
+ struct ctd_target_info *ctd_target;
+ struct ctd_dev_info *ctd_device;
+
+ error = -ENOMEM;
+ ctd_host = shost_priv(sdev->host);
+ sdev->host->max_cmd_len = EMCCTD_V010_MAX_CDB_SIZE;
+
+ ctd_device = (struct ctd_dev_info *)
+ kzalloc(sizeof(struct ctd_dev_info), GFP_ATOMIC);
+ if (ctd_device == NULL)
+ goto ctd_slave_alloc_end;
+
+ ctd_target = &ctd_host->target[sdev->id];
+ if (ctd_target->starget)
+ ctd_device->ctd_target_detect = &ctd_target->ctd_detect;
+
+ if (ctd_device->ctd_target_detect) {
+ error = 0;
+ ctd_device->ctd_host = ctd_host;
+ ctd_device->ctd_target = ctd_target;
+ sdev->hostdata = ctd_device;
+ queue_flag_set_unlocked(QUEUE_FLAG_BIDI, sdev->request_queue);
+ } else {
+ kfree(ctd_device);
+ error = -ENODEV;
+ }
+
+ctd_slave_alloc_end:
+
+ if (error) {
+ ctd_dprintk_crit("channel = %x id= %x error = %x\n",
+ sdev->channel, sdev->id, error);
+ }
+ return error;
+}
+
+static void
+ctd_slave_destroy(struct scsi_device *sdev)
+{
+ struct ctd_dev_info *dev_info;
+
+ dev_info = sdev->hostdata;
+ kfree(dev_info);
+}
+
+static enum blk_eh_timer_return
+ctd_timeout_handler(struct scsi_cmnd *cmd)
+{
+ int requeue_error;
+ unsigned long flags;
+ enum blk_eh_timer_return error;
+ struct ctd_host_info *ctd_host;
+ struct ctd_pci_private *ctd_private;
+ struct ctd_request_private *request;
+ unsigned long long tsc_val;
+
+ requeue_error = FAILED;
+ error = BLK_EH_NOT_HANDLED;
+ ctd_host = shost_priv(cmd->device->host);
+ ctd_private = pci_get_drvdata(ctd_host->pci_dev);
+
+
+ request = (struct ctd_request_private *)cmd->host_scribble;
+
+ tsc_val = ctd_read_tsc();
+
+ tsc_val -= request->io_start_time;
+
+ ctd_dprintk_crit("cmnd -> %p request -> %p, tsc -> %lld\n",
+ cmd, request, tsc_val);
+
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ if (request && request->io_timeout < EMCCTD_MAX_RETRY) {
+ struct ctd_dev_info *ctd_device;
+
+ switch (request->io_state) {
+
+ /* check if the IO in the queued_io_list */
+ case CTD_IO_REQUEST_QUEUED:
+ /* check if the IO is already requeued */
+ case CTD_IO_REQUEST_REQUEUED:
+ /* remove the old IO context from the requeued_io_list
+ * or queued_io_list
+ */
+ list_del(&request->list);
+ atomic_long_dec(&ctd_private->hw_stats.active_io_count);
+
+ ctd_device = cmd->device->hostdata;
+
+ if (!(ctd_device->ctd_target_detect->emc_ctd_v010_detect_flags &
+ EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) {
+ ctd_dprintk_crit("device diconnected\n");
+ } else {
+ union emc_ctd_v010_message ctd_message;
+ struct emc_ctd_v010_scsi_phase *ctd_phase;
+
+ memset(&ctd_message, 0x0,
+ sizeof(union emc_ctd_v010_message));
+ ctd_phase =
+ (struct emc_ctd_v010_scsi_phase *) &
+ ctd_message.emc_ctd_v010_message_scsi_phase;
+
+ /* Need to acertain if this is how an IO is
+ * aborted by the specification
+ */
+
+ /*
+ * OPT-438489 the phase flag needs to be
+ * initialized with PHASE_FLAG_TARGET If EMC
+ * CTD V010 SCSI PHASE FLAG TARGET is set, the
+ * message receiver is the target, otherwise the
+ * message receiver is the initiator. If EMC CTD
+ * V010 SCSI PHASE FLAG ABORT is set, the SCSI
+ * command is aborted.
+ */
+ ctd_phase->emc_ctd_v010_scsi_phase_flags =
+ EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT |
+ EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET;
+
+ ctd_phase->emc_ctd_v010_scsi_phase_opaque_tx =
+ (uintptr_t)request;
+ ctd_phase->emc_ctd_v010_scsi_phase_opaque_rx =
+ (-1);
+ ctd_phase->emc_ctd_scsi_phase_header_what =
+ EMC_CTD_V010_WHAT_SCSI_PHASE;
+ ctd_phase->emc_ctd_scsi_phase_header_minor =
+ EMCCTD_V010_PROTOCOL_MINOR_VERSION;
+ ctd_phase->emc_ctd_scsi_phase_header_address =
+ ctd_device->ctd_target_detect->emc_ctd_detect_header_address;
+
+ requeue_error =
+ ctd_hw_enqueue_request(&ctd_message,
+ ctd_private);
+ }
+
+
+ if (requeue_error != SUCCESS) {
+ /* add the IO context to requeued_io_list */
+ /* Client would try to abort the request in next
+ * timeout (after 20 seconds)
+ */
+ request->io_state = CTD_IO_REQUEST_REQUEUED;
+ list_add_tail(&request->list,
+ &ctd_private->requeued_io_list);
+ request->io_timeout++;
+ error = BLK_EH_RESET_TIMER;
+ } else {
+ request->cmnd = NULL;
+ cmd->host_scribble = NULL;
+
+ request->io_state = CTD_IO_REQUEST_ABORTED;
+ request->purge_lifetime =
+ jiffies + EMCCTD_OPAQUE_PURGE_WAITTIME;
+ list_add_tail(&request->list,
+ &ctd_private->aborted_io_list);
+ atomic_long_inc(&ctd_private->hw_stats.abort_sent);
+
+ /* error propagation to the SCSI midlayer */
+ scsi_translate_sam_code(cmd,
+ SAM_STAT_TASK_ABORTED);
+ scsi_set_resid(cmd, scsi_bufflen(cmd));
+
+ /* indicate no more requeue of this particular
+ * IO is needed
+ */
+ error = BLK_EH_HANDLED;
+ }
+ break;
+
+ default:
+ ctd_dprintk_crit(
+ "request @ %p in unexpected state %x\n",
+ request, request->io_state);
+ break;
+ }
+ } else if (request) {
+ ctd_dprintk_crit("cmd %p timeout completed io_state %x\n",
+ cmd, request->io_state);
+
+ /* remove the old IO context from the requeued_io_list */
+ list_del(&request->list);
+
+ /*
+ * break the context between the cmnd and the request and
+ * request on the requeue_io_list cannot be reused until
+ * server replies for the same
+ */
+ request->cmnd = NULL;
+ cmd->host_scribble = NULL;
+
+ /* error propagation to the SCSI midlayer */
+ scsi_translate_sam_code(cmd, SAM_STAT_TASK_ABORTED);
+ scsi_set_resid(cmd, scsi_bufflen(cmd));
+
+ /* we can deallocate the context only once we receive
+ * any reply from the server
+ */
+ request->io_state = CTD_IO_REQUEST_REPLY_AWAITED;
+
+ /* indicate no more requeue of this particular IO is needed */
+ error = BLK_EH_HANDLED;
+ } else {
+ ctd_dprintk_crit(
+ "cmnd -> %p request -> NULL error !!!\n", cmd);
+ }
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+
+ return error;
+}
+
+
+static int
+ctd_ITnexus_handler(struct ctd_pci_private *ctd_private)
+{
+ int i;
+ int error;
+ struct ctd_target_info *ctd_target;
+ struct ctd_host_info *ctd_host;
+
+ error = 0;
+
+ ctd_host = (struct ctd_host_info *)ctd_private->host_private;
+
+ ctd_dprintk_crit("ctd_private -> %p\n", ctd_private);
+
+ for (i = 0; i < EMCCTD_MAX_ID; i++) {
+ union emc_ctd_v010_message ctd_message;
+ struct emc_ctd_v010_detect *ctd_detect;
+
+ ctd_target = &ctd_host->target[i];
+
+ switch (ctd_target->detect_completed) {
+
+ case EMCCTD_TARGET_DETECT_NOT_COMPLETED:
+ if (!ctd_target->ctd_detect.emc_ctd_v010_detect_flags)
+ break;
+
+ /* The id defined by the SCSI Midlayer should match the
+ * index as this routine is indirectly invoked by the
+ * delayed work mechanism
+ */
+
+ ctd_dprintk_crit("ctd_target -> %p index = %x\n",
+ ctd_target, i);
+ ctd_dprintk_crit("key -> %llx header -> %x\n",
+ ctd_target->ctd_detect.emc_ctd_v010_detect_key,
+ ctd_target->ctd_detect.emc_ctd_detect_header_address);
+
+ memset(&ctd_message, 0x0,
+ sizeof(union emc_ctd_v010_message));
+
+ ctd_detect = &ctd_message.emc_ctd_v010_message_detect;
+ ctd_detect->emc_ctd_v010_detect_flags = 0x0;
+ ctd_detect->emc_ctd_v010_detect_key =
+ ctd_target->ctd_detect.emc_ctd_v010_detect_key;
+ ctd_detect->emc_ctd_detect_header_what =
+ EMC_CTD_V010_WHAT_DETECT;
+ ctd_detect->emc_ctd_detect_header_minor =
+ EMCCTD_V010_PROTOCOL_MINOR_VERSION;
+ ctd_detect->emc_ctd_detect_header_address =
+ ctd_target->ctd_detect.emc_ctd_detect_header_address;
+
+ if (ctd_hw_enqueue_request(&ctd_message,
+ ctd_private) == SUCCESS) {
+ ctd_dprintk_crit("ctd_target -> %p\n",
+ ctd_target);
+ ctd_target->detect_completed =
+ EMCCTD_TARGET_DETECT_COMPLETED;
+
+ atomic_long_inc(
+ &ctd_private->hw_stats.what_out);
+ } else {
+ ctd_dprintk_crit(
+ "ctd_target -> %p ctd_hw_enqueue_request failure\n",
+ ctd_target);
+ error = -EAGAIN;
+ break;
+ }
+
+ case EMCCTD_TARGET_DETECT_COMPLETED:
+ /* Disconnect case ... we need to remove the associated
+ * objects from SCSI midlayer
+ */
+ if (!ctd_target->ctd_detect.emc_ctd_v010_detect_flags) {
+ ctd_dprintk_crit("ctd_target -> %p\n",
+ ctd_target);
+
+ ctd_clear_io_queue(ctd_private);
+
+ if (ctd_target->starget) {
+ /* the following attempts to clean the SCSI
+ * midlayer objects
+ */
+ scsi_target_block(
+ &ctd_target->starget->dev);
+ scsi_target_unblock(
+ &ctd_target->starget->dev,
+ SDEV_TRANSPORT_OFFLINE);
+ /*
+ * Target object might still be active
+ * in case its not reaped completely (as
+ * with LVM) thus might be reused when
+ * the link reconnects back (OPT 443532)
+ */
+ scsi_remove_target(
+ &ctd_target->starget->dev);
+ } else {
+ ctd_dprintk_crit(
+ "starget already null\n");
+ }
+
+ /* declare the link dead and buried */
+ ctd_target->detect_completed =
+ EMCCTD_TARGET_DETECT_NOT_COMPLETED;
+ memset(&ctd_target->ctd_detect, 0x0,
+ sizeof(struct emc_ctd_v010_detect));
+
+ wake_up(&lun_discovery_event_barrier);
+ }
+ /* Connect case ... need to scan and create the needed
+ * objects in the SCSI midlayer
+ */
+ else {
+ ctd_dprintk_crit("ctd_target -> %p\n",
+ ctd_target);
+ scsi_scan_target(&ctd_host->shost->shost_gendev,
+ 0, i, SCAN_WILD_CARD, 1);
+ lun_discovery_complete = 1;
+ wake_up(&lun_discovery_event_barrier);
+ }
+ break;
+ default:
+ ctd_dprintk_crit("ctd_target -> %p detect unknown -> %x\n",
+ ctd_target, ctd_target->detect_completed);
+ }
+ }
+
+ return error;
+}
+
+/* This function posts the detect event into adapter specific list */
+static int
+ctd_post_event(union emc_ctd_v010_message *io_msg,
+ struct ctd_pci_private *ctd_private)
+{
+ int error;
+ struct ctd_event_io_element *event;
+
+ error = -ENOMEM;
+
+ event = (struct ctd_event_io_element *)
+ kzalloc(sizeof(struct ctd_event_io_element),
+ GFP_ATOMIC);
+ if (event) {
+ error = 0;
+ event->io_msg = *io_msg;
+ spin_lock(&ctd_private->event_io_lock);
+ list_add_tail(&event->list, &ctd_private->event_io_list);
+ spin_unlock(&ctd_private->event_io_lock);
+ } else {
+ ctd_dprintk_crit("kzalloc failure\n");
+ }
+ return error;
+}
+
+/* Thread Handler Function. This consumes the events posted into its queue, and
+ * takes respective action
+ */
+static int
+ctd_event_handler(void *ctd_thread_args)
+{
+ struct ctd_event_io_element *event;
+ struct ctd_pci_private *ctd_private;
+
+ ctd_private = (struct ctd_pci_private *)ctd_thread_args;
+
+ while (!kthread_should_stop()) {
+ schedule_timeout_interruptible(HZ);
+
+ event = NULL;
+
+ spin_lock(&ctd_private->event_io_lock);
+ if (!list_empty(&ctd_private->event_io_list)) {
+ event = list_first_entry(&ctd_private->event_io_list,
+ struct ctd_event_io_element, list);
+ list_del(&event->list);
+ }
+ spin_unlock(&ctd_private->event_io_lock);
+
+
+ if (event) {
+ int error;
+ emc_ctd_uint8_t what;
+ union emc_ctd_v010_message *io_msg;
+ struct emc_ctd_v010_detect *io_detect;
+
+ io_msg = &event->io_msg;
+ what = io_msg->emc_ctd_scsi_message_header_what;
+
+ if (what != EMC_CTD_V010_WHAT_DETECT) {
+ ctd_dprintk_crit("event -> %p what -> %x\n",
+ event, what);
+ } else {
+ error = -ENODEV;
+ io_detect =
+ &io_msg->emc_ctd_v010_message_detect;
+
+ if (io_detect->emc_ctd_v010_detect_flags ==
+ 0x0) {
+ error = ctd_handle_disconnect(
+ io_detect, ctd_private);
+ } else {
+ if (io_detect->emc_ctd_v010_detect_flags &
+ EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET) {
+ ctd_dprintk(
+ "header addr -> %x key -> %llx\n",
+ io_detect->emc_ctd_detect_header_address,
+ io_detect->emc_ctd_v010_detect_key);
+ error = ctd_handle_target_addition(io_detect,
+ ctd_private);
+ }
+ if (io_detect->emc_ctd_v010_detect_flags &
+ EMC_CTD_V010_DETECT_FLAG_SCSI_INITIATOR) {
+ ctd_dprintk("\n");
+ error = ctd_handle_source_addition(io_detect,
+ ctd_private);
+ }
+ }
+ if (!error) {
+ int retry = EMCCTD_DETECT_RETRY_MAX;
+
+ error = ctd_ITnexus_handler(ctd_private);
+ /* In case of ITnexus_handler failure,
+ * pause for 2 seconds before retrying
+ * the operation again
+ */
+ while (error && retry) {
+ schedule_timeout_interruptible(HZ * 2);
+ error = ctd_ITnexus_handler(ctd_private);
+ retry--;
+ } while (error && retry);
+
+ }
+ }
+ kfree(event);
+ }
+ }
+ return 0;
+}
+
+static int
+ctd_init_event_thread(struct ctd_pci_private *ctd_private)
+{
+ int error;
+
+ INIT_LIST_HEAD(&ctd_private->event_io_list);
+ spin_lock_init(&ctd_private->event_io_lock);
+
+ /* Create the daemon thread to handle detect requests */
+ ctd_private->ctd_event_thread = kthread_create(ctd_event_handler,
+ (void *)ctd_private, "emcctd_event_thread");
+ error = (!IS_ERR(ctd_private->ctd_event_thread)) ? 0 : -EBUSY;
+ if (!error) {
+ wake_up_process(ctd_private->ctd_event_thread);
+ } else {
+ ctd_dprintk_crit(
+ "FAILURE, ctd_private -> %p\n", ctd_private);
+ }
+ return error;
+}
+
+static void
+ctd_destroy_event_thread(struct ctd_pci_private *ctd_private)
+{
+ if (ctd_private->ctd_event_thread)
+ kthread_stop(ctd_private->ctd_event_thread);
+}
+
+static void
+ctd_init_scsi_host_private(struct Scsi_Host *shost, struct pci_dev *pci_dev)
+{
+ int i;
+ struct ctd_host_info *ctd_host_info;
+ struct ctd_pci_private *ctd_private;
+
+ ctd_private = pci_get_drvdata(pci_dev);
+
+ ctd_dprintk("ctd_private -> %p\n", ctd_private);
+
+ ctd_host_info = shost_priv(shost);
+ memset(ctd_host_info, 0x0, sizeof(struct ctd_host_info));
+ for (i = 0; i < EMCCTD_MAX_ID; i++) {
+ struct ctd_target_info *ctd_target;
+
+ ctd_target = &ctd_host_info->target[i];
+ /* nothing to do;
+ * ctd_target->ctd_detect.emc_ctd_v010_detect_flags already zero
+ */
+ }
+
+
+ ctd_host_info->shost = shost;
+ ctd_host_info->pci_dev = pci_dev;
+
+ shost->can_queue = ctd_private->pci_request_queue_size;
+ shost->cmd_per_lun = min(emcctd_cmd_per_lun, shost->can_queue);
+ shost->max_lun = emcctd_max_luns;
+ shost->max_id = EMCCTD_MAX_ID;
+
+ ctd_private->host_private = ctd_host_info;
+
+ ctd_dprintk("scsi_ctd_host = %p\n", ctd_host_info);
+}
+
+
+static int
+ctd_scsi_layer_init(struct pci_dev *pci_dev)
+{
+ int error;
+ struct Scsi_Host *scsi_ctd_host;
+ struct ctd_pci_private *ctd_private = NULL;
+
+ ctd_dprintk("pci_dev -> %p\n", pci_dev);
+
+ scsi_ctd_host = scsi_host_alloc(&scsi_ctd_template,
+ sizeof(struct ctd_host_info));
+ if (scsi_ctd_host == NULL) {
+ error = -ENOMEM;
+ goto ctd_scsi_layer_init_complete;
+ }
+
+ ctd_init_scsi_host_private(scsi_ctd_host, pci_dev);
+
+ ctd_private = pci_get_drvdata(pci_dev);
+
+ error = ctd_init_event_thread(ctd_private);
+ if (error)
+ goto ctd_scsi_layer_init_complete;
+
+ /*
+ * register the HBA to the Linux SCSI stack
+ */
+ error = scsi_add_host(scsi_ctd_host, &pci_dev->dev);
+
+ctd_scsi_layer_init_complete:
+ if (error) {
+ ctd_dprintk_crit("failure, error = %x\n", error);
+ if (scsi_ctd_host != NULL)
+ scsi_host_put(scsi_ctd_host);
+
+ if (ctd_private)
+ ctd_destroy_event_thread(ctd_private);
+ }
+ return error;
+}
+
+static void
+ctd_clear_io_queue(struct ctd_pci_private *ctd_private)
+{
+ struct list_head iochain;
+ unsigned long flags;
+ struct ctd_request_private *request, *request_next;
+
+ ctd_dprintk_crit("ctd_private -> %p\n", ctd_private);
+
+ INIT_LIST_HEAD(&iochain);
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ /* post reset need to cleanup the aborted io
+ * as no reply is expected on them
+ */
+ /* request is still kept as REPLY_AWAITED to
+ * handle any response post connect
+ */
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->aborted_io_list, list) {
+ list_del(&request->list);
+ request->io_state = CTD_IO_REQUEST_REPLY_AWAITED;
+ }
+
+ /* rifle thru queued and requeued IO list and mark them for abort,
+ * the completion to the upper layers is handled by the timeout logic
+ * invoked from the SCSI midlayer
+ */
+ /* request is still kept as REPLY_AWAITED to handle any response post
+ * connect
+ */
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->queued_io_list, list) {
+ list_del(&request->list);
+ list_add(&request->list, &iochain);
+ request->cmnd->host_scribble = NULL;
+ request->io_state = CTD_IO_REQUEST_REPLY_AWAITED;
+
+ /* These requests shall be aborted to upper layer, so treat them
+ * as abort_sent
+ */
+ atomic_long_inc(&ctd_private->hw_stats.abort_sent);
+ atomic_long_dec(&ctd_private->hw_stats.active_io_count);
+ }
+
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->requeued_io_list, list) {
+ list_del(&request->list);
+ list_add(&request->list, &iochain);
+ request->cmnd->host_scribble = NULL;
+ request->io_state = CTD_IO_REQUEST_REPLY_AWAITED;
+
+ /* These requests shall be aborted to upper layer, so treat them
+ * as abort_sent
+ */
+ atomic_long_inc(&ctd_private->hw_stats.abort_sent);
+ atomic_long_dec(&ctd_private->hw_stats.active_io_count);
+ }
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+
+ list_for_each_entry_safe(request, request_next,
+ &iochain, list) {
+ struct scsi_cmnd *cmnd;
+
+ list_del(&request->list);
+
+ ctd_dprintk_crit(
+ "cmnd -> %p request -> %p CTD_IO_REQUEST_REPLY_AWAITED\n",
+ request->cmnd, request);
+
+ cmnd = request->cmnd;
+ request->cmnd = NULL;
+
+ /* error propagation to the SCSI midlayer */
+ scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED);
+ scsi_set_resid(cmnd, scsi_bufflen(cmnd));
+ cmnd->scsi_done(cmnd);
+ }
+ ctd_dprintk("ctd_private -> %p\n", ctd_private);
+}
+
+static int
+ctd_scsi_layer_cleanup(struct pci_dev *pci_dev)
+{
+ int error;
+ struct ctd_pci_private *ctd_private;
+ struct ctd_host_info *ctd_host_info;
+
+ error = 0;
+
+ ctd_private = pci_get_drvdata(pci_dev);
+
+ ctd_dprintk("ctd_private pci_dev -> %p %p\n", ctd_private, pci_dev);
+
+ ctd_check_response_queue((unsigned long)pci_dev);
+
+ ctd_clear_io_queue(ctd_private);
+
+ ctd_destroy_event_thread(ctd_private);
+
+ flush_scheduled_work();
+
+ ctd_host_info = ctd_private->host_private;
+
+ scsi_remove_host(ctd_host_info->shost);
+
+ scsi_host_put(ctd_host_info->shost);
+
+ return error;
+}
+
+#ifdef CONFIG_PM
+static int
+ctd_pci_suspend(struct pci_dev *pci_dev, pm_message_t state)
+{
+ pci_save_state(pci_dev);
+ pci_set_power_state(pci_dev, PCI_D3hot);
+ return 0;
+}
+
+static int
+ctd_pci_resume(struct pci_dev *pci_dev)
+{
+ pci_restore_state(pci_dev);
+ pci_set_power_state(pci_dev, PCI_D0);
+ return 0;
+}
+#endif
+
+static void
+ctd_pci_remove(struct pci_dev *pci_dev)
+{
+ struct ctd_pci_private *ctd_private;
+
+ ctd_dprintk("pic_dev -> %p\n", pci_dev);
+
+ ctd_private = pci_get_drvdata(pci_dev);
+
+ ctd_private->hw_state = CTD_HW_STATE_DISABLED;
+
+ ctd_scsi_layer_cleanup(pci_dev);
+
+#if !defined(__VMKLNX__)
+ ctd_proc_remove(pci_dev);
+#endif
+ free_irq(pci_dev->irq, pci_dev);
+
+ pci_disable_msi(pci_dev);
+
+ if (ctd_private->ioaddr_txrx_rings)
+ pci_iounmap(pci_dev, ctd_private->ioaddr_txrx_rings);
+
+ if (ctd_private->ioaddr_fast_registers)
+ pci_iounmap(pci_dev, ctd_private->ioaddr_fast_registers);
+
+ if (ctd_private->ioaddr_slow_registers)
+ pci_iounmap(pci_dev, ctd_private->ioaddr_slow_registers);
+
+ tasklet_kill(&ctd_private->isr_tasklet);
+
+ ctd_release_io_pool(ctd_private);
+
+ kfree(ctd_private);
+
+ pci_release_regions(pci_dev);
+ pci_set_drvdata(pci_dev, NULL);
+ pci_disable_device(pci_dev);
+}
+
+static void
+ctd_check_error_condition(struct pci_dev *pci_dev)
+{
+#define EMCCTD_MAX_CACHED_ERROR 14
+ int i, j;
+ int error;
+ union emc_ctd_v010_message message;
+ static emc_ctd_uint32_t internal_errors_1_14[EMCCTD_MAX_CACHED_ERROR];
+ struct ctd_pci_private *ctd_private = pci_get_drvdata(pci_dev);
+
+ if (ctd_private->pci_fast_registers->emc_ctd_v010_fregs_error_flag) {
+ for (i = 0; i < EMCCTD_MAX_CACHED_ERROR; i++) {
+ if (internal_errors_1_14[i] !=
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i]) {
+
+ internal_errors_1_14[i] =
+ ctd_private->pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i];
+
+ error = i + 1;
+
+ for (j = 0; j < EMC_CTD_V010_LOG_ERROR_TX_SIZE; j++) {
+ if (ctd_private->pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_error[j] ==
+ error) {
+ memcpy(&message,
+ &ctd_private->pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_message[j],
+ sizeof(message));
+ ctd_dprintk_crit(
+ "header addr -> %x error -> %s\n",
+ message.emc_ctd_scsi_message_header_address,
+ (error ==
+ EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED ?
+ "EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED" :
+ error ==
+ EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT ?
+ "EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT" :
+ error ==
+ EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED ?
+ "EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED" :
+ error ==
+ EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER ?
+ "EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER" :
+ error ==
+ EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE ?
+ "EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE" :
+ error ==
+ EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN ?
+ "EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN" :
+ "EMC_CTD_V010_ERROR_NULL"));
+ }
+ }
+ }
+ }
+ }
+}
+/*
+ * ctd_check_response_queue
+ *
+ * Bottom half of interrupt handler.
+ */
+static void
+ctd_check_response_queue(unsigned long instance_addr)
+{
+ struct pci_dev *pci_dev = (struct pci_dev *)instance_addr;
+ struct ctd_pci_private *ctd_private = pci_get_drvdata(pci_dev);
+ union emc_ctd_v010_message io_response;
+
+
+ /* empty response queue */
+ while (ctd_hw_dequeue_response(&io_response, ctd_private) ==
+ SUCCESS) {
+ /* handle the response */
+ ctd_handle_response(&io_response, ctd_private);
+ }
+}
+
+
+static irqreturn_t
+ctd_isr(int irq, void *opaque)
+{
+ struct pci_dev *pci_dev = (struct pci_dev *)opaque;
+ struct ctd_pci_private *ctd_private = pci_get_drvdata(pci_dev);
+
+ atomic_long_inc(&ctd_private->hw_stats.interrupts);
+
+ /* schedule work for later */
+ tasklet_schedule(&ctd_private->isr_tasklet);
+
+ return IRQ_HANDLED;
+}
+
+static int
+ctd_request_msi(struct pci_dev *pci_dev)
+{
+ int err = -EFAULT;
+
+ if (pci_dev->irq) {
+ err = pci_enable_msi(pci_dev);
+ if (!err) {
+ err = request_irq(pci_dev->irq, ctd_isr,
+ IRQF_SHARED,
+ pci_name(pci_dev), pci_dev);
+ if (err < 0) {
+ ctd_dprintk_crit("request_irq failure !!!\n");
+ pci_disable_msi(pci_dev);
+ err = -EBUSY;
+ }
+ }
+ }
+ return err;
+}
+
+static struct ctd_request_private *
+ctd_acquire_request(struct ctd_pci_private *ctd_private)
+{
+ unsigned long flags;
+ struct ctd_request_private *ctd_request;
+
+ ctd_request = NULL;
+
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+
+ /* check if any request in the aborted io list can be reused */
+ if (!list_empty(&ctd_private->aborted_io_list)) {
+ struct ctd_request_private *request, *request_next;
+
+ list_for_each_entry_safe(request, request_next,
+ &ctd_private->aborted_io_list, list) {
+ /* aborted_io_list is in chronologically order thus
+ * failure of time_after() indicates any request after
+ * this point is not in the kill zone
+ */
+ if (time_before(jiffies, request->purge_lifetime))
+ break;
+
+ list_del(&request->list);
+ if (request->cdb_page) {
+ __free_pages(request->cdb_page,
+ request->cdb_page_order);
+ }
+ if (request->sgllist_page) {
+ __free_pages(request->sgllist_page,
+ request->sgllist_page_order);
+ }
+ memset(request, 0x0,
+ sizeof(struct ctd_request_private));
+ list_add(&request->list, &ctd_private->io_pool);
+ atomic_long_inc(&ctd_private->hw_stats.free_io_entries);
+ }
+ }
+
+ if (!list_empty(&ctd_private->io_pool)) {
+ ctd_request = list_first_entry(&ctd_private->io_pool,
+ struct ctd_request_private, list);
+ list_del(&ctd_request->list);
+ ctd_request->io_state = CTD_IO_REQUEST_QUEUED;
+ }
+
+ if (ctd_request)
+ atomic_long_dec(&ctd_private->hw_stats.free_io_entries);
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+
+ return ctd_request;
+}
+
+static void
+ctd_release_request(struct ctd_request_private *ctd_request,
+ struct ctd_pci_private *ctd_private)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags);
+ memset(ctd_request, 0x0, sizeof(struct ctd_request_private));
+ list_add(&ctd_request->list, &ctd_private->io_pool);
+
+ atomic_long_inc(&ctd_private->hw_stats.free_io_entries);
+
+ spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags);
+}
+
+static void
+ctd_release_io_pool(struct ctd_pci_private *ctd_private)
+{
+ kfree(ctd_private->io_map);
+}
+
+static int
+ctd_alloc_io_pool(struct ctd_pci_private *ctd_private, unsigned int pool_size)
+{
+ int i;
+ int error;
+
+ error = -ENOMEM;
+
+ INIT_LIST_HEAD(&ctd_private->io_pool);
+ INIT_LIST_HEAD(&ctd_private->queued_io_list);
+ INIT_LIST_HEAD(&ctd_private->aborted_io_list);
+ INIT_LIST_HEAD(&ctd_private->requeued_io_list);
+
+ spin_lock_init(&ctd_private->io_mgmt_lock);
+
+ ctd_private->io_map = kcalloc(pool_size,
+ sizeof(struct ctd_request_private), GFP_KERNEL);
+
+ /*
+ * in case of allocation failure try with one fourth the size before
+ * throwing the towel
+ */
+
+ if (ctd_private->io_map == NULL) {
+ pool_size = pool_size >> 2;
+ ctd_private->io_map = kcalloc(pool_size,
+ sizeof(struct ctd_request_private), GFP_KERNEL);
+ if (ctd_private->io_map == NULL) {
+ ctd_dprintk_crit(
+ "io_pool allocation failure for pool_size -> %d\n",
+ pool_size);
+ goto ctd_alloc_io_pool_complete;
+ }
+ }
+
+ ctd_private->io_map_end = ctd_private->io_map + pool_size;
+
+ for (i = 0; i < pool_size; i++) {
+ struct ctd_request_private *request_context =
+ ctd_private->io_map + i;
+ memset(request_context, 0x0,
+ sizeof(struct ctd_request_private));
+ list_add(&request_context->list, &ctd_private->io_pool);
+ }
+ ctd_dprintk_crit(
+ "ctd_private -> %p, pool_size -> %x, io_map -> %p, io_map_end-> %p\n",
+ ctd_private, pool_size,
+ ctd_private->io_map, ctd_private->io_map_end);
+ error = 0;
+ ctd_private->hw_stats.free_io_entries.counter = pool_size;
+ctd_alloc_io_pool_complete:
+ return error;
+}
+
+static int
+ctd_pci_probe(
+ struct pci_dev *pci_dev,
+ const struct pci_device_id *id
+)
+{
+ struct ctd_pci_private *ctd_private;
+ int ctd_proc_initialized;
+ int ctd_scsi_initialized;
+ int ctd_regions_initialized;
+ int err;
+
+ err = -ENODEV;
+ ctd_proc_initialized = FAILED;
+ ctd_scsi_initialized = FAILED;
+ ctd_regions_initialized = FAILED;
+
+ ctd_private = (struct ctd_pci_private *)
+ kzalloc(sizeof(struct ctd_pci_private), GFP_ATOMIC);
+
+ if (ctd_private == NULL) {
+ ctd_dprintk_crit("kzalloc Failure\n");
+ goto ctd_pci_probe_complete;
+ }
+
+ ctd_private->pci_dev = pci_dev;
+
+ /* enable the device */
+ err = pci_enable_device(pci_dev);
+ if (err) {
+ ctd_dprintk_crit("pci_enable_device Failure\n");
+ goto ctd_pci_probe_complete;
+ }
+ pci_set_master(pci_dev);
+
+ err = pci_request_regions(pci_dev, "ctd-pci");
+ if (err) {
+ ctd_dprintk_crit("pci_request_regions Failure\n");
+ goto ctd_pci_probe_complete;
+ }
+
+ ctd_regions_initialized = SUCCESS;
+
+ ctd_dprintk("ctd_private pci_dev -> %p %p\n", ctd_private, pci_dev);
+
+#define EMC_CTD_TXRX_MSG_SIZE 128
+
+ if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_RINGS)) {
+ ctd_private->ioaddr_txrx_rings = ioremap(
+ pci_resource_start(pci_dev,
+ EMC_CTD_V010_BAR_RINGS),
+ pci_resource_len(pci_dev,
+ EMC_CTD_V010_BAR_RINGS)
+ );
+ ctd_private->txrx_ringsize =
+ (pci_resource_len(pci_dev,
+ EMC_CTD_V010_BAR_RINGS) >> 1)
+ / EMC_CTD_TXRX_MSG_SIZE;
+
+ ctd_dprintk_crit(
+ "physical addr = %llx ioaddr_txrx_rings = %p , ring size = %x\n",
+ pci_resource_start(pci_dev,
+ EMC_CTD_V010_BAR_RINGS),
+ ctd_private->ioaddr_txrx_rings,
+ ctd_private->txrx_ringsize);
+
+ }
+ if (ctd_private->ioaddr_txrx_rings == NULL) {
+ err = -ENOMEM;
+ ctd_dprintk_crit("ioremap failure\n");
+ goto ctd_pci_probe_complete;
+ } else {
+ ctd_private->pci_request_array = ctd_private->ioaddr_txrx_rings;
+ ctd_private->pci_response_array =
+ ctd_private->ioaddr_txrx_rings +
+ ((pci_resource_len(pci_dev,
+ EMC_CTD_V010_BAR_RINGS)) >> 1);
+ }
+
+
+ if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_FREGS)) {
+ ctd_private->ioaddr_fast_registers =
+ ioremap(pci_resource_start(pci_dev,
+ EMC_CTD_V010_BAR_FREGS),
+ pci_resource_len(pci_dev,
+ EMC_CTD_V010_BAR_FREGS));
+
+ ctd_dprintk_crit(
+ "physical addr = %llx ioaddr_fast_registers = %p\n",
+ pci_resource_start(pci_dev,
+ EMC_CTD_V010_BAR_FREGS),
+ ctd_private->ioaddr_fast_registers);
+ }
+ if (ctd_private->ioaddr_fast_registers == NULL) {
+ err = -ENOMEM;
+ ctd_dprintk_crit("ioremap failure\n");
+ goto ctd_pci_probe_complete;
+ } else {
+ ctd_private->pci_fast_registers =
+ ctd_private->ioaddr_fast_registers;
+ }
+
+ if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_SREGS)) {
+ ctd_private->ioaddr_slow_registers =
+ ioremap(pci_resource_start(pci_dev,
+ EMC_CTD_V010_BAR_SREGS),
+ pci_resource_len(pci_dev,
+ EMC_CTD_V010_BAR_SREGS));
+
+ ctd_dprintk_crit(
+ "physical addr = %llx ioaddr_slow_registers = %p\n",
+ pci_resource_start(pci_dev,
+ EMC_CTD_V010_BAR_SREGS),
+ ctd_private->ioaddr_slow_registers);
+ }
+ if (ctd_private->ioaddr_slow_registers == NULL) {
+ err = -ENOMEM;
+ ctd_dprintk_crit("ioremap failure\n");
+ goto ctd_pci_probe_complete;
+ } else {
+ ctd_private->pci_slow_registers =
+ ctd_private->ioaddr_slow_registers;
+ }
+
+ /* reset the device */
+ ctd_private->pci_device_reset_register = 0XFF;
+
+ ctd_private->pci_request_queue_size = ctd_private->txrx_ringsize;
+ ctd_private->pci_response_queue_size = ctd_private->txrx_ringsize;
+
+ err = ctd_alloc_io_pool(ctd_private,
+ ctd_private->pci_request_queue_size);
+ if (err) {
+ ctd_dprintk_crit("ctd_alloc_io_pool failure\n");
+ goto ctd_pci_probe_complete;
+ }
+
+ pci_set_drvdata(pci_dev, ctd_private);
+
+ spin_lock_init(&ctd_private->isr_lock);
+
+ /* setup tasklet for scanning response queue */
+ tasklet_init(&ctd_private->isr_tasklet,
+ ctd_check_response_queue, (unsigned long)pci_dev);
+
+ ctd_private->hw_state = CTD_HW_STATE_INITIALIZED;
+
+ pci_set_master(pci_dev);
+
+#if !defined(__VMKLNX__)
+ err = ctd_proc_init(pci_dev);
+ if (err) {
+ ctd_dprintk_crit("ctd_proc_init failure\n");
+ goto ctd_pci_probe_complete;
+ }
+ ctd_proc_initialized = SUCCESS;
+#endif
+ err = ctd_scsi_layer_init(pci_dev);
+ if (err) {
+ ctd_dprintk_crit("ctd_scsi_layer_init failure\n");
+ goto ctd_pci_probe_complete;
+ }
+ ctd_scsi_initialized = SUCCESS;
+
+ err = ctd_request_msi(pci_dev);
+ if (err) {
+ ctd_dprintk_crit("ctd_request_msi failure\n");
+ goto ctd_pci_probe_complete;
+ }
+
+ /* after we reset the device, but before we enabled MSI, some messages
+ * may have been received. check for them
+ */
+ tasklet_schedule(&ctd_private->isr_tasklet);
+
+ctd_pci_probe_complete:
+ if (err) {
+ if (ctd_private) {
+ tasklet_kill(&ctd_private->isr_tasklet);
+
+ if (ctd_scsi_initialized == SUCCESS)
+ ctd_scsi_layer_cleanup(pci_dev);
+
+ if (ctd_proc_initialized == SUCCESS)
+ ctd_proc_remove(pci_dev);
+
+ if (ctd_private->ioaddr_txrx_rings)
+ pci_iounmap(pci_dev,
+ ctd_private->ioaddr_txrx_rings);
+
+ if (ctd_private->ioaddr_fast_registers)
+ pci_iounmap(pci_dev,
+ ctd_private->ioaddr_fast_registers);
+
+ if (ctd_private->ioaddr_slow_registers)
+ pci_iounmap(pci_dev,
+ ctd_private->ioaddr_slow_registers);
+
+ if (ctd_regions_initialized == SUCCESS)
+ pci_release_regions(pci_dev);
+
+ ctd_release_io_pool(ctd_private);
+ kfree(ctd_private);
+ }
+ pci_set_drvdata(pci_dev, NULL);
+ pci_disable_device(pci_dev);
+ }
+ return err;
+}
+
+#if !defined(__VMKLNX__)
+static const struct file_operations ctd_proc_fops = {
+ .open = ctd_proc_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = seq_release,
+};
+static int
+ctd_proc_open(struct inode *inode, struct file *file)
+{
+ return single_open(file, ctd_proc_show, PDE_DATA(inode));
+}
+
+int
+ctd_proc_init(struct pci_dev *pci_dev)
+{
+ int err;
+ static int hw_index;
+ char hw_name[MAX_PROC_FILE_NAMELEN];
+ struct ctd_pci_private *ctd_private;
+ struct proc_dir_entry *pde;
+
+ ctd_private = (struct ctd_pci_private *) pci_get_drvdata(pci_dev);
+ ctd_private->hw_index = hw_index;
+
+ memset(hw_name, 0x0, sizeof(hw_name));
+ snprintf(hw_name, sizeof(hw_name), "emcctd_stats_%d", hw_index++);
+
+ err = -EPERM;
+
+ do {
+ if (ctd_proc_directory == NULL)
+ break;
+ pde = proc_create_data(hw_name, S_IFREG | S_IRUGO | S_IWUSR,
+ ctd_proc_directory,
+ &ctd_proc_fops,
+ ctd_private);
+
+ if (pde == NULL) {
+ ctd_dprintk_crit(
+ "create_proc_read_entry failure for %s\n",
+ hw_name);
+ break;
+ }
+ err = 0;
+ } while (0);
+ return err;
+}
+
+void
+ctd_proc_remove(struct pci_dev *pci_dev)
+{
+ int hw_index;
+ char hw_name[MAX_PROC_FILE_NAMELEN];
+ struct ctd_pci_private *ctd_private;
+
+ ctd_private = (struct ctd_pci_private *) pci_get_drvdata(pci_dev);
+
+ hw_index = ctd_private->hw_index;
+ memset(hw_name, 0x0, sizeof(hw_name));
+ snprintf(hw_name, sizeof(hw_name), "emc/emcctd_stats_%d", hw_index);
+
+ ctd_dprintk("removing %s\n", hw_name);
+
+ remove_proc_entry(hw_name, NULL);
+
+}
+#endif
+
+static int __init ctd_pci_init(void)
+{
+ int err;
+
+ ctd_dprintk_crit("Loading emcctd\n");
+ init_waitqueue_head(&lun_discovery_event_barrier);
+
+ ctd_proc_directory = proc_mkdir("emc", NULL);
+
+ err = pci_register_driver(&ctd_pci_driver);
+
+ if (err) {
+ remove_proc_entry("emc", NULL);
+ } else {
+ /* wait for 20 seconds or less to allow the luns to appear
+ * before exiting from insmod
+ */
+ wait_event_interruptible_timeout(lun_discovery_event_barrier,
+ lun_discovery_complete, ((HZ) * 20));
+ }
+
+ return err;
+}
+
+module_init(ctd_pci_init);
+
+static void __exit ctd_pci_exit(void)
+{
+ pci_unregister_driver(&ctd_pci_driver);
+ remove_proc_entry("emc", NULL);
+}
+
+module_exit(ctd_pci_exit);
diff --git a/drivers/scsi/emcctd/emcctd.h b/drivers/scsi/emcctd/emcctd.h
new file mode 100644
index 0000000..3fba931
--- /dev/null
+++ b/drivers/scsi/emcctd/emcctd.h
@@ -0,0 +1,232 @@
+/*
+ * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem.
+ *
+ * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA.
+ *
+ * Authors:
+ * fredette, matt <matt.fredette@....com>
+ * Pirotte, Serge <serge.pirotte@....com>
+ * Singh Animesh <Animesh.Singh@....com>
+ * Singhal, Maneesh <Maneesh.Singhal@....com>
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ */
+
+#ifndef _EMCCTD_H_
+#define _EMCCTD_H_
+
+#define DRV_NAME "emcctd"
+
+#define EMCCTD_V010_PROTOCOL_MINOR_VERSION 0x0
+
+/* please refer to emc_ctd_v010_scsi_command_cdb in emc_ctd_interface.h */
+#define EMCCTD_V010_MAX_CDB_SIZE 16
+
+#define EMCCTD_MAX_LUN 16384
+#define EMCCTD_MAX_ID 16
+#define EMCCTD_MAX_RETRY 5
+#define EMCCTD_CMD_PER_LUN 16
+#define EMCCTD_THIS_ID (-1)
+#define EMCCTD_REQUEST_TIMEOUT (60 * HZ)
+#define EMCCTD_OPAQUE_PURGE_WAITTIME (10 * HZ)
+
+#define EMCCTD_DEVICE_RESET_PAUSE 3
+#define EMCCTD_DETECT_RETRY_MAX 3
+
+#define ctd_dprintk(__m_fmt, ...) \
+do { \
+ if (ctd_debug) \
+ pr_info("%s:%d:"__m_fmt, __func__, __LINE__, ##__VA_ARGS__); \
+} while (0)
+
+#define ctd_dprintk_crit(__m_fmt, ...) \
+ pr_crit("%s:%d:"__m_fmt, __func__, __LINE__, ##__VA_ARGS__)
+
+#define EMCCTD_TARGET_DETECT_COMPLETED 1
+#define EMCCTD_TARGET_DETECT_NOT_COMPLETED 0
+
+struct ctd_target_info {
+ unsigned int detect_completed;
+ struct scsi_target *starget;
+ struct emc_ctd_v010_detect ctd_detect;
+};
+
+#define emc_ctd_detect_header_address \
+ emc_ctd_v010_detect_header.emc_ctd_v010_header_address
+#define emc_ctd_detect_header_minor \
+ emc_ctd_v010_detect_header.emc_ctd_v010_header_minor
+#define emc_ctd_detect_header_what \
+ emc_ctd_v010_detect_header.emc_ctd_v010_header_what
+
+#define emc_ctd_scsi_command_header_address \
+ emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_address
+#define emc_ctd_scsi_command_header_minor \
+ emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_minor
+#define emc_ctd_scsi_command_header_what \
+ emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_what
+
+#define emc_ctd_scsi_response_header_address \
+ emc_ctd_v010_scsi_response_header.emc_ctd_v010_header_address
+#define emc_ctd_scsi_response_header_minor \
+ emc_ctd_v010_scsi_response_header.emc_ctd_v010_header_minor
+#define emc_ctd_scsi_response_header_what \
+ emc_ctd_v010_scsi_response_header.emc_ctd_v010_header_what
+
+#define emc_ctd_scsi_phase_header_address \
+ emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_address
+#define emc_ctd_scsi_phase_header_minor \
+ emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_minor
+#define emc_ctd_scsi_phase_header_what \
+ emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_what
+
+#define emc_ctd_scsi_message_header_address \
+ emc_ctd_v010_message_header.emc_ctd_v010_header_address
+#define emc_ctd_scsi_message_header_minor \
+ emc_ctd_v010_message_header.emc_ctd_v010_header_minor
+#define emc_ctd_scsi_message_header_what \
+ emc_ctd_v010_message_header.emc_ctd_v010_header_what
+
+#define emc_ctd_scsi_command_sgl \
+ emc_ctd_v010_scsi_command_u.emc_ctd_v010_scsi_command_sgl
+#define emc_ctd_scsi_response_extra \
+ emc_ctd_v010_scsi_response_u.emc_ctd_v010_scsi_response_extra
+
+#define ctd_detect_name_bytes \
+ ctd_detect.emc_ctd_v010_detect_name.emc_ctd_v010_name_bytes
+
+struct ctd_host_info {
+ struct Scsi_Host *shost;
+ struct pci_dev *pci_dev;
+ struct ctd_target_info target[EMCCTD_MAX_ID];
+};
+
+struct ctd_dev_info {
+ struct ctd_host_info *ctd_host;
+ struct ctd_target_info *ctd_target;
+ struct emc_ctd_v010_detect *ctd_target_detect;
+};
+
+#define PROC_STAT_SCSI_TS_MAX 10
+#define MAX_PROC_FILE_NAMELEN 128
+#define CTD_MAX_IO_STATS 200
+
+struct ctd_hw_stats {
+ atomic_long_t interrupts;
+ atomic_long_t requests_sent;
+ atomic_long_t responses_received;
+ atomic_long_t active_io_count;
+ atomic_long_t abort_sent;
+ atomic_long_t abort_received;
+ atomic_long_t what_in;
+ atomic_long_t what_out;
+ atomic_long_t free_io_entries;
+ unsigned long long io_stats[CTD_MAX_IO_STATS];
+ unsigned int io_stats_index;
+};
+
+enum ctd_io_request_state {
+ CTD_IO_REQUEST_FREE,
+ CTD_IO_REQUEST_QUEUED,
+ CTD_IO_REQUEST_REQUEUED,
+ CTD_IO_REQUEST_ABORTED,
+ CTD_IO_REQUEST_COMPLETED,
+ CTD_IO_REQUEST_REPLY_AWAITED,
+ CTD_IO_REQUEST_INVALID
+};
+
+enum ctd_hw_state {
+ CTD_HW_STATE_UNINITIALIZED,
+ CTD_HW_STATE_INITIALIZED,
+ CTD_HW_STATE_UNDER_RESET,
+ CTD_HW_STATE_DISABLED,
+ CTD_HW_STATE_INVALID
+};
+
+struct ctd_request_private {
+ struct list_head list;
+ enum ctd_io_request_state io_requeue_state;
+ unsigned int io_timeout;
+ enum ctd_io_request_state io_state;
+ struct scsi_cmnd *cmnd;
+ struct page *cdb_page;
+ unsigned int cdb_page_order;
+ struct page *sgllist_page;
+ unsigned int sgllist_page_order;
+ unsigned long purge_lifetime;
+ unsigned long long io_start_time;
+};
+
+struct ctd_pci_private {
+ struct pci_dev *pci_dev;
+ void *host_private;
+
+ void __iomem *ioaddr_txrx_rings;
+ void __iomem *ioaddr_fast_registers;
+ void __iomem *ioaddr_slow_registers;
+
+ emc_ctd_uint32_t txrx_ringsize;
+ emc_ctd_uint32_t pci_request_queue_size;
+ emc_ctd_uint32_t pci_response_queue_size;
+ struct emc_ctd_v010_fregs *pci_fast_registers;
+ struct emc_ctd_v010_sregs *pci_slow_registers;
+ union emc_ctd_v010_message *pci_response_array;
+ union emc_ctd_v010_message *pci_request_array;
+
+ struct tasklet_struct isr_tasklet;
+ spinlock_t isr_lock;
+ unsigned int hw_index;
+ struct ctd_hw_stats hw_stats;
+ enum ctd_hw_state hw_state;
+ struct list_head queued_io_list;
+ struct list_head aborted_io_list;
+ struct list_head requeued_io_list;
+ struct list_head io_pool;
+ struct ctd_request_private *io_map;
+ struct ctd_request_private *io_map_end;
+ spinlock_t io_mgmt_lock;
+
+ struct task_struct *ctd_event_thread;
+ struct list_head event_io_list;
+ spinlock_t event_io_lock;
+};
+
+#define request_producer_index \
+ pci_fast_registers->emc_ctd_v010_fregs_tx_index_producer
+#define request_consumer_index \
+ pci_fast_registers->emc_ctd_v010_fregs_tx_index_consumer
+#define response_producer_index \
+ pci_fast_registers->emc_ctd_v010_fregs_rx_index_producer
+#define response_consumer_index \
+ pci_fast_registers->emc_ctd_v010_fregs_rx_index_consumer
+#define pci_device_reset_register \
+ pci_slow_registers->emc_ctd_v010_sregs_reset
+
+#define pci_device_name_bytes \
+ pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes
+
+struct ctd_event_io_element {
+ struct list_head list;
+ union emc_ctd_v010_message io_msg;
+};
+
+static inline unsigned long long ctd_read_tsc(void)
+{
+ unsigned long long local_tsc;
+
+ local_tsc = rdtsc();
+
+ return local_tsc;
+}
+
+#endif /* _EMCCTD_H_ */
+
+/* vi: set ts=8 sw=8 noet: */
--
1.8.5.2
Powered by blists - more mailing lists