lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [day] [month] [year] [list]
Message-Id: <1253311760.4498.159.camel@haakon2.linux-iscsi.org>
Date:	Fri, 18 Sep 2009 15:09:20 -0700
From:	"Nicholas A. Bellinger" <nab@...ux-iscsi.org>
To:	LKML <linux-kernel@...r.kernel.org>,
	linux-scsi <linux-scsi@...r.kernel.org>
Cc:	Andrew Morton <akpm@...ux-foundation.org>,
	Greg KH <greg@...ah.com>,
	Douglas Gilbert <dgilbert@...erlog.com>,
	James Bottomley <James.Bottomley@...senPartnership.com>,
	Hannes Reinecke <hare@...e.de>,
	FUJITA Tomonori <fujita.tomonori@....ntt.co.jp>,
	Mike Christie <michaelc@...wisc.edu>,
	Joel Becker <joel.becker@...cle.com>,
	"Martin K. Petersen" <martin.petersen@...cle.com>,
	Christoph Hellwig <hch@....de>,
	Linus Torvalds <torvalds@...ux-foundation.org>,
	Alan Stern <stern@...land.harvard.edu>,
	Boaz Harrosh <bharrosh@...asas.com>,
	Florian Haas <florian.haas@...bit.com>,
	Philipp Reisner <philipp.reisner@...bit.com>,
	Lars Ellenberg <lars.ellenberg@...bit.com>,
	Daniel Walker <dwalker@...o99.com>
Subject: [RFC PATCH 16/19] Target_Core_Mod Generic Transport Core

[RFC PATCH 16/19] Target_Core_Mod Generic Transport Core

The patch adds the core generic target engine functionality including processing of
se_cmd_t, generation of se_task_ts, control patch CDB emulation, sense data, memory allocation,
and memory mapping from fabric module allocated memory using linked list struct scatterlist
style memory based on underlying storage object hardware.

The Generic Target Core layer is above the Linux/SCSI, Linux/BLOCK and Linux/VFS subsystems,
and communicates with these subsystems (and local RAMDISK plugins) using an API of function pointers
in se_subsystem_api_t.  It provides processing capabilties and emulation for SCSI logic and CDBs in
a generic manner to storage object LUNs provided from Linux storage subsystems backend storage devices.

Signed-off-by: Nicholas A. Bellinger <nab@...ux-iscsi.org>

 
diff --git a/drivers/target/target_core_transport.c b/drivers/target/target_core_transport.c
new file mode 100644
index 0000000..3825628
--- /dev/null
+++ b/drivers/target/target_core_transport.c
@@ -0,0 +1,8001 @@
+/*******************************************************************************
+ * Filename:  target_core_transport.c
+ *
+ * This file contains the Generic Target Engine Core.
+ *
+ * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
+ * Copyright (c) 2005, 2006, 2007 SBE, Inc.
+ * Copyright (c) 2007-2009 Rising Tide Software, Inc.
+ * Copyright (c) 2008-2009 Linux-iSCSI.org
+ *
+ * Nicholas A. Bellinger <nab@...nel.org>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+ *
+ ******************************************************************************/
+
+
+#define TARGET_CORE_TRANSPORT_C
+
+#include <linux/version.h>
+#include <linux/net.h>
+#include <linux/delay.h>
+#include <linux/string.h>
+#include <linux/timer.h>
+#include <linux/slab.h>
+#include <linux/blkdev.h>
+#include <linux/spinlock.h>
+#include <linux/smp_lock.h>
+#include <linux/kthread.h>
+#include <linux/in.h>
+#include <linux/cdrom.h>
+#include <net/sock.h>
+#include <net/tcp.h>
+#include <scsi/scsi.h>
+#include <scsi/scsi_cmnd.h>
+#include <scsi/libsas.h> /* For TASK_ATTR_* */
+
+#include <target/target_core_base.h>
+#include <target/target_core_device.h>
+#include <target/target_core_hba.h>
+#include <target/target_core_scdb.h>
+#include <target/target_core_pr.h>
+#include <target/target_core_alua.h>
+#include <target/target_core_tmr.h>
+#include <target/target_core_ua.h>
+#include <target/target_core_transport.h>
+#include <target/target_core_plugin.h>
+#include <target/target_core_seobj.h>
+#include <target/target_core_transport_plugin.h>
+#include <target/target_core_fabric_ops.h>
+#include <target/target_core_configfs.h>
+
+#undef TARGET_CORE_TRANSPORT_C
+
+/* #define DEBUG_CDB_HANDLER */
+#ifdef DEBUG_CDB_HANDLER
+#define DEBUG_CDB_H(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CDB_H(x...)
+#endif
+
+/* #define DEBUG_CMD_MAP */
+#ifdef DEBUG_CMD_MAP
+#define DEBUG_CMD_M(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CMD_M(x...)
+#endif
+
+/* #define DEBUG_MEM_ALLOC */
+#ifdef DEBUG_MEM_ALLOC
+#define DEBUG_MEM(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_MEM(x...)
+#endif
+
+/* #define DEBUG_MEM2_ALLOC */
+#ifdef DEBUG_MEM2_ALLOC
+#define DEBUG_MEM2(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_MEM2(x...)
+#endif
+
+/* #define DEBUG_SG_CALC */
+#ifdef DEBUG_SG_CALC
+#define DEBUG_SC(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_SC(x...)
+#endif
+
+/* #define DEBUG_SE_OBJ */
+#ifdef DEBUG_SE_OBJ
+#define DEBUG_SO(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_SO(x...)
+#endif
+
+/* #define DEBUG_CMD_VOL */
+#ifdef DEBUG_CMD_VOL
+#define DEBUG_VOL(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_VOL(x...)
+#endif
+
+/* #define DEBUG_CMD_STOP */
+#ifdef DEBUG_CMD_STOP
+#define DEBUG_CS(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CS(x...)
+#endif
+
+/* #define DEBUG_PASSTHROUGH */
+#ifdef DEBUG_PASSTHROUGH
+#define DEBUG_PT(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_PT(x...)
+#endif
+
+/* #define DEBUG_TASK_STOP */
+#ifdef DEBUG_TASK_STOP
+#define DEBUG_TS(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TS(x...)
+#endif
+
+/* #define DEBUG_TRANSPORT_STOP */
+#ifdef DEBUG_TRANSPORT_STOP
+#define DEBUG_TRANSPORT_S(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TRANSPORT_S(x...)
+#endif
+
+/* #define DEBUG_TASK_FAILURE */
+#ifdef DEBUG_TASK_FAILURE
+#define DEBUG_TF(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TF(x...)
+#endif
+
+/* #define DEBUG_DEV_OFFLINE */
+#ifdef DEBUG_DEV_OFFLINE
+#define DEBUG_DO(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_DO(x...)
+#endif
+
+/* #define DEBUG_TASK_STATE */
+#ifdef DEBUG_TASK_STATE
+#define DEBUG_TSTATE(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TSTATE(x...)
+#endif
+
+/* #define DEBUG_STATUS_THR */
+#ifdef DEBUG_STATUS_THR
+#define DEBUG_ST(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_ST(x...)
+#endif
+
+/* #define DEBUG_TASK_TIMEOUT */
+#ifdef DEBUG_TASK_TIMEOUT
+#define DEBUG_TT(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TT(x...)
+#endif
+
+/* #define DEBUG_GENERIC_REQUEST_FAILURE */
+#ifdef DEBUG_GENERIC_REQUEST_FAILURE
+#define DEBUG_GRF(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_GRF(x...)
+#endif
+
+/* #define DEBUG_SAM_TASK_ATTRS */
+#ifdef DEBUG_SAM_TASK_ATTRS
+#define DEBUG_STA(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_STA(x...)
+#endif
+
+se_global_t *se_global;
+EXPORT_SYMBOL(se_global);
+
+struct kmem_cache *se_cmd_cache;
+struct kmem_cache *se_task_cache;
+struct kmem_cache *se_tmr_req_cache;
+struct kmem_cache *se_sess_cache;
+struct kmem_cache *se_hba_cache;
+struct kmem_cache *se_ua_cache;
+struct kmem_cache *t10_pr_reg_cache;
+struct kmem_cache *t10_alua_lu_gp_cache;
+struct kmem_cache *t10_alua_lu_gp_mem_cache;
+struct kmem_cache *t10_alua_tg_pt_gp_cache;
+struct kmem_cache *t10_alua_tg_pt_gp_mem_cache;
+
+static int transport_generic_write_pending(se_cmd_t *);
+static int transport_processing_thread(void *);
+
+static char *transport_passthrough_get_fabric_name(void)
+{
+	return "PT";
+}
+
+static u32 transport_passthrough_get_task_tag(se_cmd_t *cmd)
+{
+	return 0;
+}
+
+static int transport_passthrough_get_cmd_state(se_cmd_t *cmd)
+{
+	return 0;
+}
+
+static void transport_passthrough_release_cmd_direct(se_cmd_t *cmd)
+{
+	return;
+}
+
+static u16 transport_passthrough_set_fabric_sense_len(
+	se_cmd_t *cmd,
+	u32 sense_len)
+{
+	return 0;
+}
+
+static u16 transport_passthrough_get_fabric_sense_len(void)
+{
+	return 0;
+}
+
+struct target_core_fabric_ops passthrough_fabric_ops = {
+	.release_cmd_direct	= transport_passthrough_release_cmd_direct,
+	.get_fabric_name	= transport_passthrough_get_fabric_name,
+	.get_task_tag		= transport_passthrough_get_task_tag,
+	.get_cmd_state		= transport_passthrough_get_cmd_state,
+	.set_fabric_sense_len	= transport_passthrough_set_fabric_sense_len,
+	.get_fabric_sense_len	= transport_passthrough_get_fabric_sense_len,
+};
+
+int init_se_global(void)
+{
+	se_global_t *global;
+
+	global = kzalloc(sizeof(se_global_t), GFP_KERNEL);
+	if (!(global)) {
+		printk(KERN_ERR "Unable to allocate memory for se_global_t\n");
+		return -1;
+	}
+
+	INIT_LIST_HEAD(&global->g_lu_gps_list);
+	INIT_LIST_HEAD(&global->g_se_tpg_list);
+	INIT_LIST_HEAD(&global->g_hba_list);
+	INIT_LIST_HEAD(&global->g_se_dev_list);
+	spin_lock_init(&global->g_device_lock);
+	spin_lock_init(&global->hba_lock);
+	spin_lock_init(&global->se_tpg_lock);
+	spin_lock_init(&global->lu_gps_lock);
+	spin_lock_init(&global->plugin_class_lock);
+
+	se_cmd_cache = kmem_cache_create("se_cmd_cache",
+			sizeof(se_cmd_t), __alignof__(se_cmd_t), 0, NULL);
+	if (!(se_cmd_cache)) {
+		printk(KERN_ERR "kmem_cache_create for se_cmd_t failed\n");
+		goto out;
+	}
+	se_task_cache = kmem_cache_create("se_task_cache",
+			sizeof(se_task_t), __alignof__(se_task_t), 0, NULL);
+	if (!(se_task_cache)) {
+		printk(KERN_ERR "kmem_cache_create for se_task_t failed\n");
+		goto out;
+	}
+	se_tmr_req_cache = kmem_cache_create("se_tmr_cache",
+			sizeof(se_tmr_req_t), __alignof__(se_tmr_req_t),
+			0, NULL);
+	if (!(se_tmr_req_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for se_tmr_req_t"
+				" failed\n");
+		goto out;
+	}
+	se_sess_cache = kmem_cache_create("se_sess_cache",
+			sizeof(se_session_t), __alignof__(se_session_t),
+			0, NULL);
+	if (!(se_sess_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for se_session_t"
+				" failed\n");
+		goto out;
+	}
+	se_hba_cache = kmem_cache_create("se_hba_cache",
+			sizeof(se_hba_t), __alignof__(se_hba_t),
+			0, NULL);
+	if (!(se_hba_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for se_hba_t"
+				" failed\n");
+		goto out;
+	}
+	se_ua_cache = kmem_cache_create("se_ua_cache",
+			sizeof(se_ua_t), __alignof__(se_ua_t), 0, NULL);
+	if (!(se_ua_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for se_ua_t failed\n");
+		goto out;
+	}
+	t10_pr_reg_cache = kmem_cache_create("t10_pr_reg_cache",
+			sizeof(t10_pr_registration_t),
+			__alignof__(t10_pr_registration_t), 0, NULL);
+	if (!(t10_pr_reg_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for t10_pr_registration_t"
+				" failed\n");
+		goto out;
+	}
+	t10_alua_lu_gp_cache = kmem_cache_create("t10_alua_lu_gp_cache",
+			sizeof(t10_alua_lu_gp_t), __alignof__(t10_alua_lu_gp_t),
+			0, NULL);
+	if (!(t10_alua_lu_gp_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for t10_alua_lu_gp_cache"
+				" failed\n");
+		goto out;
+	}
+	t10_alua_lu_gp_mem_cache = kmem_cache_create("t10_alua_lu_gp_mem_cache",
+			sizeof(t10_alua_lu_gp_member_t),
+			__alignof__(t10_alua_lu_gp_member_t), 0, NULL);
+	if (!(t10_alua_lu_gp_mem_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for t10_alua_lu_gp_mem_"
+				"cache failed\n");
+		goto out;
+	}
+	t10_alua_tg_pt_gp_cache = kmem_cache_create("t10_alua_tg_pt_gp_cache",
+			sizeof(t10_alua_tg_pt_gp_t),
+			__alignof__(t10_alua_tg_pt_gp_t), 0, NULL);
+	if (!(t10_alua_tg_pt_gp_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for t10_alua_tg_pt_gp_"
+				"cache failed\n");
+		goto out;
+	}
+	t10_alua_tg_pt_gp_mem_cache = kmem_cache_create(
+			"t10_alua_tg_pt_gp_mem_cache",
+			sizeof(t10_alua_tg_pt_gp_member_t),
+			__alignof__(t10_alua_tg_pt_gp_member_t),
+			0, NULL);
+	if (!(t10_alua_tg_pt_gp_mem_cache)) {
+		printk(KERN_ERR "kmem_cache_create() for t10_alua_tg_pt_gp_"
+				"mem_t failed\n");
+		goto out;
+	}
+
+	global->plugin_class_list = kzalloc((sizeof(se_plugin_class_t) *
+				MAX_PLUGIN_CLASSES), GFP_KERNEL);
+	if (!(global->plugin_class_list)) {
+		printk(KERN_ERR "Unable to allocate global->"
+			"plugin_class_list\n");
+		goto out;
+	}
+
+	se_global = global;
+
+	return 0;
+out:
+	kfree(global->plugin_class_list);
+	if (se_cmd_cache)
+		kmem_cache_destroy(se_cmd_cache);
+	if (se_task_cache)
+		kmem_cache_destroy(se_task_cache);
+	if (se_tmr_req_cache)
+		kmem_cache_destroy(se_tmr_req_cache);
+	if (se_sess_cache)
+		kmem_cache_destroy(se_sess_cache);
+	if (se_hba_cache)
+		kmem_cache_destroy(se_hba_cache);
+	if (se_ua_cache)
+		kmem_cache_destroy(se_ua_cache);
+	if (t10_pr_reg_cache)
+		kmem_cache_destroy(t10_pr_reg_cache);
+	if (t10_alua_lu_gp_cache)
+		kmem_cache_destroy(t10_alua_lu_gp_cache);
+	if (t10_alua_lu_gp_mem_cache)
+		kmem_cache_destroy(t10_alua_lu_gp_mem_cache);
+	if (t10_alua_tg_pt_gp_cache)
+		kmem_cache_destroy(t10_alua_tg_pt_gp_cache);
+	if (t10_alua_tg_pt_gp_mem_cache)
+		kmem_cache_destroy(t10_alua_tg_pt_gp_mem_cache);
+	kfree(global);
+	return -1;
+}
+
+void release_se_global(void)
+{
+	se_global_t *global;
+
+	global = se_global;
+	if (!(global))
+		return;
+
+	kfree(global->plugin_class_list);
+	kmem_cache_destroy(se_cmd_cache);
+	kmem_cache_destroy(se_task_cache);
+	kmem_cache_destroy(se_tmr_req_cache);
+	kmem_cache_destroy(se_sess_cache);
+	kmem_cache_destroy(se_hba_cache);
+	kmem_cache_destroy(se_ua_cache);
+	kmem_cache_destroy(t10_pr_reg_cache);
+	kmem_cache_destroy(t10_alua_lu_gp_cache);
+	kmem_cache_destroy(t10_alua_lu_gp_mem_cache);
+	kmem_cache_destroy(t10_alua_tg_pt_gp_cache);
+	kmem_cache_destroy(t10_alua_tg_pt_gp_mem_cache);
+	kfree(global);
+
+	se_global = NULL;
+}
+
+#ifdef DEBUG_DEV
+
+/* warning FIXME: PLUGIN API TODO */
+int __iscsi_debug_dev(se_device_t *dev)
+{
+	int fail_task = 0;
+	fd_dev_t *fd_dev;
+	iblock_dev_t *ib_dev;
+	rd_dev_t *rd_dev;
+	struct scsi_device *sd;
+
+	spin_lock(&se_global->debug_dev_lock);
+	switch (dev->se_hba->type) {
+	case PSCSI:
+		sd = (struct scsi_device *) dev->dev_ptr;
+		if (dev->dev_flags & DF_DEV_DEBUG) {
+			printk(KERN_INFO "HBA[%u] - Failing PSCSI Task for"
+				" %d/%d/%d\n", dev->se_hba->hba_id,
+				sd->channel, sd->id, sd->lun);
+			fail_task = 1;
+		}
+		break;
+	case IBLOCK:
+		ib_dev = (iblock_dev_t *) dev->dev_ptr;
+		if (dev->dev_flags & DF_DEV_DEBUG) {
+			printk(KERN_INFO "HBA[%u] - Failing IBLOCK Task for"
+				" %u/%u\n", dev->se_hba->hba_id,
+				ib_dev->ibd_major, ib_dev->ibd_minor);
+			fail_task = 1;
+		}
+		break;
+	case FILEIO:
+		fd_dev = (fd_dev_t *) dev->dev_ptr;
+		if (dev->dev_flags & DF_DEV_DEBUG) {
+			printk(KERN_INFO "HBA[%u] - Failing FILEIO Task for"
+				" %u\n", dev->se_hba->hba_id,
+				fd_dev->fd_dev_id);
+			fail_task = 1;
+		}
+		break;
+	case RAMDISK_DR:
+	case RAMDISK_MCP:
+		rd_dev = (rd_dev_t *) dev->dev_ptr;
+		if (dev->dev_flags & DF_DEV_DEBUG) {
+			printk(KERN_INFO "HBA[%u] - Failing RAMDISK Task for"
+				" %u\n", dev->se_hba->hba_id,
+				rd_dev->rd_dev_id);
+			fail_task = 1;
+		}
+		break;
+	default:
+		if (dev->dev_flags & DF_DEV_DEBUG) {
+			printk(KERN_INFO "HBA[%u] - Failing unknown Task\n",
+				dev->se_hba->hba_id);
+			fail_task = 1;
+		}
+		break;
+	}
+	spin_unlock(&se_global->debug_dev_lock);
+
+	return fail_task;
+}
+
+#endif /* DEBUG_DEV */
+
+/* #warning FIXME: transport_get_iqn_sn() for se_global_t */
+unsigned char *transport_get_iqn_sn(void)
+{
+	/*
+	 * Assume that for production WWN information will come through
+	 * ConfigFS at /sys/kernel/config/target/core/$HBA/$DEV/vpd_unit_serial
+	 */
+	return "1234567890";
+}
+
+void transport_init_queue_obj(se_queue_obj_t *qobj)
+{
+	atomic_set(&qobj->queue_cnt, 0);
+	INIT_LIST_HEAD(&qobj->qobj_list);
+	init_waitqueue_head(&qobj->thread_wq);
+	init_completion(&qobj->thread_create_comp);
+	init_completion(&qobj->thread_done_comp);
+	spin_lock_init(&qobj->cmd_queue_lock);
+}
+
+void transport_load_plugins(void)
+{
+	int ret = 0;
+
+#ifdef PARALLEL_SCSI
+	plugin_register((void *)&pscsi_template, pscsi_template.type,
+			pscsi_template.name, PLUGIN_TYPE_TRANSPORT,
+			pscsi_template.get_plugin_info, NULL, NULL, &ret);
+#endif
+#ifdef STGT_PLUGIN
+	plugin_register((void *)&stgt_template, stgt_template.type,
+			stgt_template.name, PLUGIN_TYPE_TRANSPORT,
+			stgt_template.get_plugin_info,
+			stgt_template.plugin_init,
+			stgt_template.plugin_free, &ret);
+#endif
+#ifdef PYX_IBLOCK
+	plugin_register((void *)&iblock_template, iblock_template.type,
+			iblock_template.name, PLUGIN_TYPE_TRANSPORT,
+			iblock_template.get_plugin_info, NULL, NULL, &ret);
+#endif
+#ifdef PYX_RAMDISK
+	plugin_register((void *)&rd_dr_template, rd_dr_template.type,
+			rd_dr_template.name, PLUGIN_TYPE_TRANSPORT,
+			rd_dr_template.get_plugin_info, NULL, NULL, &ret);
+	plugin_register((void *)&rd_mcp_template, rd_mcp_template.type,
+			rd_mcp_template.name, PLUGIN_TYPE_TRANSPORT,
+			rd_mcp_template.get_plugin_info, NULL, NULL, &ret);
+#endif
+#ifdef PYX_FILEIO
+	plugin_register((void *)&fileio_template, fileio_template.type,
+			fileio_template.name, PLUGIN_TYPE_TRANSPORT,
+			fileio_template.get_plugin_info, NULL, NULL, &ret);
+#endif
+}
+
+se_plugin_t *transport_core_get_plugin_by_name(const char *name)
+{
+	se_plugin_class_t *pc;
+	se_plugin_t *p;
+	int i;
+
+	pc = plugin_get_class(PLUGIN_TYPE_TRANSPORT);
+	if (!(pc))
+		return NULL;
+
+	for (i = 0; i < MAX_PLUGINS; i++) {
+		p = &pc->plugin_array[i];
+
+		if (!p->plugin_obj)
+			continue;
+
+		if (!(strncmp(name, p->plugin_name, strlen(p->plugin_name))))
+			return p;
+	}
+
+	return NULL;
+}
+
+void transport_check_dev_params_delim(char *ptr, char **cur)
+{
+	char *ptr2;
+
+	if (ptr) {
+		ptr2 = strstr(ptr, ",");
+		if ((ptr2)) {
+			*ptr2 = '\0';
+			*cur = (ptr2 + 1); /* Skip over comma */
+		} else
+			*cur = NULL;
+	}
+}
+
+se_session_t *transport_init_session(void)
+{
+	se_session_t *se_sess;
+
+	se_sess = kmem_cache_zalloc(se_sess_cache, GFP_KERNEL);
+	if (!(se_sess)) {
+		printk(KERN_ERR "Unable to allocate se_session_t from"
+				" se_sess_cache\n");
+		return ERR_PTR(-ENOMEM);
+	}
+	INIT_LIST_HEAD(&se_sess->sess_list);
+
+	return se_sess;
+}
+EXPORT_SYMBOL(transport_init_session);
+
+/*
+ * Called with spin_lock_bh(&se_portal_group_t->session_lock called.
+ */
+void __transport_register_session(
+	se_portal_group_t *se_tpg,
+	se_node_acl_t *se_nacl,
+	se_session_t *se_sess,
+	void *fabric_sess_ptr)
+{
+	se_sess->se_tpg = se_tpg;
+	se_sess->fabric_sess_ptr = fabric_sess_ptr;
+	/*
+	 * Used by se_node_acl_t's under ConfigFS to locate active se_session-t
+	 *
+	 * Only set for se_session_t's that will actually be moving I/O.
+	 * eg: *NOT* discovery sessions.
+	 */
+	if (se_nacl) {
+		spin_lock_bh(&se_nacl->nacl_sess_lock);
+		se_nacl->nacl_sess = se_sess;
+		spin_unlock_bh(&se_nacl->nacl_sess_lock);
+	}
+	list_add_tail(&se_sess->sess_list, &se_tpg->tpg_sess_list);
+
+	printk(KERN_INFO "TARGET_CORE[%s]: Registered fabric_sess_ptr: %p\n",
+		TPG_TFO(se_tpg)->get_fabric_name(), se_sess->fabric_sess_ptr);
+}
+EXPORT_SYMBOL(__transport_register_session);
+
+void transport_register_session(
+	se_portal_group_t *se_tpg,
+	se_node_acl_t *se_nacl,
+	se_session_t *se_sess,
+	void *fabric_sess_ptr)
+{
+	spin_lock_bh(&se_tpg->session_lock);
+	__transport_register_session(se_tpg, se_nacl, se_sess, fabric_sess_ptr);
+	spin_unlock_bh(&se_tpg->session_lock);
+}
+EXPORT_SYMBOL(transport_register_session);
+
+void transport_deregister_session_configfs(se_session_t *se_sess)
+{
+	se_node_acl_t *se_nacl;
+
+	/*
+	 * Used by se_node_acl_t's under ConfigFS to locate active se_session_t
+	 */
+	se_nacl = se_sess->se_node_acl;
+	if ((se_nacl)) {
+		spin_lock_bh(&se_nacl->nacl_sess_lock);
+		se_nacl->nacl_sess = NULL;
+		spin_unlock_bh(&se_nacl->nacl_sess_lock);
+	}
+}
+EXPORT_SYMBOL(transport_deregister_session_configfs);
+
+void transport_free_session(se_session_t *se_sess)
+{
+	kmem_cache_free(se_sess_cache, se_sess);
+}
+EXPORT_SYMBOL(transport_free_session);
+
+void transport_deregister_session(se_session_t *se_sess)
+{
+	se_portal_group_t *se_tpg = se_sess->se_tpg;
+	se_node_acl_t *se_nacl;
+
+	if (!(se_tpg)) {
+		transport_free_session(se_sess);
+		return;
+	}
+
+	spin_lock_bh(&se_tpg->session_lock);
+	list_del(&se_sess->sess_list);
+	se_sess->se_tpg = NULL;
+	se_sess->fabric_sess_ptr = NULL;
+	spin_unlock_bh(&se_tpg->session_lock);
+
+	/*
+	 * Determine if we need to do extra work for this initiator node's
+	 * se_node_acl_t if it had been previously dynamically generated.
+	 */
+	se_nacl = se_sess->se_node_acl;
+	if ((se_nacl)) {
+		spin_lock_bh(&se_tpg->acl_node_lock);
+		if (se_nacl->nodeacl_flags & NAF_DYNAMIC_NODE_ACL) {
+			if (!(TPG_TFO(se_tpg)->tpg_check_demo_mode_cache(
+					se_tpg))) {
+				list_del(&se_nacl->acl_list);
+				se_tpg->num_node_acls--;
+				spin_unlock_bh(&se_tpg->acl_node_lock);
+				core_free_device_list_for_node(se_nacl, se_tpg);
+				TPG_TFO(se_tpg)->tpg_release_fabric_acl(se_tpg,
+						se_nacl);
+				kfree(se_nacl);
+				spin_lock_bh(&se_tpg->acl_node_lock);
+			}
+		}
+		spin_unlock_bh(&se_tpg->acl_node_lock);
+	}
+
+	transport_free_session(se_sess);
+
+	printk(KERN_INFO "TARGET_CORE[%s]: Deregistered fabric_sess\n",
+		TPG_TFO(se_tpg)->get_fabric_name());
+}
+EXPORT_SYMBOL(transport_deregister_session);
+
+/*
+ * Called with T_TASK(cmd)->t_state_lock held.
+ */
+static void transport_all_task_dev_remove_state(se_cmd_t *cmd)
+{
+	se_device_t *dev;
+	se_task_t *task;
+	unsigned long flags;
+
+	if (!T_TASK(cmd))
+		return;
+
+	list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+		dev = task->se_dev;
+		if (!(dev))
+			continue;
+
+		if (atomic_read(&task->task_active))
+			continue;
+
+		if (!(atomic_read(&task->task_state_active)))
+			continue;
+
+		spin_lock_irqsave(&dev->execute_task_lock, flags);
+		list_del(&task->t_state_list);
+		DEBUG_TSTATE("Removed ITT: 0x%08x dev: %p task[%p]\n",
+			CMD_TFO(cmd)->tfo_get_task_tag(cmd), dev, task);
+		spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+		atomic_set(&task->task_state_active, 0);
+		atomic_dec(&T_TASK(cmd)->t_task_cdbs_ex_left);
+	}
+}
+
+/*
+ * Called with T_TASK(cmd)->t_state_lock held.
+ */
+void transport_task_dev_remove_state(se_task_t *task, se_device_t *dev)
+{
+	se_cmd_t *cmd = task->task_se_cmd;
+	unsigned long flags;
+
+	/*
+	 * We cannot remove the task from the state list while said task is
+	 * still active and probably timed out.
+	 */
+	if (atomic_read(&task->task_active)) {
+#if 0
+		printk(KERN_ERR "Skipping Removal of state for ITT: 0x%08x"
+			" dev: %p task[%p]\n"
+			CMD_TFO(task->task_se_cmd)->tfo_get_task_tag(
+			task->task_se_cmd), dev, task);
+#endif
+		return;
+	}
+
+	if (atomic_read(&task->task_state_active)) {
+		spin_lock_irqsave(&dev->execute_task_lock, flags);
+		list_del(&task->t_state_list);
+		DEBUG_TSTATE("Removed ITT: 0x%08x dev: %p task[%p]\n",
+			CMD_TFO(task->task_se_cmd)->tfo_get_task_tag(
+			task->task_se_cmd), dev, task);
+		spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+		atomic_set(&task->task_state_active, 0);
+		atomic_dec(&T_TASK(cmd)->t_task_cdbs_ex_left);
+	}
+}
+
+static void transport_passthrough_check_stop(se_cmd_t *cmd)
+{
+	if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+		return;
+
+	if (!cmd->transport_passthrough_done) {
+		if (cmd->callback) {
+			cmd->callback(cmd, cmd->callback_arg,
+				transport_passthrough_complete(cmd));
+		} else
+			complete(&T_TASK(cmd)->t_transport_passthrough_comp);
+
+		return;
+	}
+
+	cmd->transport_passthrough_done(cmd);
+}
+
+/*	transport_cmd_check_stop():
+ *
+ *	'transport_off = 1' determines if t_transport_active should be cleared.
+ *	'transport_off = 2' determines if task_dev_state should be removed.
+ *
+ *	A non-zero u8 t_state sets cmd->t_state.
+ *	Returns 1 when command is stopped, else 0.
+ */
+static int transport_cmd_check_stop(
+	se_cmd_t *cmd,
+	int transport_off,
+	u8 t_state)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	/*
+	 * Determine if IOCTL context caller in requesting the stopping of this
+	 * command for LUN shutdown purposes.
+	 */
+	if (atomic_read(&T_TASK(cmd)->transport_lun_stop)) {
+		DEBUG_CS("%s:%d atomic_read(&T_TASK(cmd)->transport_lun_stop)"
+			" == TRUE for ITT: 0x%08x\n", __func__, __LINE__,
+			CMD_TFO(cmd)->get_task_tag(cmd));
+
+		cmd->deferred_t_state = cmd->t_state;
+		cmd->t_state = TRANSPORT_DEFERRED_CMD;
+		atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+		if (transport_off == 2)
+			transport_all_task_dev_remove_state(cmd);
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+		complete(&T_TASK(cmd)->transport_lun_stop_comp);
+		return 1;
+	}
+	/*
+	 * Determine if frontend context caller is requesting the stopping of
+	 * this command for frontend excpections.
+	 */
+	if (atomic_read(&T_TASK(cmd)->t_transport_stop)) {
+		DEBUG_CS("%s:%d atomic_read(&T_TASK(cmd)->t_transport_stop) =="
+			" TRUE for ITT: 0x%08x\n", __func__, __LINE__,
+			CMD_TFO(cmd)->get_task_tag(cmd));
+
+		cmd->deferred_t_state = cmd->t_state;
+		cmd->t_state = TRANSPORT_DEFERRED_CMD;
+		if (transport_off == 2)
+			transport_all_task_dev_remove_state(cmd);
+
+		/*
+		 * Clear se_cmd_t->se_lun before the transport_off == 2 handoff
+		 * to FE.
+		 */
+		if ((transport_off == 2) && !(cmd->se_cmd_flags &
+						SCF_CMD_PASSTHROUGH))
+			cmd->se_lun = NULL;
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+		complete(&T_TASK(cmd)->t_transport_stop_comp);
+		return 1;
+	}
+	if (transport_off) {
+		atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+		if (transport_off == 2)
+			transport_all_task_dev_remove_state(cmd);
+
+		/*
+		 * Clear se_cmd_t->se_lun before the transport_off == 2 handoff
+		 * to FE.
+		 */
+		if ((transport_off == 2) && !(cmd->se_cmd_flags &
+						SCF_CMD_PASSTHROUGH))
+			cmd->se_lun = NULL;
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+		return 0;
+	} else if (t_state)
+		cmd->t_state = t_state;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	return 0;
+}
+
+static void transport_lun_remove_cmd(se_cmd_t *cmd)
+{
+	se_lun_t *lun = SE_LUN(cmd);
+	unsigned long flags;
+
+	if (!lun)
+		return;
+	/*
+	 * Do not track passthrough se_cmd_t for now..
+	 */
+	if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+		return;
+
+	spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+	if (atomic_read(&T_TASK(cmd)->transport_lun_active)) {
+		list_del(&cmd->se_lun_list);
+		atomic_set(&T_TASK(cmd)->transport_lun_active, 0);
+#if 0
+		printk(KERN_INFO "Removed ITT: 0x%08x from LUN LIST[%d]\n"
+			CMD_TFO(cmd)->get_task_tag(cmd), lun->unpacked_lun);
+#endif
+	}
+	spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
+}
+
+void transport_cmd_finish_abort(se_cmd_t *cmd, int remove)
+{
+	transport_remove_cmd_from_queue(cmd,
+		CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+			cmd->se_orig_obj_ptr));
+
+	transport_lun_remove_cmd(cmd);
+
+	if (!(transport_cmd_check_stop(cmd, 1, 0))) {
+		transport_passthrough_check_stop(cmd);
+		return;
+	}
+	if (remove)
+		transport_generic_remove(cmd, 0, 0);
+}
+
+void transport_cmd_finish_abort_tmr(se_cmd_t *cmd)
+{
+	transport_remove_cmd_from_queue(cmd,
+			CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+			cmd->se_orig_obj_ptr));
+
+	if (!(transport_cmd_check_stop(cmd, 1, 0))) {
+		transport_passthrough_check_stop(cmd);
+		return;
+	}
+	transport_generic_remove(cmd, 0, 0);
+}
+
+int transport_add_cmd_to_queue(
+	se_cmd_t *cmd,
+	se_queue_obj_t *qobj,
+	u8 t_state)
+{
+	se_queue_req_t *qr;
+	unsigned long flags;
+
+	qr = kzalloc(sizeof(se_queue_req_t), GFP_ATOMIC);
+	if (!(qr)) {
+		printk(KERN_ERR "Unable to allocate memory for"
+				" se_queue_req_t\n");
+		return -1;
+	}
+	INIT_LIST_HEAD(&qr->qr_list);
+
+	qr->cmd = (void *)cmd;
+	qr->state = t_state;
+
+	if (t_state) {
+		spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+		cmd->t_state = t_state;
+		atomic_set(&T_TASK(cmd)->t_transport_active, 1);
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+	}
+
+	spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
+	list_add_tail(&qr->qr_list, &qobj->qobj_list);
+	atomic_inc(&T_TASK(cmd)->t_transport_queue_active);
+	spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+	atomic_inc(&qobj->queue_cnt);
+	wake_up_interruptible(&qobj->thread_wq);
+	return 0;
+}
+
+static int transport_add_cmd_to_dev_queue(se_cmd_t *cmd, u8 t_state)
+{
+	se_device_t *dev = cmd->se_dev;
+
+	return transport_add_cmd_to_queue(cmd, dev->dev_queue_obj, t_state);
+}
+
+/*
+ * Called with se_queue_obj_t->cmd_queue_lock held.
+ */
+se_queue_req_t *__transport_get_qr_from_queue(se_queue_obj_t *qobj)
+{
+	se_cmd_t *cmd;
+	se_queue_req_t *qr = NULL;
+
+	if (list_empty(&qobj->qobj_list))
+		return NULL;
+
+	list_for_each_entry(qr, &qobj->qobj_list, qr_list)
+		break;
+
+	if (qr->cmd) {
+		cmd = (se_cmd_t *)qr->cmd;
+		atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
+	}
+	list_del(&qr->qr_list);
+	atomic_dec(&qobj->queue_cnt);
+
+	return qr;
+}
+
+se_queue_req_t *transport_get_qr_from_queue(se_queue_obj_t *qobj)
+{
+	se_cmd_t *cmd;
+	se_queue_req_t *qr;
+	unsigned long flags;
+
+	spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
+	if (list_empty(&qobj->qobj_list)) {
+		spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+		return NULL;
+	}
+
+	list_for_each_entry(qr, &qobj->qobj_list, qr_list)
+		break;
+
+	if (qr->cmd) {
+		cmd = (se_cmd_t *)qr->cmd;
+		atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
+	}
+	list_del(&qr->qr_list);
+	atomic_dec(&qobj->queue_cnt);
+	spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+	return qr;
+}
+
+void transport_remove_cmd_from_queue(se_cmd_t *cmd, se_queue_obj_t *qobj)
+{
+	se_cmd_t *q_cmd;
+	se_queue_req_t *qr = NULL, *qr_p = NULL;
+	unsigned long flags;
+
+	spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
+	if (!(atomic_read(&T_TASK(cmd)->t_transport_queue_active))) {
+		spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+		return;
+	}
+
+	list_for_each_entry_safe(qr, qr_p, &qobj->qobj_list, qr_list) {
+		q_cmd = (se_cmd_t *)qr->cmd;
+		if (q_cmd != cmd)
+			continue;
+
+		atomic_dec(&T_TASK(q_cmd)->t_transport_queue_active);
+		atomic_dec(&qobj->queue_cnt);
+		list_del(&qr->qr_list);
+		kfree(qr);
+	}
+	spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+	if (atomic_read(&T_TASK(cmd)->t_transport_queue_active)) {
+		printk(KERN_ERR "ITT: 0x%08x t_transport_queue_active: %d\n",
+			CMD_TFO(cmd)->get_task_tag(cmd),
+			atomic_read(&T_TASK(cmd)->t_transport_queue_active));
+	}
+}
+
+void transport_complete_cmd(se_cmd_t *cmd, int success)
+{
+	int t_state;
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (!success) {
+		cmd->transport_error_status = PYX_TRANSPORT_LU_COMM_FAILURE;
+		t_state = TRANSPORT_COMPLETE_FAILURE;
+	} else {
+		t_state = TRANSPORT_COMPLETE_OK;
+	}
+	atomic_set(&T_TASK(cmd)->t_transport_complete, 1);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	cmd->transport_add_cmd_to_queue(cmd, t_state);
+}
+
+/*	transport_complete_task():
+ *
+ *	Called from interrupt and non interrupt context depending
+ *	on the transport plugin.
+ */
+void transport_complete_task(se_task_t *task, int success)
+{
+	se_cmd_t *cmd = TASK_CMD(task);
+	se_device_t *dev = task->se_dev;
+	int t_state;
+	unsigned long flags;
+#if 0
+	printk(KERN_INFO "task: %p CDB: 0x%02x obj_ptr: %p\n", task,
+			T_TASK(cmd)->t_task_cdb[0], dev);
+#endif
+	if (dev) {
+		spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
+		atomic_inc(&dev->depth_left);
+		atomic_inc(&SE_HBA(dev)->left_queue_depth);
+		spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+	}
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	atomic_set(&task->task_active, 0);
+
+	/*
+	 * See if any sense data exists, if so set the TASK_SENSE flag.
+	 * Also check for any other post completion work that needs to be
+	 * done by the plugins.
+	 */
+	if (!dev)
+		goto check_task_stop;
+
+	if (TRANSPORT(dev)->transport_complete(task) != 0) {
+		cmd->se_cmd_flags |= SCF_TRANSPORT_TASK_SENSE;
+		task->task_sense = 1;
+		success = 1;
+	}
+
+	/*
+	 * See if we are waiting for outstanding se_task_t
+	 * to complete for an exception condition
+	 */
+check_task_stop:
+	if (atomic_read(&task->task_stop)) {
+		/*
+		 * Decrement T_TASK(cmd)->t_se_count if this task had
+		 * previously thrown its timeout exception handler.
+		 */
+		if (atomic_read(&task->task_timeout)) {
+			atomic_dec(&T_TASK(cmd)->t_se_count);
+			atomic_set(&task->task_timeout, 0);
+		}
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+		complete(&task->task_stop_comp);
+		return;
+	}
+	/*
+	 * If the task's timeout handler has fired, use the t_task_cdbs_timeout
+	 * left counter to determine when the se_cmd_t is ready to be queued to
+	 * the processing thread.
+	 */
+	if (atomic_read(&task->task_timeout)) {
+		if (!(atomic_dec_and_test(
+				&T_TASK(cmd)->t_task_cdbs_timeout_left))) {
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+				flags);
+			return;
+		}
+		t_state = TRANSPORT_COMPLETE_TIMEOUT;
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+		cmd->transport_add_cmd_to_queue(cmd, t_state);
+		return;
+	}
+	atomic_dec(&T_TASK(cmd)->t_task_cdbs_timeout_left);
+
+#ifdef DEBUG_DEV
+	if (dev) {
+		if (__iscsi_debug_dev(dev) != 0) {
+			success = 0;
+			task->task_scsi_status = 1;
+			cmd->transport_error_status =
+				PYX_TRANSPORT_LU_COMM_FAILURE;
+		}
+	}
+#endif /* DEBUG_DEV */
+
+	/*
+	 * Decrement the outstanding t_task_cdbs_left count.  The last
+	 * se_task_t from se_cmd_t will complete itself into the
+	 * device queue depending upon int success.
+	 */
+	if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_left))) {
+		if (!success)
+			T_TASK(cmd)->t_tasks_failed = 1;
+
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+
+	if (!success || T_TASK(cmd)->t_tasks_failed) {
+		t_state = TRANSPORT_COMPLETE_FAILURE;
+		if (!task->task_error_status) {
+			task->task_error_status =
+				PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+			cmd->transport_error_status =
+				PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+		}
+	} else {
+		atomic_set(&T_TASK(cmd)->t_transport_complete, 1);
+		t_state = TRANSPORT_COMPLETE_OK;
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	cmd->transport_add_cmd_to_queue(cmd, t_state);
+}
+
+/*
+ * Called by transport_add_tasks_from_cmd() once a se_cmd_t's
+ * se_task_t list are ready to be added to the active execution list
+ * se_device_t
+
+ * Called with se_dev_t->execute_task_lock called.
+ */
+static inline int transport_add_task_check_sam_attr(
+	se_task_t *task,
+	se_task_t *task_prev,
+	se_device_t *dev)
+{
+	/*
+	 * No SAM Task attribute emulation enabled, add to tail of
+	 * execution queue
+	 */
+	if (dev->dev_task_attr_type != SAM_TASK_ATTR_EMULATED) {
+		list_add_tail(&task->t_execute_list, &dev->execute_task_list);
+		return 0;
+	}
+	/*
+	 * HEAD_OF_QUEUE attribute for received CDB, which means
+	 * the first task that is associated with a se_cmd_t goes to
+	 * head of the se_device_t->execute_task_list, and task_prev
+	 * after that for each subsequent task
+	 */
+	if (task->task_se_cmd->sam_task_attr == TASK_ATTR_HOQ) {
+		list_add(&task->t_execute_list,
+				(task_prev != NULL) ?
+				&task_prev->t_execute_list :
+				&dev->execute_task_list);
+
+		DEBUG_STA("Set HEAD_OF_QUEUE for task CDB: 0x%02x"
+				" in execution queue\n",
+				T_TASK(task->task_se_cmd)->t_task_cdb[0]);
+		return 1;
+	}
+	/*
+	 * For ORDERED, SIMPLE or UNTAGGED attribute tasks once they have been
+	 * transitioned from Dermant -> Active state, and are added to the end
+	 * of the se_device_t->execute_task_list
+	 */
+	list_add_tail(&task->t_execute_list, &dev->execute_task_list);
+	return 0;
+}
+
+/*	__transport_add_task_to_execute_queue():
+ *
+ *	Called with se_dev_t->execute_task_lock called.
+ */
+static void __transport_add_task_to_execute_queue(
+	se_task_t *task,
+	se_task_t *task_prev,
+	se_device_t *dev)
+{
+	int head_of_queue;
+
+	head_of_queue = transport_add_task_check_sam_attr(task, task_prev, dev);
+	atomic_inc(&dev->execute_tasks);
+
+	if (atomic_read(&task->task_state_active))
+		return;
+	/*
+	 * Determine if this task needs to go to HEAD_OF_QUEUE for the
+	 * state list as well.  Running with SAM Task Attribute emulation
+	 * will always return head_of_queue == 0 here
+	 */
+	if (head_of_queue)
+		list_add(&task->t_state_list, (task_prev) ?
+				&task_prev->t_state_list :
+				&dev->state_task_list);
+	else
+		list_add_tail(&task->t_state_list, &dev->state_task_list);
+
+	atomic_set(&task->task_state_active, 1);
+
+	DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
+		CMD_TFO(task->task_se_cmd)->get_task_tag(task->task_se_cmd),
+		task, dev);
+}
+
+static void transport_add_tasks_to_state_queue(se_cmd_t *cmd)
+{
+	se_device_t *dev;
+	se_task_t *task;
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+		dev = task->se_dev;
+
+		if (atomic_read(&task->task_state_active))
+			continue;
+
+		spin_lock(&dev->execute_task_lock);
+		list_add_tail(&task->t_state_list, &dev->state_task_list);
+		atomic_set(&task->task_state_active, 1);
+
+		DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
+			CMD_TFO(task->task_se_cmd)->get_task_tag(
+			task->task_se_cmd), task, dev);
+
+		spin_unlock(&dev->execute_task_lock);
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+/*	transport_add_tasks_from_cmd():
+ *
+ *
+ */
+void transport_add_tasks_from_cmd(se_cmd_t *cmd)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	se_task_t *task, *task_prev = NULL;
+	unsigned long flags;
+
+	spin_lock_irqsave(&dev->execute_task_lock, flags);
+	list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+		if (atomic_read(&task->task_execute_queue))
+			continue;
+		/*
+		 * __transport_add_task_to_execute_queue() handles the
+		 * SAM Task Attribute emulation if enabled
+		 */
+		__transport_add_task_to_execute_queue(task, task_prev, dev);
+		atomic_set(&task->task_execute_queue, 1);
+		task_prev = task;
+	}
+	spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+	return;
+}
+
+/*	transport_get_task_from_execute_queue():
+ *
+ *	Called with dev->execute_task_lock held.
+ */
+se_task_t *transport_get_task_from_execute_queue(se_device_t *dev)
+{
+	se_task_t *task;
+
+	if (list_empty(&dev->execute_task_list))
+		return NULL;
+
+	list_for_each_entry(task, &dev->execute_task_list, t_execute_list)
+		break;
+
+	list_del(&task->t_execute_list);
+	atomic_dec(&dev->execute_tasks);
+
+	return task;
+}
+
+/*	transport_remove_task_from_execute_queue():
+ *
+ *
+ */
+static void transport_remove_task_from_execute_queue(
+	se_task_t *task,
+	se_device_t *dev)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&dev->execute_task_lock, flags);
+	list_del(&task->t_execute_list);
+	atomic_dec(&dev->execute_tasks);
+	spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+}
+
+/*	transport_check_device_tcq():
+ *
+ *
+ */
+int transport_check_device_tcq(
+	se_device_t *dev,
+	u32 unpacked_lun,
+	u32 device_tcq)
+{
+	if (device_tcq > dev->queue_depth) {
+		printk(KERN_ERR "Attempting to set storage device queue depth"
+			" to %d while transport maximum is %d on LUN: %u,"
+			" ignoring request\n", device_tcq, dev->queue_depth,
+			unpacked_lun);
+		return -1;
+	} else if (!device_tcq) {
+		printk(KERN_ERR "Attempting to set storage device queue depth"
+			" to 0 on LUN: %u, ignoring request\n", unpacked_lun);
+		return -1;
+	}
+
+	dev->queue_depth = device_tcq;
+	atomic_set(&dev->depth_left, dev->queue_depth);
+	printk(KERN_INFO "Reset Device Queue Depth to %u for Logical Unit"
+		" Number: %u\n", dev->queue_depth, unpacked_lun);
+
+	return 0;
+}
+EXPORT_SYMBOL(transport_check_device_tcq);
+
+unsigned char *transport_dump_cmd_direction(se_cmd_t *cmd)
+{
+	switch (cmd->data_direction) {
+	case SE_DIRECTION_NONE:
+		return "NONE";
+	case SE_DIRECTION_READ:
+		return "READ";
+	case SE_DIRECTION_WRITE:
+		return "WRITE";
+	case SE_DIRECTION_BIDI:
+		return "BIDI";
+	default:
+		break;
+	}
+
+	return "UNKNOWN";
+}
+
+void transport_dump_dev_state(
+	se_device_t *dev,
+	char *b,
+	int *bl)
+{
+	*bl += sprintf(b + *bl, "Status: ");
+	switch (dev->dev_status) {
+	case TRANSPORT_DEVICE_ACTIVATED:
+		*bl += sprintf(b + *bl, "ACTIVATED");
+		break;
+	case TRANSPORT_DEVICE_DEACTIVATED:
+		*bl += sprintf(b + *bl, "DEACTIVATED");
+		break;
+	case TRANSPORT_DEVICE_SHUTDOWN:
+		*bl += sprintf(b + *bl, "SHUTDOWN");
+		break;
+	case TRANSPORT_DEVICE_OFFLINE_ACTIVATED:
+	case TRANSPORT_DEVICE_OFFLINE_DEACTIVATED:
+		*bl += sprintf(b + *bl, "OFFLINE");
+		break;
+	default:
+		*bl += sprintf(b + *bl, "UNKNOWN=%d", dev->dev_status);
+		break;
+	}
+
+	*bl += sprintf(b + *bl, "  Execute/Left/Max Queue Depth: %d/%d/%d",
+		atomic_read(&dev->execute_tasks), atomic_read(&dev->depth_left),
+		dev->queue_depth);
+	*bl += sprintf(b + *bl, "  SectorSize: %u  MaxSectors: %u\n",
+		DEV_ATTRIB(dev)->block_size, DEV_ATTRIB(dev)->max_sectors);
+	*bl += sprintf(b + *bl, "        ");
+}
+
+void transport_dump_dev_info(
+	se_device_t *dev,
+	se_lun_t *lun,
+	unsigned long long total_bytes,
+	char *b,        /* Pointer to info buffer */
+	int *bl)
+{
+	se_subsystem_api_t *t;
+	int ret = 0;
+
+	t = (se_subsystem_api_t *)plugin_get_obj(PLUGIN_TYPE_TRANSPORT,
+			dev->type, &ret);
+	if (!t || (ret != 0))
+		return;
+
+	t->get_dev_info(dev, b, bl);
+	*bl += sprintf(b + *bl, "        ");
+	*bl += sprintf(b + *bl, "Type: %s ",
+		scsi_device_type(TRANSPORT(dev)->get_device_type(dev)));
+	*bl += sprintf(b + *bl, "ANSI SCSI revision: %02x  ",
+		TRANSPORT(dev)->get_device_rev(dev));
+
+	if (DEV_OBJ_API(dev)->get_t10_wwn) {
+		t10_wwn_t *wwn = DEV_OBJ_API(dev)->get_t10_wwn((void *)dev);
+
+		*bl += sprintf(b + *bl, "Unit Serial: %s  ",
+			((strlen(wwn->unit_serial) != 0) ?
+			(char *)wwn->unit_serial : "None"));
+	}
+	*bl += sprintf(b + *bl, "%s", "DIRECT");
+
+	if ((DEV_OBJ_API(dev)->check_count(&dev->dev_access_obj)) ||
+	    (DEV_OBJ_API(dev)->check_count(&dev->dev_feature_obj)))
+		*bl += sprintf(b + *bl, "  ACCESSED\n");
+	else if (DEV_OBJ_API(dev)->check_count(&dev->dev_export_obj))
+		*bl += sprintf(b + *bl, "  EXPORTED\n");
+	else
+		*bl += sprintf(b + *bl, "  FREE\n");
+
+	if (lun) {
+		*bl += sprintf(b + *bl, "        Core Host ID: %u LUN: %u",
+			dev->se_hba->hba_id, lun->unpacked_lun);
+		if (!(TRANSPORT(dev)->get_device_type(dev))) {
+			*bl += sprintf(b + *bl, "  Active Cmds: %d  Total Bytes"
+				": %llu\n", atomic_read(&dev->active_cmds),
+			total_bytes);
+		} else {
+			*bl += sprintf(b + *bl, "  Active Cmds: %d\n",
+				atomic_read(&dev->active_cmds));
+		}
+	} else {
+		if (!(TRANSPORT(dev)->get_device_type(dev))) {
+			*bl += sprintf(b + *bl, "        Core Host ID: %u"
+				"  Active Cmds: %d  Total Bytes: %llu\n",
+				dev->se_hba->hba_id,
+				atomic_read(&dev->active_cmds), total_bytes);
+		} else {
+			*bl += sprintf(b + *bl, "        CoreI Host ID: %u"
+				"  Active Cmds: %d\n", dev->se_hba->hba_id,
+				atomic_read(&dev->active_cmds));
+		}
+	}
+}
+
+/*	transport_release_all_cmds():
+ *
+ *
+ */
+static void transport_release_all_cmds(se_device_t *dev)
+{
+	se_cmd_t *cmd = NULL;
+	se_queue_req_t *qr = NULL, *qr_p = NULL;
+	int bug_out = 0, t_state;
+	unsigned long flags;
+
+	spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+	list_for_each_entry_safe(qr, qr_p, &dev->dev_queue_obj->qobj_list,
+				qr_list) {
+
+		cmd = (se_cmd_t *)qr->cmd;
+		t_state = qr->state;
+		list_del(&qr->qr_list);
+		kfree(qr);
+		spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock,
+				flags);
+
+		printk(KERN_ERR "Releasing %s ITT: 0x%08x, i_state: %u,"
+			" t_state: %u directly\n",
+			(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH) ?
+			"Passthrough" : "Normal",
+			CMD_TFO(cmd)->get_task_tag(cmd),
+			CMD_TFO(cmd)->get_cmd_state(cmd), t_state);
+
+		transport_release_fe_cmd(cmd);
+		bug_out = 1;
+
+		spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+	}
+	spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock, flags);
+#if 0
+	if (bug_out)
+		BUG();
+#endif
+}
+
+/*	transport_dev_write_pending_nop():
+ *
+ *
+ */
+static int transport_dev_write_pending_nop(se_task_t *task)
+{
+	return 0;
+}
+
+static int transport_get_inquiry(
+	se_obj_lun_type_t *obj_api,
+	t10_wwn_t *wwn,
+	void *obj_ptr)
+{
+	se_cmd_t *cmd;
+	unsigned char *buf;
+	int i;
+	unsigned char cdb[SCSI_CDB_SIZE];
+
+	memset(cdb, 0, SCSI_CDB_SIZE);
+	cdb[0] = INQUIRY;
+	cdb[3] = (INQUIRY_LEN >> 8) & 0xff;
+	cdb[4] = (INQUIRY_LEN & 0xff);
+
+	cmd = transport_allocate_passthrough(&cdb[0],  SE_DIRECTION_READ,
+			0, NULL, 0, INQUIRY_LEN, obj_api, obj_ptr);
+	if (!(cmd))
+		return -1;
+
+	if (transport_generic_passthrough(cmd) < 0) {
+		transport_passthrough_release(cmd);
+		return -1;
+	}
+
+	buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+	/*
+	 * Save the basic Vendor, Model and Revision in passed t10_wwn_t.
+	 * We will obtain the VPD in a seperate passthrough operation.
+	 */
+	memcpy((void *)&wwn->vendor[0], (void *)&buf[8],
+			sizeof(wwn->vendor));
+	memcpy((void *)&wwn->model[0], (void *)&buf[16],
+			sizeof(wwn->model));
+	memcpy((void *)&wwn->revision[0], (void *)&buf[32],
+			sizeof(wwn->revision));
+
+	printk("  Vendor: ");
+	for (i = 8; i < 16; i++)
+		if (buf[i] >= 0x20 && i < buf[4] + 5)
+			printk("%c", buf[i]);
+		else
+			printk(" ");
+
+	printk("  Model: ");
+	for (i = 16; i < 32; i++)
+		if (buf[i] >= 0x20 && i < buf[4] + 5)
+			printk("%c", buf[i]);
+		else
+			printk(" ");
+
+	printk("  Revision: ");
+	for (i = 32; i < 36; i++)
+		if (buf[i] >= 0x20 && i < buf[4] + 5)
+			printk("%c", buf[i]);
+		else
+			printk(" ");
+
+	printk("\n");
+
+	i = buf[0] & 0x1f;
+
+	printk("  Type:   %s ", scsi_device_type(i));
+	printk("                 ANSI SCSI revision: %02x",
+				buf[2] & 0x07);
+	if ((buf[2] & 0x07) == 1 && (buf[3] & 0x0f) == 1)
+		printk(" CCS\n");
+	else
+		printk("\n");
+
+	transport_passthrough_release(cmd);
+	return 0;
+}
+
+static int transport_get_inquiry_vpd_serial(
+	se_obj_lun_type_t *obj_api,
+	t10_wwn_t *wwn,
+	void *obj_ptr)
+{
+	unsigned char *buf;
+	se_cmd_t *cmd;
+	unsigned char cdb[SCSI_CDB_SIZE];
+
+	memset(cdb, 0, SCSI_CDB_SIZE);
+	cdb[0] = INQUIRY;
+	cdb[1] = 0x01; /* Query VPD */
+	cdb[2] = 0x80; /* Unit Serial Number */
+	cdb[3] = (INQUIRY_VPD_SERIAL_LEN >> 8) & 0xff;
+	cdb[4] = (INQUIRY_VPD_SERIAL_LEN & 0xff);
+
+	cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+			0, NULL, 0, INQUIRY_VPD_SERIAL_LEN, obj_api, obj_ptr);
+	if (!(cmd))
+		return -1;
+
+	if (transport_generic_passthrough(cmd) < 0) {
+		transport_passthrough_release(cmd);
+		return -1;
+	}
+
+	buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+
+	printk(KERN_INFO "T10 VPD Unit Serial Number: %s\n", &buf[4]);
+	snprintf(&wwn->unit_serial[0], INQUIRY_VPD_SERIAL_LEN, "%s", &buf[4]);
+
+	transport_passthrough_release(cmd);
+	return 0;
+}
+
+static const char hex_str[] = "0123456789abcdef";
+
+void transport_dump_vpd_proto_id(
+	t10_vpd_t *vpd,
+	unsigned char *p_buf,
+	int p_buf_len)
+{
+	unsigned char buf[VPD_TMP_BUF_SIZE];
+	int len;
+
+	memset(buf, 0, VPD_TMP_BUF_SIZE);
+	len = sprintf(buf, "T10 VPD Protocol Identifier: ");
+
+	switch (vpd->protocol_identifier) {
+	case 0x00:
+		sprintf(buf+len, "Fibre Channel\n");
+		break;
+	case 0x10:
+		sprintf(buf+len, "Parallel SCSI\n");
+		break;
+	case 0x20:
+		sprintf(buf+len, "SSA\n");
+		break;
+	case 0x30:
+		sprintf(buf+len, "IEEE 1394\n");
+		break;
+	case 0x40:
+		sprintf(buf+len, "SCSI Remote Direct Memory Access"
+				" Protocol\n");
+		break;
+	case 0x50:
+		sprintf(buf+len, "Internet SCSI (iSCSI)\n");
+		break;
+	case 0x60:
+		sprintf(buf+len, "SAS Serial SCSI Protocol\n");
+		break;
+	case 0x70:
+		sprintf(buf+len, "Automation/Drive Interface Transport"
+				" Protocol\n");
+		break;
+	case 0x80:
+		sprintf(buf+len, "AT Attachment Interface ATA/ATAPI\n");
+		break;
+	default:
+		sprintf(buf+len, "Unknown 0x%02x\n",
+				vpd->protocol_identifier);
+		break;
+	}
+
+	if (p_buf)
+		strncpy(p_buf, buf, p_buf_len);
+	else
+		printk(KERN_INFO "%s", buf);
+}
+
+void transport_set_vpd_proto_id(t10_vpd_t *vpd, unsigned char *page_83)
+{
+	/*
+	 * Check if the Protocol Identifier Valid (PIV) bit is set..
+	 *
+	 * from spc3r23.pdf section 7.5.1
+	 */
+	if (page_83[1] & 0x80) {
+		vpd->protocol_identifier = (page_83[0] & 0xf0);
+		vpd->protocol_identifier_set = 1;
+		transport_dump_vpd_proto_id(vpd, NULL, 0);
+	}
+}
+
+int transport_dump_vpd_assoc(
+	t10_vpd_t *vpd,
+	unsigned char *p_buf,
+	int p_buf_len)
+{
+	unsigned char buf[VPD_TMP_BUF_SIZE];
+	int ret = 0, len;
+
+	memset(buf, 0, VPD_TMP_BUF_SIZE);
+	len = sprintf(buf, "T10 VPD Identifier Association: ");
+
+	switch (vpd->association) {
+	case 0x00:
+		sprintf(buf+len, "addressed logical unit\n");
+		break;
+	case 0x10:
+		sprintf(buf+len, "target port\n");
+		break;
+	case 0x20:
+		sprintf(buf+len, "SCSI target device\n");
+		break;
+	default:
+		sprintf(buf+len, "Unknown 0x%02x\n", vpd->association);
+		ret = -1;
+		break;
+	}
+
+	if (p_buf)
+		strncpy(p_buf, buf, p_buf_len);
+	else
+		printk("%s", buf);
+
+	return ret;
+}
+
+static int transport_set_vpd_assoc(t10_vpd_t *vpd, unsigned char *page_83)
+{
+	/*
+	 * The VPD identification association..
+	 *
+	 * from spc3r23.pdf Section 7.6.3.1 Table 297
+	 */
+	vpd->association = (page_83[1] & 0x30);
+	return transport_dump_vpd_assoc(vpd, NULL, 0);
+}
+
+int transport_dump_vpd_ident_type(
+	t10_vpd_t *vpd,
+	unsigned char *p_buf,
+	int p_buf_len)
+{
+	unsigned char buf[VPD_TMP_BUF_SIZE];
+	int ret = 0, len;
+
+	memset(buf, 0, VPD_TMP_BUF_SIZE);
+	len = sprintf(buf, "T10 VPD Identifier Type: ");
+
+	switch (vpd->device_identifier_type) {
+	case 0x00:
+		sprintf(buf+len, "Vendor specific\n");
+		break;
+	case 0x01:
+		sprintf(buf+len, "T10 Vendor ID based\n");
+		break;
+	case 0x02:
+		sprintf(buf+len, "EUI-64 based\n");
+		break;
+	case 0x03:
+		sprintf(buf+len, "NAA\n");
+		break;
+	case 0x04:
+		sprintf(buf+len, "Relative target port identifier\n");
+		break;
+	case 0x08:
+		sprintf(buf+len, "SCSI name string\n");
+		break;
+	default:
+		sprintf(buf+len, "Unsupported: 0x%02x\n",
+				vpd->device_identifier_type);
+		ret = -1;
+		break;
+	}
+
+	if (p_buf)
+		strncpy(p_buf, buf, p_buf_len);
+	else
+		printk("%s", buf);
+
+	return ret;
+}
+
+int transport_set_vpd_ident_type(t10_vpd_t *vpd, unsigned char *page_83)
+{
+	/*
+	 * The VPD identifier type..
+	 *
+	 * from spc3r23.pdf Section 7.6.3.1 Table 298
+	 */
+	vpd->device_identifier_type = (page_83[1] & 0x0f);
+	return transport_dump_vpd_ident_type(vpd, NULL, 0);
+}
+
+int transport_dump_vpd_ident(
+	t10_vpd_t *vpd,
+	unsigned char *p_buf,
+	int p_buf_len)
+{
+	unsigned char buf[VPD_TMP_BUF_SIZE];
+	int ret = 0;
+
+	memset(buf, 0, VPD_TMP_BUF_SIZE);
+
+	switch (vpd->device_identifier_code_set) {
+	case 0x01: /* Binary */
+		sprintf(buf, "T10 VPD Binary Device Identifier: %s\n",
+			&vpd->device_identifier[0]);
+		break;
+	case 0x02: /* ASCII */
+		sprintf(buf, "T10 VPD ASCII Device Identifier: %s\n",
+			&vpd->device_identifier[0]);
+		break;
+	case 0x03: /* UTF-8 */
+		sprintf(buf, "T10 VPD UTF-8 Device Identifier: %s\n",
+			&vpd->device_identifier[0]);
+		break;
+	default:
+		sprintf(buf, "T10 VPD Device Identifier encoding unsupported:"
+			" 0x%02x", vpd->device_identifier_code_set);
+		ret = -1;
+		break;
+	}
+
+	if (p_buf)
+		strncpy(p_buf, buf, p_buf_len);
+	else
+		printk("%s", buf);
+
+	return ret;
+}
+
+int transport_set_vpd_ident(t10_vpd_t *vpd, unsigned char *page_83)
+{
+	int j = 0, i = 4; /* offset to start of the identifer */
+
+	/*
+	 * The VPD Code Set (encoding)
+	 *
+	 * from spc3r23.pdf Section 7.6.3.1 Table 296
+	 */
+	vpd->device_identifier_code_set = (page_83[0] & 0x0f);
+	switch (vpd->device_identifier_code_set) {
+	case 0x01: /* Binary */
+		vpd->device_identifier[j++] =
+				hex_str[vpd->device_identifier_type];
+		while (i < (4 + page_83[3])) {
+			vpd->device_identifier[j++] =
+				hex_str[(page_83[i] & 0xf0) >> 4];
+			vpd->device_identifier[j++] =
+				hex_str[page_83[i] & 0x0f];
+			i++;
+		}
+		break;
+	case 0x02: /* ASCII */
+	case 0x03: /* UTF-8 */
+		while (i < (4 + page_83[3]))
+			vpd->device_identifier[j++] = page_83[i++];
+
+		break;
+	default:
+		break;
+	}
+
+	return transport_dump_vpd_ident(vpd, NULL, 0);
+}
+
+static int transport_get_inquiry_vpd_device_ident(
+	se_obj_lun_type_t *obj_api,
+	t10_wwn_t *wwn,
+	void *obj_ptr)
+{
+	unsigned char *buf, *page_83;
+	se_cmd_t *cmd;
+	t10_vpd_t *vpd;
+	unsigned char cdb[SCSI_CDB_SIZE];
+	int ident_len, page_len, off = 4, ret = 0;
+
+	memset(cdb, 0, SCSI_CDB_SIZE);
+	cdb[0] = INQUIRY;
+	cdb[1] = 0x01; /* Query VPD */
+	cdb[2] = 0x83; /* Device Identifier */
+	cdb[3] = (INQUIRY_VPD_DEVICE_IDENTIFIER_LEN >> 8) & 0xff;
+	cdb[4] = (INQUIRY_VPD_DEVICE_IDENTIFIER_LEN & 0xff);
+
+	cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+			0, NULL, 0, INQUIRY_VPD_DEVICE_IDENTIFIER_LEN,
+			obj_api, obj_ptr);
+	if (!(cmd))
+		return -1;
+
+	if (transport_generic_passthrough(cmd) < 0) {
+		transport_passthrough_release(cmd);
+		return -1;
+	}
+
+	buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+	page_len = (buf[2] << 8) | buf[3];
+	printk("T10 VPD Page Length: %d\n", page_len);
+
+	while (page_len > 0) {
+		/* Grab a pointer to the Identification descriptor */
+		page_83 = &buf[off];
+		ident_len = page_83[3];
+		if (!(ident_len)) {
+			printk(KERN_ERR "page_83[3]: identifier"
+					" length zero!\n");
+			break;
+		}
+		printk(KERN_INFO "T10 VPD Identifer Length: %d\n", ident_len);
+
+		vpd = kzalloc(sizeof(t10_vpd_t), GFP_KERNEL);
+		if (!(vpd)) {
+			printk(KERN_ERR "Unable to allocate memory for"
+					" t10_vpd_t\n");
+			ret = -1;
+			goto out;
+		}
+		INIT_LIST_HEAD(&vpd->vpd_list);
+
+		transport_set_vpd_proto_id(vpd, page_83);
+		transport_set_vpd_assoc(vpd, page_83);
+
+		if (transport_set_vpd_ident_type(vpd, page_83) < 0) {
+			off += (ident_len + 4);
+			page_len -= (ident_len + 4);
+			kfree(vpd);
+			continue;
+		}
+		if (transport_set_vpd_ident(vpd, page_83) < 0) {
+			off += (ident_len + 4);
+			page_len -= (ident_len + 4);
+			kfree(vpd);
+			continue;
+		}
+
+		list_add_tail(&vpd->vpd_list, &wwn->t10_vpd_list);
+		off += (ident_len + 4);
+		page_len -= (ident_len + 4);
+	}
+out:
+	transport_passthrough_release(cmd);
+	return 0;
+}
+
+int transport_rescan_evpd_device_ident(
+	se_device_t *dev)
+{
+	se_release_vpd_for_dev(dev);
+	transport_get_inquiry_vpd_device_ident(DEV_OBJ_API(dev),
+			DEV_T10_WWN(dev), (void *)dev);
+	return 0;
+}
+
+static int transport_get_read_capacity(se_device_t *dev)
+{
+	unsigned char cdb[SCSI_CDB_SIZE], *buf;
+	u32 blocks, v1, v2;
+	se_cmd_t *cmd;
+	unsigned long long blocks_long;
+
+	memset(cdb, 0, SCSI_CDB_SIZE);
+	cdb[0] = 0x25; /* READ_CAPACITY */
+
+	cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+			0, NULL, 0, READ_CAP_LEN, DEV_OBJ_API(dev),
+			(void *)dev);
+	if (!(cmd))
+		return -1;
+
+	if (transport_generic_passthrough(cmd) < 0) {
+		transport_passthrough_release(cmd);
+		return -1;
+	}
+
+	buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+	blocks = (buf[0] << 24) | (buf[1] << 16) | (buf[2] << 8) | buf[3];
+
+	transport_passthrough_release(cmd);
+
+	if (blocks != 0xFFFFFFFF) {
+		dev->dev_sectors_total = blocks;
+		dev->dev_generate_cdb = &split_cdb_RW_10;
+		return 0;
+	}
+
+	printk(KERN_INFO "READ_CAPACITY returned 0xFFFFFFFF, issuing"
+			" SAI_READ_CAPACITY_16\n");
+
+	memset(cdb, 0, SCSI_CDB_SIZE);
+	cdb[0] = 0x9e; /* SERVICE_ACTION_IN */
+	cdb[1] = 0x10; /* SAI_READ_CAPACITY_16 */
+	cdb[13] = 12;
+
+	cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+			0, NULL, 0, 12, DEV_OBJ_API(dev), (void *)dev);
+	if (!(cmd))
+		return -1;
+
+	if (transport_generic_passthrough(cmd) < 0) {
+		transport_passthrough_release(cmd);
+		return -1;
+	}
+
+	buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+	v1 = (buf[0] << 24) | (buf[1] << 16) | (buf[2] << 8) | buf[3];
+	v2 = (buf[4] << 24) | (buf[5] << 16) | (buf[6] << 8) | buf[7];
+	blocks_long = ((unsigned long long)v2 | (unsigned long long)v1 << 32);
+
+	transport_passthrough_release(cmd);
+
+	dev->dev_sectors_total = blocks_long;
+	dev->dev_generate_cdb = &split_cdb_RW_16;
+
+	return 0;
+}
+
+static void core_setup_task_attr_emulation(se_device_t *dev)
+{
+	/*
+	 * If this device is from Target_Core_Mod/pSCSI, disable the
+	 * SAM Task Attribute emulation.
+	 *
+	 * This is currently not available in upsream Linux/SCSI Target
+	 * mode code, and is assumed to be disabled while using TCM/pSCSI.
+	 */
+	if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
+		dev->dev_task_attr_type = SAM_TASK_ATTR_PASSTHROUGH;
+		return;
+	}
+
+	dev->dev_task_attr_type = SAM_TASK_ATTR_EMULATED;
+	DEBUG_STA("%s: Using SAM_TASK_ATTR_EMULATED for SPC: 0x%02x"
+		" device\n", TRANSPORT(dev)->name,
+		TRANSPORT(dev)->get_device_rev(dev));
+}
+
+/*	transport_add_device_to_core_hba():
+ *
+ *	Note that some plugins (IBLOCK) will pass device_flags ==
+ *	DF_CLAIMED_BLOCKDEV signifying OS that a dependent block_device
+ *	has been claimed.  In exception cases we will release said
+ *	block_device ourselves.
+ */
+se_device_t *transport_add_device_to_core_hba(
+	se_hba_t *hba,
+	se_subsystem_api_t *transport,
+	se_subsystem_dev_t *se_dev,
+	u32 device_flags,
+	void *transport_dev)
+{
+	int ret = 0;
+	se_device_t  *dev;
+
+	dev = kzalloc(sizeof(se_device_t), GFP_KERNEL);
+	if (!(dev)) {
+		printk(KERN_ERR "Unable to allocate memory for se_dev_t\n");
+		return NULL;
+	}
+	dev->dev_queue_obj = kzalloc(sizeof(se_queue_obj_t), GFP_KERNEL);
+	if (!(dev->dev_queue_obj)) {
+		printk(KERN_ERR "Unable to allocate memory for"
+				" dev->dev_queue_obj\n");
+		kfree(dev);
+		return NULL;
+	}
+	transport_init_queue_obj(dev->dev_queue_obj);
+
+	dev->dev_status_queue_obj = kzalloc(sizeof(se_queue_obj_t),
+					GFP_KERNEL);
+	if (!(dev->dev_status_queue_obj)) {
+		printk(KERN_ERR "Unable to allocate memory for"
+				" dev->dev_status_queue_obj\n");
+		kfree(dev->dev_queue_obj);
+		kfree(dev);
+		return NULL;
+	}
+	transport_init_queue_obj(dev->dev_status_queue_obj);
+
+	dev->dev_flags		= device_flags;
+	dev->dev_status		|= TRANSPORT_DEVICE_DEACTIVATED;
+	dev->type		= transport->type;
+	dev->dev_ptr		= (void *) transport_dev;
+	dev->se_hba		= hba;
+	dev->se_sub_dev		= se_dev;
+	dev->transport		= transport;
+	atomic_set(&dev->active_cmds, 0);
+	INIT_LIST_HEAD(&dev->dev_list);
+	INIT_LIST_HEAD(&dev->dev_sep_list);
+	INIT_LIST_HEAD(&dev->dev_tmr_list);
+	INIT_LIST_HEAD(&dev->execute_task_list);
+	INIT_LIST_HEAD(&dev->delayed_cmd_list);
+	INIT_LIST_HEAD(&dev->ordered_cmd_list);
+	INIT_LIST_HEAD(&dev->state_task_list);
+	spin_lock_init(&dev->execute_task_lock);
+	spin_lock_init(&dev->delayed_cmd_lock);
+	spin_lock_init(&dev->ordered_cmd_lock);
+	spin_lock_init(&dev->state_task_lock);
+	spin_lock_init(&dev->dev_alua_lock);
+	spin_lock_init(&dev->dev_reservation_lock);
+	spin_lock_init(&dev->dev_status_lock);
+	spin_lock_init(&dev->dev_status_thr_lock);
+	spin_lock_init(&dev->se_port_lock);
+	spin_lock_init(&dev->se_tmr_lock);
+
+	dev->queue_depth	= TRANSPORT(dev)->get_queue_depth(dev);
+	atomic_set(&dev->depth_left, dev->queue_depth);
+	atomic_set(&dev->dev_ordered_id, 0);
+
+	se_dev_set_default_attribs(dev);
+
+	dev->write_pending = (transport->write_pending) ?
+		transport->write_pending : &transport_dev_write_pending_nop;
+
+#ifdef SNMP_SUPPORT
+	dev->dev_index = scsi_get_new_index(SCSI_DEVICE_INDEX);
+	dev->creation_time = get_jiffies_64();
+	spin_lock_init(&dev->stats_lock);
+#endif /* SNMP_SUPPORT */
+
+	spin_lock(&hba->device_lock);
+	list_add_tail(&dev->dev_list, &hba->hba_dev_list);
+	hba->dev_count++;
+	spin_unlock(&hba->device_lock);
+
+	/*
+	 * Get this se_device_t's API from the device object plugin.
+	 */
+	dev->dev_obj_api = se_obj_get_api(TRANSPORT_LUN_TYPE_DEVICE);
+	if (!(dev->dev_obj_api))
+		goto out;
+	/*
+	 * Setup the SAM Task Attribute emulation for se_device_t
+	 */
+	core_setup_task_attr_emulation(dev);
+	/*
+	 * Setup the Reservations infrastructure for se_device_t
+	 */
+	core_setup_reservations(dev);
+	/*
+	 * Setup the Asymmetric Logical Unit Assignment for se_device_t
+	 */
+	if (core_setup_alua(dev) < 0)
+		goto out;
+	/*
+	 * Startup the se_device_t processing thread
+	 */
+	if (transport_generic_activate_device(dev) < 0)
+		goto out;
+	/*
+	 * This HBA is reserved for internal storage engine / feature plugin use
+	 */
+	if (hba->hba_flags & HBA_FLAGS_INTERNAL_USE) {
+		ret = transport_get_read_capacity(dev);
+		if (ret < 0)
+			goto out;
+
+		return dev;
+	}
+
+	ret = transport_get_inquiry(DEV_OBJ_API(dev),
+			DEV_T10_WWN(dev), (void *)dev);
+	if (ret < 0)
+		goto out;
+	/*
+	 * Locate VPD WWN Information used for various purposes within
+	 * the Storage Engine.
+	 */
+	if (!(transport_get_inquiry_vpd_serial(DEV_OBJ_API(dev),
+			DEV_T10_WWN(dev), (void *)dev))) {
+		/*
+		 * If VPD Unit Serial returned GOOD status, try
+		 * VPD Device Identification page (0x83).
+		 */
+		transport_get_inquiry_vpd_device_ident(DEV_OBJ_API(dev),
+			DEV_T10_WWN(dev), (void *)dev);
+	}
+
+	/*
+	 * Only perform the volume scan for peripheral type TYPE_DISK
+	 */
+	if (TRANSPORT(dev)->get_device_type(dev) != 0)
+		return dev;
+
+	/*
+	 * Get the sector count via READ_CAPACITY
+	 */
+	ret = transport_get_read_capacity(dev);
+	if (ret < 0)
+		goto out;
+out:
+	if (!ret)
+		return dev;
+
+	/*
+	 * Release claim to OS dependant block_device that may have been
+	 * set by plugin with passed dev_flags.
+	 */
+	transport_generic_release_phydevice(dev, 0);
+
+	/*
+	 * Release newly allocated state for se_device_t
+	 */
+	transport_generic_deactivate_device(dev);
+
+	spin_lock(&hba->device_lock);
+	list_del(&dev->dev_list);
+	hba->dev_count--;
+	spin_unlock(&hba->device_lock);
+
+	se_release_vpd_for_dev(dev);
+
+	kfree(dev->dev_status_queue_obj);
+	kfree(dev->dev_queue_obj);
+	kfree(dev);
+
+	return NULL;
+}
+
+/*	transport_generic_activate_device():
+ *
+ *
+ */
+int transport_generic_activate_device(se_device_t *dev)
+{
+	char name[16];
+
+	if (TRANSPORT(dev)->activate_device)
+		TRANSPORT(dev)->activate_device(dev);
+
+	memset(name, 0, 16);
+	snprintf(name, 16, "LIO_%s", TRANSPORT(dev)->name);
+
+	dev->process_thread = kthread_run(transport_processing_thread,
+			(void *)dev, name);
+	if (IS_ERR(dev->process_thread)) {
+		printk(KERN_ERR "Unable to create kthread: %s\n", name);
+		return -1;
+	}
+
+	wait_for_completion(&dev->dev_queue_obj->thread_create_comp);
+
+	return 0;
+}
+
+/*	transport_generic_deactivate_device():
+ *
+ *
+ */
+void transport_generic_deactivate_device(se_device_t *dev)
+{
+	if (TRANSPORT(dev)->deactivate_device)
+		TRANSPORT(dev)->deactivate_device(dev);
+
+	kthread_stop(dev->process_thread);
+
+	wait_for_completion(&dev->dev_queue_obj->thread_done_comp);
+}
+
+/*	transport_generic_claim_phydevice()
+ *
+ *	Obtain exclusive access to OS dependant block-device via
+ *	Storage Transport Plugin API.
+ *
+ *	In Linux v2.6 this means calling fs/block_dev.c:bd_claim()
+ *	that is called in an plugin dependent method for claiming
+ *	struct block_device.
+ *
+ *	Returns 0 - Already claimed or not able to claim
+ *	Returns 1 - Successfuly claimed
+ *	Returns < 0 - Error
+ */
+int transport_generic_claim_phydevice(se_device_t *dev)
+{
+	int ret;
+	se_hba_t *hba;
+
+	/*
+	 * This function pointer is present when handling access
+	 * control to a OS dependant block subsystem.
+	 */
+	if (!TRANSPORT(dev)->claim_phydevice)
+		return 0;
+
+	if (dev->dev_flags & DF_READ_ONLY)
+		return 0;
+
+	if (dev->dev_flags & DF_CLAIMED_BLOCKDEV)
+		return 0;
+
+	hba = dev->se_hba;
+	if (!(hba)) {
+		printk(KERN_ERR "se_device_t->se_hba is NULL!\n");
+		return -1;
+	}
+
+	ret = TRANSPORT(dev)->claim_phydevice(hba, dev);
+	if (ret < 0)
+		return ret;
+
+	dev->dev_flags |= DF_CLAIMED_BLOCKDEV;
+
+	return 1;
+}
+EXPORT_SYMBOL(transport_generic_claim_phydevice);
+
+/*	transport_generic_release_phydevice():
+ *
+ *	Release exclusive access from OS dependant block-device via
+ *	Storage Transport Plugin API.
+ *
+ *	In Linux v2.6 this means calling fs/block_dev.c:bd_release()
+ *	see iscsi_target_pscsi.c and iscsi_target_iblock.c functions for
+ *	se_subsystem_api_t->[claim,release]_phydevice()
+ */
+void transport_generic_release_phydevice(se_device_t *dev, int check_pscsi)
+{
+	if (!TRANSPORT(dev)->release_phydevice)
+		return;
+
+	if (dev->dev_flags & DF_READ_ONLY) {
+		if (check_pscsi &&
+		   (TRANSPORT(dev)->transport_type !=
+		    TRANSPORT_PLUGIN_PHBA_PDEV))
+				return;
+
+		TRANSPORT(dev)->release_phydevice(dev);
+		return;
+	}
+
+	if (!(dev->dev_flags & DF_CLAIMED_BLOCKDEV))
+		return;
+
+	if (!dev->dev_ptr) {
+		printk(KERN_ERR "se_device_t->dev_ptr is NULL!\n");
+		BUG();
+	}
+
+	if (check_pscsi) {
+		if (TRANSPORT(dev)->transport_type !=
+		    TRANSPORT_PLUGIN_PHBA_PDEV)
+			return;
+
+		if (dev->dev_flags & DF_PERSISTENT_CLAIMED_BLOCKDEV)
+			return;
+	}
+
+	TRANSPORT(dev)->release_phydevice(dev);
+	dev->dev_flags &= ~DF_CLAIMED_BLOCKDEV;
+}
+
+/*	transport_generic_free_device():
+ *
+ *
+ */
+void transport_generic_free_device(se_device_t *dev)
+{
+	if (!(dev->dev_ptr))
+		return;
+
+	transport_generic_deactivate_device(dev);
+
+	transport_generic_release_phydevice(dev, 0);
+
+	if (TRANSPORT(dev)->free_device)
+		TRANSPORT(dev)->free_device(dev->dev_ptr);
+}
+EXPORT_SYMBOL(transport_generic_free_device);
+
+int transport_allocate_iovecs_for_cmd(
+	se_cmd_t *cmd,
+	u32 iov_count)
+{
+	cmd->iov_data = kzalloc(iov_count * sizeof(struct iovec), GFP_ATOMIC);
+	if (!(cmd->iov_data)) {
+		printk(KERN_ERR "Unable to allocate memory for"
+			" iscsi_cmd_t->iov_data.\n");
+		return -1;
+	}
+	cmd->orig_iov_data_count = iov_count;
+
+	return 0;
+}
+EXPORT_SYMBOL(transport_allocate_iovecs_for_cmd);
+
+/*	transport_generic_allocate_iovecs():
+ *
+ *	Called from transport_generic_new_cmd() in Transport Processing Thread.
+ */
+static int transport_generic_allocate_iovecs(
+	se_cmd_t *cmd)
+{
+	u32 iov_count;
+
+	iov_count = T_TASK(cmd)->t_task_se_num;
+	if (!(iov_count))
+		iov_count = 1;
+#if 0
+	printk(KERN_INFO "Allocated %d iovecs for ITT: 0x%08x t_task_se_num:"
+		" %u\n", iov_count, CMD_TFO(cmd)->get_task_tag(cmd),
+		T_TASK(cmd)->t_task_se_num);
+#endif
+	iov_count += TRANSPORT_IOV_DATA_BUFFER;
+
+	if (transport_allocate_iovecs_for_cmd(cmd, iov_count))
+		return -1;
+
+	return 0;
+}
+
+/*	transport_generic_prepare_cdb():
+ *
+ *	Since the Initiator sees iSCSI devices as LUNs,  the SCSI CDB will
+ *	contain the iSCSI LUN in bits 7-5 of byte 1 as per SAM-2.
+ *	The point of this is since we are mapping iSCSI LUNs to
+ *	SCSI Target IDs having a non-zero LUN in the CDB will throw the
+ *	devices and HBAs for a loop.
+ */
+static inline void transport_generic_prepare_cdb(
+	unsigned char *cdb)
+{
+	switch (cdb[0]) {
+	case READ_10: /* SBC - RDProtect */
+	case READ_12: /* SBC - RDProtect */
+	case READ_16: /* SBC - RDProtect */
+	case SEND_DIAGNOSTIC: /* SPC - SELF-TEST Code */
+	case VERIFY: /* SBC - VRProtect */
+	case VERIFY_16: /* SBC - VRProtect */
+	case WRITE_VERIFY: /* SBC - VRProtect */
+	case WRITE_VERIFY_12: /* SBC - VRProtect */
+		break;
+	default:
+		cdb[1] &= 0x1f; /* clear logical unit number */
+		break;
+	}
+}
+
+/*	transport_check_device_cdb_sector_count():
+ *
+ *	returns:
+ *	0 on supported request sector count.
+ *	1 on unsupported request sector count.
+ */
+static inline int transport_check_device_cdb_sector_count(
+	se_obj_lun_type_t *se_obj_api,
+	void *se_obj_ptr,
+	u32 sectors)
+{
+	u32 max_sectors;
+
+	max_sectors = se_obj_api->max_sectors(se_obj_ptr);
+	if (!(max_sectors)) {
+		printk(KERN_ERR "TRANSPORT->get_max_sectors returned zero!\n");
+		return 1;
+	}
+
+	if (sectors > max_sectors)
+		return -1;
+
+	return 0;
+}
+
+/*	transport_generic_get_task():
+ *
+ *
+ */
+static se_task_t *transport_generic_get_task(
+	se_transform_info_t *ti,
+	se_cmd_t *cmd,
+	void *se_obj_ptr,
+	se_obj_lun_type_t *se_obj_api)
+{
+	se_task_t *task;
+	unsigned long flags;
+
+	task = kmem_cache_zalloc(se_task_cache, GFP_KERNEL);
+	if (!(task)) {
+		printk(KERN_ERR "Unable to allocate se_task_t\n");
+		return NULL;
+	}
+
+	INIT_LIST_HEAD(&task->t_list);
+	INIT_LIST_HEAD(&task->t_execute_list);
+	INIT_LIST_HEAD(&task->t_state_list);
+	init_completion(&task->task_stop_comp);
+	task->task_no = T_TASK(cmd)->t_task_no++;
+	task->task_se_cmd = cmd;
+
+	DEBUG_SO("se_obj_ptr: %p\n", se_obj_ptr);
+	DEBUG_SO("se_obj_api: %p\n", se_obj_api);
+	DEBUG_SO("Plugin: %s\n", se_obj_api->obj_plugin->plugin_name);
+
+	task->transport_req = se_obj_api->get_transport_req(se_obj_ptr, task);
+	if (!(task->transport_req))
+		return NULL;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	list_add_tail(&task->t_list, &T_TASK(cmd)->t_task_list);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	task->se_obj_api = se_obj_api;
+	task->se_obj_ptr = se_obj_ptr;
+
+	return task;
+}
+
+int transport_generic_obj_start(
+	se_transform_info_t *ti,
+	se_obj_lun_type_t *obj_api,
+	void *p,
+	unsigned long long starting_lba)
+{
+	ti->ti_lba = starting_lba;
+	ti->ti_obj_api = obj_api;
+	ti->ti_obj_ptr = p;
+
+	return 0;
+}
+
+static int transport_process_data_sg_transform(
+	se_cmd_t *cmd,
+	se_transform_info_t *ti)
+{
+	/*
+	 * Already handled in transport_generic_get_cdb_count()
+	 */
+	return 0;
+}
+
+/*	transport_process_control_sg_transform():
+ *
+ *
+ */
+static int transport_process_control_sg_transform(
+	se_cmd_t *cmd,
+	se_transform_info_t *ti)
+{
+	unsigned char *cdb;
+	se_task_t *task;
+	se_mem_t *se_mem, *se_mem_lout = NULL;
+	int ret;
+	u32 se_mem_cnt = 0, task_offset = 0;
+
+	list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_list, se_list)
+		break;
+
+	if (!se_mem) {
+		printk(KERN_ERR "se_mem is NULL!\n");
+		return -1;
+	}
+
+	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
+				ti->se_obj_api);
+	if (!(task))
+		return -1;
+
+	task->transport_map_task = ti->se_obj_api->get_map_SG(
+			ti->se_obj_ptr, cmd->data_direction);
+
+	cdb = ti->se_obj_api->get_cdb(ti->se_obj_ptr, task);
+	if (cdb)
+		memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+
+	task->task_size = cmd->data_length;
+	task->task_sg_num = 1;
+
+	atomic_inc(&T_TASK(cmd)->t_fe_count);
+	atomic_inc(&T_TASK(cmd)->t_se_count);
+
+	ret = ti->se_obj_api->do_se_mem_map(ti->se_obj_ptr, task,
+			T_TASK(cmd)->t_mem_list, NULL, se_mem, &se_mem_lout,
+			&se_mem_cnt, &task_offset);
+	if (ret < 0)
+		return ret;
+
+	DEBUG_CDB_H("task_no[%u]: SCF_SCSI_CONTROL_SG_IO_CDB task_size: %d\n",
+			task->task_no, task->task_size);
+	return 0;
+}
+
+/*	transport_process_control_nonsg_transform():
+ *
+ *
+ */
+static int transport_process_control_nonsg_transform(
+	se_cmd_t *cmd,
+	se_transform_info_t *ti)
+{
+	unsigned char *cdb;
+	se_task_t *task;
+
+	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
+				ti->se_obj_api);
+	if (!(task))
+		return -1;
+
+	task->transport_map_task = ti->se_obj_api->get_map_non_SG(
+			ti->se_obj_ptr, cmd->data_direction);
+
+	cdb = ti->se_obj_api->get_cdb(ti->se_obj_ptr, task);
+	if (cdb)
+		memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+
+	task->task_size = cmd->data_length;
+	task->task_sg_num = 0;
+
+	atomic_inc(&T_TASK(cmd)->t_fe_count);
+	atomic_inc(&T_TASK(cmd)->t_se_count);
+
+	DEBUG_CDB_H("task_no[%u]: SCF_SCSI_CONTROL_NONSG_IO_CDB task_size:"
+			" %d\n", task->task_no, task->task_size);
+	return 0;
+}
+
+/*	transport_process_non_data_transform():
+ *
+ *
+ */
+static int transport_process_non_data_transform(
+	se_cmd_t *cmd,
+	se_transform_info_t *ti)
+{
+	unsigned char *cdb;
+	se_task_t *task;
+
+	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
+				ti->se_obj_api);
+	if (!(task))
+		return -1;
+
+	task->transport_map_task = ti->se_obj_api->get_map_none(ti->se_obj_ptr);
+
+	cdb = ti->se_obj_api->get_cdb(ti->se_obj_ptr, task);
+	if (cdb)
+		memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+
+	task->task_size = cmd->data_length;
+	task->task_sg_num = 0;
+
+	atomic_inc(&T_TASK(cmd)->t_fe_count);
+	atomic_inc(&T_TASK(cmd)->t_se_count);
+
+	DEBUG_CDB_H("task_no[%u]: SCF_SCSI_NON_DATA_CDB task_size: %d\n",
+			task->task_no, task->task_size);
+	return 0;
+}
+
+static int transport_generic_cmd_sequencer(se_cmd_t *, unsigned char *);
+
+void transport_device_setup_cmd(se_cmd_t *cmd)
+{
+	cmd->transport_add_cmd_to_queue = &transport_add_cmd_to_dev_queue;
+	cmd->se_dev = SE_LUN(cmd)->se_dev;
+}
+
+se_cmd_t *__transport_alloc_se_cmd(
+	struct target_core_fabric_ops *tfo,
+	se_session_t *se_sess,
+	void *fabric_cmd_ptr,
+	u32 data_length,
+	int data_direction,
+	int task_attr)
+{
+	se_cmd_t *cmd;
+
+	if (data_direction == SE_DIRECTION_BIDI) {
+		printk(KERN_ERR "SCSI BiDirectional mode not supported yet\n");
+		return ERR_PTR(-ENOSYS);
+	}
+
+	cmd = kmem_cache_zalloc(se_cmd_cache, GFP_KERNEL);
+	if (!(cmd)) {
+		printk(KERN_ERR "kmem_cache_alloc() failed for se_cmd_cache\n");
+		return ERR_PTR(-ENOMEM);
+	}
+	INIT_LIST_HEAD(&cmd->se_lun_list);
+	INIT_LIST_HEAD(&cmd->se_delayed_list);
+	INIT_LIST_HEAD(&cmd->se_ordered_list);
+
+	cmd->t_task = kzalloc(sizeof(se_transport_task_t), GFP_KERNEL);
+	if (!(cmd->t_task)) {
+		printk(KERN_ERR "Unable to allocate cmd->t_task\n");
+		kmem_cache_free(se_cmd_cache, cmd);
+		return NULL;
+	}
+
+	cmd->sense_buffer = kzalloc(
+			TRANSPORT_SENSE_BUFFER + tfo->get_fabric_sense_len(),
+			GFP_KERNEL);
+	if (!(cmd->sense_buffer)) {
+		printk(KERN_ERR "Unable to allocate memory for"
+			" cmd->sense_buffer\n");
+		kfree(cmd->t_task);
+		kmem_cache_free(se_cmd_cache, cmd);
+		return NULL;
+	}
+	INIT_LIST_HEAD(&T_TASK(cmd)->t_task_list);
+	init_completion(&T_TASK(cmd)->transport_lun_fe_stop_comp);
+	init_completion(&T_TASK(cmd)->transport_lun_stop_comp);
+	init_completion(&T_TASK(cmd)->t_transport_stop_comp);
+	init_completion(&T_TASK(cmd)->t_transport_passthrough_comp);
+	init_completion(&T_TASK(cmd)->t_transport_passthrough_wcomp);
+	spin_lock_init(&T_TASK(cmd)->t_state_lock);
+
+	cmd->se_tfo = tfo;
+	cmd->se_sess = se_sess;
+	cmd->se_fabric_cmd_ptr = fabric_cmd_ptr;
+	cmd->data_length = data_length;
+	cmd->data_direction = data_direction;
+	cmd->sam_task_attr = task_attr;
+
+	return cmd;
+}
+
+int transport_check_alloc_task_attr(se_cmd_t *cmd)
+{
+	/*
+	 * Check if SAM Task Attribute emulation is enabled for this
+	 * se_device_t storage object
+	 */
+	if (SE_DEV(cmd)->dev_task_attr_type != SAM_TASK_ATTR_EMULATED)
+		return 0;
+
+	if (cmd->sam_task_attr == TASK_ATTR_ACA) {
+		DEBUG_STA("SAM Task Attribute ACA"
+			" emulation is not supported\n");
+		return -1;
+	}
+	/*
+	 * Used to determine when ORDERED commands should go from
+	 * Dormant to Active status.
+	 */
+	cmd->se_ordered_id = atomic_inc_return(&SE_DEV(cmd)->dev_ordered_id);
+	smp_mb__after_atomic_inc();
+	DEBUG_STA("Allocated se_ordered_id: %u for Task Attr: 0x%02x on %s\n",
+			cmd->se_ordered_id, cmd->sam_task_attr,
+			TRANSPORT(cmd->se_dev)->name);
+	return 0;
+}
+
+se_cmd_t *transport_alloc_se_cmd(
+	struct target_core_fabric_ops *tfo_api,
+	se_session_t *se_sess,
+	void *fabric_cmd_ptr,
+	u32 data_length,
+	int data_direction,
+	int task_attr)
+{
+	return __transport_alloc_se_cmd(tfo_api, se_sess, fabric_cmd_ptr,
+				data_length, data_direction, task_attr);
+}
+EXPORT_SYMBOL(transport_alloc_se_cmd);
+
+void transport_free_se_cmd(
+	se_cmd_t *se_cmd)
+{
+	if (se_cmd->se_tmr_req)
+		core_tmr_release_req(se_cmd->se_tmr_req);
+
+	kfree(se_cmd->iov_data);
+	kfree(se_cmd->sense_buffer);
+	kfree(se_cmd->t_task);
+	kmem_cache_free(se_cmd_cache, se_cmd);
+}
+EXPORT_SYMBOL(transport_free_se_cmd);
+
+static void transport_generic_wait_for_tasks(se_cmd_t *, int, int);
+
+/*	transport_generic_allocate_tasks():
+ *
+ *	Called from fabric RX Thread.
+ */
+int transport_generic_allocate_tasks(
+	se_cmd_t *cmd,
+	unsigned char *cdb)
+{
+	int non_data_cdb;
+
+	transport_generic_prepare_cdb(cdb);
+
+	/*
+	 * This is needed for early exceptions.
+	 */
+	cmd->transport_wait_for_tasks = &transport_generic_wait_for_tasks;
+
+	CMD_ORIG_OBJ_API(cmd)->transport_setup_cmd(cmd->se_orig_obj_ptr, cmd);
+	/*
+	 * See if this is a CDB which follows SAM, also grab a function
+	 * pointer to see if we need to do extra work.
+	 */
+	non_data_cdb = transport_generic_cmd_sequencer(cmd, cdb);
+	if (non_data_cdb < 0)
+		return -1;
+	/*
+	 * Copy the original CDB into T_TASK(cmd).
+	 */
+	memcpy(T_TASK(cmd)->t_task_cdb, cdb, SCSI_CDB_SIZE);
+	/*
+	 * Check for SAM Task Attribute Emulation
+	 */
+	if (transport_check_alloc_task_attr(cmd) < 0) {
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = INVALID_CDB_FIELD;
+		return -2;
+	}
+#ifdef SNMP_SUPPORT
+	spin_lock(&cmd->se_lun->lun_sep_lock);
+	if (cmd->se_lun->lun_sep)
+		cmd->se_lun->lun_sep->sep_stats.cmd_pdus++;
+	spin_unlock(&cmd->se_lun->lun_sep_lock);
+#endif /* SNMP_SUPPORT */
+
+	switch (non_data_cdb) {
+	case 0:
+		DEBUG_CDB_H("Set cdb[0]: 0x%02x to "
+				"SCF_SCSI_DATA_SG_IO_CDB\n", cdb[0]);
+		cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
+
+		/*
+		 * Get the initial Logical Block Address from the Original
+		 * Command Descriptor Block that arrived on the iSCSI wire.
+		 */
+		T_TASK(cmd)->t_task_lba = (cmd->transport_get_long_lba) ?
+			cmd->transport_get_long_lba(cdb) :
+			cmd->transport_get_lba(cdb);
+
+		break;
+	case 1:
+		DEBUG_CDB_H("Set cdb[0]: 0x%02x to"
+				" SCF_SCSI_CONTROL_SG_IO_CDB\n", cdb[0]);
+		cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
+		cmd->transport_cdb_transform =
+				&transport_process_control_sg_transform;
+		break;
+	case 2:
+		DEBUG_CDB_H("Set cdb[0]: 0x%02x to "
+				"SCF_SCSI_CONTROL_NONSG_IO_CDB\n", cdb[0]);
+		cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+		cmd->transport_cdb_transform =
+				&transport_process_control_nonsg_transform;
+		break;
+	case 3:
+		DEBUG_CDB_H("Set cdb[0]: 0x%02x to "
+				"SCF_SCSI_NON_DATA_CDB\n", cdb[0]);
+		cmd->se_cmd_flags |= SCF_SCSI_NON_DATA_CDB;
+		cmd->transport_cdb_transform =
+				&transport_process_non_data_transform;
+		break;
+	case 4:
+		DEBUG_CDB_H("Set cdb[0]: 0x%02x to"
+				" SCF_SCSI_UNSUPPORTED_CDB\n", cdb[0]);
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+		return -2;
+	case 5:
+		DEBUG_CDB_H("Set cdb[0]: 0x%02x to"
+				" SCF_SCSI_RESERVATION_CONFLICT\n", cdb[0]);
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->se_cmd_flags |= SCF_SCSI_RESERVATION_CONFLICT;
+		cmd->scsi_status = SAM_STAT_RESERVATION_CONFLICT;
+		/*
+		 * For UA Interlock Code 11b, a RESERVATION CONFLICT will
+		 * establish a UNIT ATTENTION with PREVIOUS RESERVATION
+		 * CONFLICT STATUS.
+		 *
+		 * See spc4r17, section 7.4.6 Control Mode Page, Table 349
+		 */
+		if (SE_SESS(cmd) &&
+		    DEV_ATTRIB(cmd->se_dev)->emulate_ua_intlck_ctrl == 2)
+			core_scsi3_ua_allocate(SE_SESS(cmd)->se_node_acl,
+				cmd->orig_fe_lun, 0x2C,
+				ASCQ_2CH_PREVIOUS_RESERVATION_CONFLICT_STATUS);
+		return -2;
+	case 6:
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = INVALID_CDB_FIELD;
+		return -2;
+	case 7:
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = ILLEGAL_REQUEST;
+		return -2;
+	case 8:
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = CHECK_CONDITION_UNIT_ATTENTION;
+		return -2;
+	case 9:
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = CHECK_CONDITION_NOT_READY;
+		return -2;
+	default:
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+		return -2;
+	}
+
+	return 0;
+}
+EXPORT_SYMBOL(transport_generic_allocate_tasks);
+
+/*	transport_generic_handle_cdb():
+ *
+ *
+ */
+int transport_generic_handle_cdb(
+	se_cmd_t *cmd)
+{
+	if (!SE_LUN(cmd)) {
+		printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
+		return -1;
+	}
+
+	cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_NEW_CMD);
+	return 0;
+}
+EXPORT_SYMBOL(transport_generic_handle_cdb);
+
+/*	transport_generic_handle_data():
+ *
+ *
+ */
+int transport_generic_handle_data(
+	se_cmd_t *cmd)
+{
+	/*
+	 * Make sure that the transport has been disabled by
+	 * transport_write_pending() before readding this se_cmd_t to the
+	 * processing queue.  If it has not yet been reset to zero by the
+	 * processing thread in cmd->transport_add_cmd_to_queue(), let other
+	 * processes run.  If a signal was received, then we assume the
+	 * connection is being failed/shutdown, so we return a failure.
+	 */
+	while (atomic_read(&T_TASK(cmd)->t_transport_active)) {
+		msleep_interruptible(10);
+		if (signal_pending(current))
+			return -1;
+	}
+	/*
+	 * If the received CDB has aleady been ABORTED by the generic
+	 * target engine, we now call transport_check_aborted_status()
+	 * to queue any delated TASK_ABORTED status for the received CDB to the
+	 * fabric module as we are expecting no futher incoming DATA OUT
+	 * sequences at this point.
+	 */
+	if (transport_check_aborted_status(cmd, 1) != 0)
+		return 0;
+
+	cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_PROCESS_WRITE);
+	return 0;
+}
+EXPORT_SYMBOL(transport_generic_handle_data);
+
+/*	transport_generic_handle_tmr():
+ *
+ *
+ */
+int transport_generic_handle_tmr(
+	se_cmd_t *cmd)
+{
+	/*
+	 * This is needed for early exceptions.
+	 */
+	cmd->transport_wait_for_tasks = &transport_generic_wait_for_tasks;
+	CMD_ORIG_OBJ_API(cmd)->transport_setup_cmd(cmd->se_orig_obj_ptr, cmd);
+
+	cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_PROCESS_TMR);
+	return 0;
+}
+EXPORT_SYMBOL(transport_generic_handle_tmr);
+
+/*	transport_stop_tasks_for_cmd():
+ *
+ *
+ */
+void transport_stop_tasks_for_cmd(se_cmd_t *cmd)
+{
+	se_task_t *task, *task_tmp;
+	unsigned long flags;
+
+	DEBUG_TS("ITT[0x%08x] - Stopping tasks\n",
+		CMD_TFO(cmd)->get_task_tag(cmd));
+
+	/*
+	 * No tasks remain in the execution queue
+	 */
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	list_for_each_entry_safe(task, task_tmp,
+				&T_TASK(cmd)->t_task_list, t_list) {
+		DEBUG_TS("task_no[%d] - Processing task %p\n",
+				task->task_no, task);
+		/*
+		 * If the se_task_t has not been sent and is not active,
+		 * remove the se_task_t from the execution queue.
+		 */
+		if (!atomic_read(&task->task_sent) &&
+		    !atomic_read(&task->task_active)) {
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+					flags);
+			transport_remove_task_from_execute_queue(task,
+					task->se_dev);
+
+			DEBUG_TS("task_no[%d] - Removed from execute queue\n",
+				task->task_no);
+			spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+			continue;
+		}
+
+		/*
+		 * If the se_task_t is active, sleep until it is returned
+		 * from the plugin.
+		 */
+		if (atomic_read(&task->task_active)) {
+			atomic_set(&task->task_stop, 1);
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+					flags);
+
+			DEBUG_TS("task_no[%d] - Waiting to complete\n",
+				task->task_no);
+			wait_for_completion(&task->task_stop_comp);
+			DEBUG_TS("task_no[%d] - Stopped successfully\n",
+				task->task_no);
+
+			spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+			atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+
+			atomic_set(&task->task_active, 0);
+			atomic_set(&task->task_stop, 0);
+		} else {
+			DEBUG_TS("task_no[%d] - Did nothing\n", task->task_no);
+		}
+
+		__transport_stop_task_timer(task, &flags);
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+static void transport_failure_reset_queue_depth(se_device_t *dev)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);;
+	atomic_inc(&dev->depth_left);
+	atomic_inc(&SE_HBA(dev)->left_queue_depth);
+	spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+}
+
+/*	transport_generic_request_failure():
+ *
+ *	Handle SAM-esque emulation for generic transport request failures.
+ */
+void transport_generic_request_failure(
+	se_cmd_t *cmd,
+	se_device_t *dev,
+	int complete,
+	int sc)
+{
+	DEBUG_GRF("-----[ Storage Engine Exception for cmd: %p ITT: 0x%08x"
+		" CDB: 0x%02x\n", cmd, CMD_TFO(cmd)->get_task_tag(cmd),
+		T_TASK(cmd)->t_task_cdb[0]);
+	DEBUG_GRF("-----[ se_obj_api: %p se_obj_ptr: %p\n", cmd->se_obj_api,
+		cmd->se_obj_ptr);
+	DEBUG_GRF("-----[ se_orig_obj_api: %p se_orig_obj_ptr: %p\n",
+		cmd->se_orig_obj_api, cmd->se_orig_obj_ptr);
+	DEBUG_GRF("-----[ i_state: %d t_state/def_t_state:"
+		" %d/%d transport_error_status: %d\n",
+		CMD_TFO(cmd)->get_cmd_state(cmd),
+		cmd->t_state, cmd->deferred_t_state,
+		cmd->transport_error_status);
+	DEBUG_GRF("-----[ t_task_cdbs: %d t_task_cdbs_left: %d"
+		" t_task_cdbs_sent: %d t_task_cdbs_ex_left: %d --"
+		" t_transport_active: %d t_transport_stop: %d"
+		" t_transport_sent: %d\n", T_TASK(cmd)->t_task_cdbs,
+		atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
+		atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
+		atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left),
+		atomic_read(&T_TASK(cmd)->t_transport_active),
+		atomic_read(&T_TASK(cmd)->t_transport_stop),
+		atomic_read(&T_TASK(cmd)->t_transport_sent));
+
+	transport_stop_all_task_timers(cmd);
+
+	if (dev)
+		transport_failure_reset_queue_depth(dev);
+	/*
+	 * For SAM Task Attribute emulation for failed se_cmd_t
+	 */
+	if (cmd->se_dev->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+		transport_complete_task_attr(cmd);
+
+	if (complete) {
+		transport_direct_request_timeout(cmd);
+		cmd->transport_error_status = PYX_TRANSPORT_LU_COMM_FAILURE;
+	}
+
+	switch (cmd->transport_error_status) {
+	case PYX_TRANSPORT_UNKNOWN_SAM_OPCODE:
+		cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+		break;
+	case PYX_TRANSPORT_REQ_TOO_MANY_SECTORS:
+		cmd->scsi_sense_reason = SECTOR_COUNT_TOO_MANY;
+		break;
+	case PYX_TRANSPORT_INVALID_CDB_FIELD:
+		cmd->scsi_sense_reason = INVALID_CDB_FIELD;
+		break;
+	case PYX_TRANSPORT_INVALID_PARAMETER_LIST:
+		cmd->scsi_sense_reason = INVALID_PARAMETER_LIST;
+		break;
+	case PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES:
+		if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)) {
+			if (!sc)
+				transport_new_cmd_failure(cmd);
+			/*
+			 * Currently for PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES,
+			 * we force this session to fall back to session
+			 * recovery.
+			 */
+			CMD_TFO(cmd)->fall_back_to_erl0(cmd->se_sess);
+			CMD_TFO(cmd)->stop_session(cmd->se_sess, 0, 0);
+
+			goto check_stop;
+		} else {
+			cmd->scsi_sense_reason =
+				LOGICAL_UNIT_COMMUNICATION_FAILURE;
+		}
+		break;
+	case PYX_TRANSPORT_LU_COMM_FAILURE:
+		cmd->scsi_sense_reason = LOGICAL_UNIT_COMMUNICATION_FAILURE;
+		break;
+	case PYX_TRANSPORT_UNKNOWN_MODE_PAGE:
+		cmd->scsi_sense_reason = UNKNOWN_MODE_PAGE;
+		break;
+	case PYX_TRANSPORT_WRITE_PROTECTED:
+		cmd->scsi_sense_reason = WRITE_PROTECTED;
+		break;
+	case PYX_TRANSPORT_RESERVATION_CONFLICT:
+		/*
+		 * No SENSE Data payload for this case, set SCSI Status
+		 * and queue the response to $FABRIC_MOD.
+		 *
+		 * Uses linux/include/scsi/scsi.h SAM status codes defs
+		 */
+		cmd->scsi_status = SAM_STAT_RESERVATION_CONFLICT;
+		/*
+		 * For UA Interlock Code 11b, a RESERVATION CONFLICT will
+		 * establish a UNIT ATTENTION with PREVIOUS RESERVATION
+		 * CONFLICT STATUS.
+		 *
+		 * See spc4r17, section 7.4.6 Control Mode Page, Table 349
+		 */
+		if (SE_SESS(cmd) &&
+		    DEV_ATTRIB(cmd->se_dev)->emulate_ua_intlck_ctrl == 2)
+			core_scsi3_ua_allocate(SE_SESS(cmd)->se_node_acl,
+				cmd->orig_fe_lun, 0x2C,
+				ASCQ_2CH_PREVIOUS_RESERVATION_CONFLICT_STATUS);
+
+		if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+			CMD_TFO(cmd)->queue_status(cmd);
+
+		goto check_stop;
+	default:
+		printk(KERN_ERR "Unknown transport error for CDB 0x%02x: %d\n",
+			T_TASK(cmd)->t_task_cdb[0],
+			cmd->transport_error_status);
+		cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+		break;
+	}
+
+	if (!sc)
+		transport_new_cmd_failure(cmd);
+	else
+		transport_send_check_condition_and_sense(cmd,
+			cmd->scsi_sense_reason, 0);
+check_stop:
+	transport_lun_remove_cmd(cmd);
+	if (!(transport_cmd_check_stop(cmd, 2, 0)))
+		transport_passthrough_check_stop(cmd);
+}
+
+void transport_direct_request_timeout(se_cmd_t *cmd)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (!(atomic_read(&T_TASK(cmd)->t_transport_timeout))) {
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+	if (atomic_read(&T_TASK(cmd)->t_task_cdbs_timeout_left)) {
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+
+	atomic_sub(atomic_read(&T_TASK(cmd)->t_transport_timeout),
+		   &T_TASK(cmd)->t_se_count);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+void transport_generic_request_timeout(se_cmd_t *cmd)
+{
+	unsigned long flags;
+
+	/*
+	 * Reset T_TASK(cmd)->t_se_count to allow transport_generic_remove()
+	 * to allow last call to free memory resources.
+	 */
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (atomic_read(&T_TASK(cmd)->t_transport_timeout) > 1) {
+		int tmp = (atomic_read(&T_TASK(cmd)->t_transport_timeout) - 1);
+
+		atomic_sub(tmp, &T_TASK(cmd)->t_se_count);
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	transport_generic_remove(cmd, 0, 0);
+}
+
+/* #define iscsi_linux_calculate_map_segment_DEBUG */
+#ifdef iscsi_linux_calculate_map_segment_DEBUG
+#define DEBUG_MAP_SEGMENTS(buf...) PYXPRINT(buf)
+#else
+#define DEBUG_MAP_SEGMENTS(buf...)
+#endif
+
+/*	transport_calculate_map_segment():
+ *
+ *
+ */
+static inline void transport_calculate_map_segment(
+	u32 *data_length,
+	se_offset_map_t *lm)
+{
+	u32 sg_offset = 0;
+	se_mem_t *se_mem = lm->map_se_mem;
+
+	DEBUG_MAP_SEGMENTS(" START Mapping se_mem: %p, Length: %d"
+		"  Remaining iSCSI Data: %u\n", se_mem, se_mem->se_len,
+		*data_length);
+	/*
+	 * Still working on pages in the current se_mem_t.
+	 */
+	if (!lm->map_reset) {
+		lm->iovec_length = (lm->sg_length > PAGE_SIZE) ?
+					PAGE_SIZE : lm->sg_length;
+		if (*data_length < lm->iovec_length) {
+			DEBUG_MAP_SEGMENTS("LINUX_MAP: Reset lm->iovec_length"
+				" to %d\n", *data_length);
+
+			lm->iovec_length = *data_length;
+		}
+		lm->iovec_base = page_address(lm->sg_page) + sg_offset;
+
+		DEBUG_MAP_SEGMENTS("LINUX_MAP: Set lm->iovec_base to %p from"
+			" lm->sg_page: %p\n", lm->iovec_base, lm->sg_page);
+		return;
+	}
+
+	/*
+	 * First run of an iscsi_linux_map_t.
+	 *
+	 * OR:
+	 *
+	 * Mapped all of the pages in the current scatterlist, move
+	 * on to the next one.
+	 */
+	lm->map_reset = 0;
+	sg_offset = se_mem->se_off;
+	lm->sg_page = se_mem->se_page;
+	lm->sg_length = se_mem->se_len;
+
+	DEBUG_MAP_SEGMENTS("LINUX_MAP1[%p]: Starting to se_mem->se_len: %u,"
+		" se_mem->se_off: %u, se_mem->se_page: %p\n", se_mem,
+		se_mem->se_len, se_mem->se_off, se_mem->se_page);;
+	/*
+	 * Get the base and length of the current page for use with the iovec.
+	 */
+recalc:
+	lm->iovec_length = (lm->sg_length > (PAGE_SIZE - sg_offset)) ?
+			   (PAGE_SIZE - sg_offset) : lm->sg_length;
+
+	DEBUG_MAP_SEGMENTS("LINUX_MAP: lm->iovec_length: %u, lm->sg_length: %u,"
+		" sg_offset: %u\n", lm->iovec_length, lm->sg_length, sg_offset);
+	/*
+	 * See if there is any iSCSI offset we need to deal with.
+	 */
+	if (!lm->current_offset) {
+		lm->iovec_base = page_address(lm->sg_page) + sg_offset;
+
+		if (*data_length < lm->iovec_length) {
+			DEBUG_MAP_SEGMENTS("LINUX_MAP1[%p]: Reset"
+				" lm->iovec_length to %d\n", se_mem,
+				*data_length);
+			lm->iovec_length = *data_length;
+		}
+
+		DEBUG_MAP_SEGMENTS("LINUX_MAP2[%p]: No current_offset,"
+			" set iovec_base to %p and set Current Page to %p\n",
+			se_mem, lm->iovec_base, lm->sg_page);
+
+		return;
+	}
+
+	/*
+	 * We know the iSCSI offset is in the next page of the current
+	 * scatterlist.  Increase the lm->sg_page pointer and try again.
+	 */
+	if (lm->current_offset >= lm->iovec_length) {
+		DEBUG_MAP_SEGMENTS("LINUX_MAP3[%p]: Next Page:"
+			" lm->current_offset: %u, iovec_length: %u"
+			" sg_offset: %u\n", se_mem, lm->current_offset,
+			lm->iovec_length, sg_offset);
+
+		lm->current_offset -= lm->iovec_length;
+		lm->sg_length -= lm->iovec_length;
+		lm->sg_page++;
+		sg_offset = 0;
+
+		DEBUG_MAP_SEGMENTS("LINUX_MAP3[%p]: ** Skipping to Next Page,"
+			" updated values: lm->current_offset: %u\n", se_mem,
+			lm->current_offset);
+
+		goto recalc;
+	}
+
+	/*
+	 * The iSCSI offset is in the current page, increment the iovec
+	 * base and reduce iovec length.
+	 */
+	lm->iovec_base = page_address(lm->sg_page);
+
+	DEBUG_MAP_SEGMENTS("LINUX_MAP4[%p]: Set lm->iovec_base to %p\n", se_mem,
+			lm->iovec_base);
+
+	lm->iovec_base += sg_offset;
+	lm->iovec_base += lm->current_offset;
+	DEBUG_MAP_SEGMENTS("****** the OLD lm->iovec_length: %u lm->sg_length:"
+		" %u\n", lm->iovec_length, lm->sg_length);
+
+	if ((lm->iovec_length - lm->current_offset) < *data_length)
+		lm->iovec_length -= lm->current_offset;
+	else
+		lm->iovec_length = *data_length;
+
+	if ((lm->sg_length - lm->current_offset) < *data_length)
+		lm->sg_length -= lm->current_offset;
+	else
+		lm->sg_length = *data_length;
+
+	lm->current_offset = 0;
+
+	DEBUG_MAP_SEGMENTS("****** the NEW lm->iovec_length %u lm->sg_length:"
+		" %u\n", lm->iovec_length, lm->sg_length);
+}
+
+/* #define iscsi_linux_get_iscsi_offset_DEBUG */
+#ifdef iscsi_linux_get_iscsi_offset_DEBUG
+#define DEBUG_GET_ISCSI_OFFSET(buf...) PYXPRINT(buf)
+#else
+#define DEBUG_GET_ISCSI_OFFSET(buf...)
+#endif
+
+/*	transport_get_iscsi_offset():
+ *
+ *
+ */
+static int transport_get_iscsi_offset(
+	se_offset_map_t *lmap,
+	se_unmap_sg_t *usg)
+{
+	u32 current_length = 0, current_iscsi_offset = lmap->iscsi_offset;
+	u32 total_offset = 0;
+	se_cmd_t *cmd = usg->se_cmd;
+	se_mem_t *se_mem;
+
+	list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_list, se_list)
+		break;
+
+	if (!se_mem) {
+		printk(KERN_ERR "Unable to locate se_mem from"
+				" T_TASK(cmd)->t_mem_list\n");
+		return -1;
+	}
+
+	/*
+	 * Locate the current offset from the passed iSCSI Offset.
+	 */
+	while (lmap->iscsi_offset != current_length) {
+		/*
+		 * The iSCSI Offset is within the current se_mem_t.
+		 *
+		 * Or:
+		 *
+		 * The iSCSI Offset is outside of the current se_mem_t.
+		 * Recalculate the values and obtain the next se_mem_t pointer.
+		 */
+		total_offset += se_mem->se_len;
+
+		DEBUG_GET_ISCSI_OFFSET("ISCSI_OFFSET: current_length: %u,"
+			" total_offset: %u, sg->length: %u\n",
+			current_length, total_offset, se_mem->se_len);
+
+		if (total_offset > lmap->iscsi_offset) {
+			current_length += current_iscsi_offset;
+			lmap->orig_offset = lmap->current_offset =
+				usg->t_offset = current_iscsi_offset;
+			DEBUG_GET_ISCSI_OFFSET("ISCSI_OFFSET: Within Current"
+				" se_mem_t: %p, current_length incremented to"
+				" %u\n", se_mem, current_length);
+		} else {
+			current_length += se_mem->se_len;
+			current_iscsi_offset -= se_mem->se_len;
+
+			DEBUG_GET_ISCSI_OFFSET("ISCSI_OFFSET: Outside of"
+				" Current se_mem: %p, current_length"
+				" incremented to %u and current_iscsi_offset"
+				" decremented to %u\n", se_mem, current_length,
+				current_iscsi_offset);
+
+			list_for_each_entry_continue(se_mem,
+					T_TASK(cmd)->t_mem_list, se_list)
+				break;
+
+			if (!se_mem) {
+				printk(KERN_ERR "Unable to locate se_mem_t\n");
+				return -1;
+			}
+		}
+	}
+	lmap->map_orig_se_mem = se_mem;
+	usg->cur_se_mem = se_mem;
+
+	return 0;
+}
+
+/* #define iscsi_OS_set_SG_iovec_ptrs_DEBUG */
+#ifdef iscsi_OS_set_SG_iovec_ptrs_DEBUG
+#define DEBUG_IOVEC_SCATTERLISTS(buf...) PYXPRINT(buf)
+
+static void iscsi_check_iovec_map(
+	u32 iovec_count,
+	u32 map_length,
+	se_map_sg_t *map_sg,
+	se_unmap_sg_t *unmap_sg)
+{
+	u32 i, iovec_map_length = 0;
+	se_cmd_t *cmd = map_sg->se_cmd;
+	struct iovec *iov = map_sg->iov;
+	se_mem_t *se_mem;
+
+	for (i = 0; i < iovec_count; i++)
+		iovec_map_length += iov[i].iov_len;
+
+	if (iovec_map_length == map_length)
+		return;
+
+	printk(KERN_INFO "Calculated iovec_map_length: %u does not match passed"
+		" map_length: %u\n", iovec_map_length, map_length);
+	printk(KERN_INFO "ITT: 0x%08x data_length: %u data_direction %d\n",
+		CMD_TFO(cmd)->get_task_tag(cmd), cmd->data_length,
+		cmd->data_direction);
+
+	iovec_map_length = 0;
+
+	for (i = 0; i < iovec_count; i++) {
+		printk(KERN_INFO "iov[%d].iov_[base,len]: %p / %u bytes------"
+			"-->\n", i, iov[i].iov_base, iov[i].iov_len);
+
+		printk(KERN_INFO "iovec_map_length from %u to %u\n",
+			iovec_map_length, iovec_map_length + iov[i].iov_len);
+		iovec_map_length += iov[i].iov_len;
+
+		printk(KERN_INFO "XXXX_map_length from %u to %u\n", map_length,
+				(map_length - iov[i].iov_len));
+		map_length -= iov[i].iov_len;
+	}
+
+	list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_list, se_list) {
+		printk(KERN_INFO "se_mem[%p]: offset: %u length: %u\n",
+			se_mem, se_mem->se_off, se_mem->se_len);
+	}
+
+	BUG();
+}
+
+#else
+#define DEBUG_IOVEC_SCATTERLISTS(buf...)
+#define iscsi_check_iovec_map(a, b, c, d)
+#endif
+
+/*	transport_generic_set_iovec_ptrs():
+ *
+ *
+ */
+static int transport_generic_set_iovec_ptrs(
+	se_map_sg_t *map_sg,
+	se_unmap_sg_t *unmap_sg)
+{
+	u32 i = 0 /* For iovecs */, j = 0 /* For scatterlists */;
+#ifdef iscsi_OS_set_SG_iovec_ptrs_DEBUG
+	u32 orig_map_length = map_sg->data_length;
+#endif
+	se_cmd_t *cmd = map_sg->se_cmd;
+	se_offset_map_t *lmap = &unmap_sg->lmap;
+	struct iovec *iov = map_sg->iov;
+
+	/*
+	 * Used for non scatterlist operations, assume a single iovec.
+	 */
+	if (!T_TASK(cmd)->t_task_se_num) {
+		DEBUG_IOVEC_SCATTERLISTS("ITT: 0x%08x No se_mem_t elements"
+			" present\n", CMD_TFO(cmd)->get_task_tag(cmd));
+		iov[0].iov_base = (unsigned char *) T_TASK(cmd)->t_task_buf +
+							map_sg->data_offset;
+		iov[0].iov_len  = map_sg->data_length;
+		return 1;
+	}
+
+	/*
+	 * Set lmap->map_reset = 1 so the first call to
+	 * transport_calculate_map_segment() sets up the initial
+	 * values for se_offset_map_t.
+	 */
+	lmap->map_reset = 1;
+
+	DEBUG_IOVEC_SCATTERLISTS("[-------------------] ITT: 0x%08x OS"
+		" Independent Network POSIX defined iovectors to SE Memory"
+		" [-------------------]\n\n", CMD_TFO(cmd)->get_task_tag(cmd));
+
+	/*
+	 * Get a pointer to the first used scatterlist based on the passed
+	 * offset. Also set the rest of the needed values in iscsi_linux_map_t.
+	 */
+	lmap->iscsi_offset = map_sg->data_offset;
+	if (map_sg->map_flags & MAP_SG_KMAP) {
+		unmap_sg->se_cmd = map_sg->se_cmd;
+		transport_get_iscsi_offset(lmap, unmap_sg);
+		unmap_sg->data_length = map_sg->data_length;
+	} else {
+		lmap->current_offset = lmap->orig_offset;
+	}
+	lmap->map_se_mem = lmap->map_orig_se_mem;
+
+	DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: Total map_sg->data_length: %d,"
+		" lmap->iscsi_offset: %d, cmd->orig_iov_data_count: %d\n",
+		map_sg->data_length, lmap->iscsi_offset,
+		cmd->orig_iov_data_count);
+
+	while (map_sg->data_length) {
+		/*
+		 * Time to get the virtual address for use with iovec pointers.
+		 * This function will return the expected iovec_base address
+		 * and iovec_length.
+		 */
+		transport_calculate_map_segment(&map_sg->data_length, lmap);
+
+		/*
+		 * Set the iov.iov_base and iov.iov_len from the current values
+		 * in iscsi_linux_map_t.
+		 */
+		iov[i].iov_base = lmap->iovec_base;
+		iov[i].iov_len = lmap->iovec_length;
+
+		/*
+		 * Subtract the final iovec length from the total length to be
+		 * mapped, and the length of the current scatterlist.  Also
+		 * perform the paranoid check to make sure we are not going to
+		 * overflow the iovecs allocated for this command in the next
+		 * pass.
+		 */
+		map_sg->data_length -= iov[i].iov_len;
+		lmap->sg_length -= iov[i].iov_len;
+
+		DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: iov[%u].iov_len: %u\n",
+				i, iov[i].iov_len);
+		DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: lmap->sg_length: from %u"
+			" to %u\n", lmap->sg_length + iov[i].iov_len,
+				lmap->sg_length);
+		DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: Changed total"
+			" map_sg->data_length from %u to %u\n",
+			map_sg->data_length + iov[i].iov_len,
+			map_sg->data_length);
+
+		if ((++i + 1) > cmd->orig_iov_data_count) {
+			printk(KERN_ERR "Current iovec count %u is greater than"
+				" se_cmd_t->orig_data_iov_count %u, cannot"
+				" continue.\n", i+1, cmd->orig_iov_data_count);
+			return -1;
+		}
+
+		/*
+		 * All done mapping this scatterlist's pages, move on to
+		 * the next scatterlist by setting lmap.map_reset = 1;
+		 */
+		if (!lmap->sg_length || !map_sg->data_length) {
+			list_for_each_entry(lmap->map_se_mem,
+					&lmap->map_se_mem->se_list, se_list)
+				break;
+
+			if (!lmap->map_se_mem) {
+				printk(KERN_ERR "Unable to locate next"
+					" lmap->map_se_mem_t entry\n");
+				return -1;
+			}
+			j++;
+
+			lmap->sg_page = NULL;
+			lmap->map_reset = 1;
+
+			DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: Done with current"
+				" scatterlist, incremented Generic scatterlist"
+				" Counter to %d and reset = 1\n", j);
+		} else
+			lmap->sg_page++;
+	}
+
+	unmap_sg->sg_count = j;
+
+	iscsi_check_iovec_map(i, orig_map_length, map_sg, unmap_sg);
+
+	return i;
+}
+
+/*	transport_generic_allocate_buf():
+ *
+ *	Called from transport_generic_new_cmd() in Transport Processing Thread.
+ */
+int transport_generic_allocate_buf(
+	se_cmd_t *cmd,
+	u32 data_length,
+	u32 dma_size)
+{
+	unsigned char *buf;
+
+	buf = kzalloc(data_length, GFP_KERNEL);
+	if (!(buf)) {
+		printk(KERN_ERR "Unable to allocate memory for buffer\n");
+		return -1;
+	}
+
+	T_TASK(cmd)->t_task_se_num = 0;
+	T_TASK(cmd)->t_task_buf = buf;
+
+	return 0;
+}
+
+/*	transport_generic_allocate_none():
+ *
+ *
+ */
+static int transport_generic_allocate_none(
+	se_cmd_t *cmd,
+	u32 data_length,
+	u32 dma_size)
+{
+	return 0;
+}
+
+/*	transport_generic_map_SG_segments():
+ *
+ *
+ */
+static void transport_generic_map_SG_segments(se_unmap_sg_t *unmap_sg)
+{
+	u32 i = 0;
+	se_cmd_t *cmd = unmap_sg->se_cmd;
+	se_mem_t *se_mem = unmap_sg->cur_se_mem;
+
+	if (!(T_TASK(cmd)->t_task_se_num))
+		return;
+
+	list_for_each_entry_continue(se_mem, T_TASK(cmd)->t_mem_list, se_list) {
+		kmap(se_mem->se_page);
+
+		if (++i == unmap_sg->sg_count)
+			break;
+	}
+}
+
+/*	transport_generic_unmap_SG_segments():
+ *
+ *
+ */
+static void transport_generic_unmap_SG_segments(se_unmap_sg_t *unmap_sg)
+{
+	u32 i = 0;
+	se_cmd_t *cmd = unmap_sg->se_cmd;
+	se_mem_t *se_mem = unmap_sg->cur_se_mem;
+
+	if (!(T_TASK(cmd)->t_task_se_num))
+		return;
+
+	list_for_each_entry_continue(se_mem, T_TASK(cmd)->t_mem_list, se_list) {
+		kunmap(se_mem->se_page);
+
+		if (++i == unmap_sg->sg_count)
+			break;
+	}
+
+	return;
+}
+
+static inline u32 transport_lba_21(unsigned char *cdb)
+{
+	return ((cdb[1] & 0x1f) << 16) | (cdb[2] << 8) | cdb[3];
+}
+
+static inline u32 transport_lba_32(unsigned char *cdb)
+{
+	return (cdb[2] << 24) | (cdb[3] << 16) | (cdb[4] << 8) | cdb[5];
+}
+
+static inline unsigned long long transport_lba_64(unsigned char *cdb)
+{
+	unsigned int __v1, __v2;
+
+	__v1 = (cdb[2] << 24) | (cdb[3] << 16) | (cdb[4] << 8) | cdb[5];
+	__v2 = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
+
+	return ((unsigned long long)__v2) | (unsigned long long)__v1 << 32;
+}
+
+/*	transport_set_supported_SAM_opcode():
+ *
+ *
+ */
+void transport_set_supported_SAM_opcode(se_cmd_t *se_cmd)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(se_cmd)->t_state_lock, flags);
+	se_cmd->se_cmd_flags |= SCF_SUPPORTED_SAM_OPCODE;
+	spin_unlock_irqrestore(&T_TASK(se_cmd)->t_state_lock, flags);
+}
+
+/*
+ * Called from interrupt context.
+ */
+void transport_task_timeout_handler(unsigned long data)
+{
+	se_task_t *task = (se_task_t *)data;
+	se_cmd_t *cmd = TASK_CMD(task);
+	unsigned long flags;
+
+	DEBUG_TT("transport task timeout fired! task: %p cmd: %p\n", task, cmd);
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (task->task_flags & TF_STOP) {
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+	task->task_flags &= ~TF_RUNNING;
+
+	/*
+	 * Determine if transport_complete_task() has already been called.
+	 */
+	if (!(atomic_read(&task->task_active))) {
+		DEBUG_TT("transport task: %p cmd: %p timeout task_active"
+				" == 0\n", task, cmd);
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+
+	atomic_inc(&T_TASK(cmd)->t_se_count);
+	atomic_inc(&T_TASK(cmd)->t_transport_timeout);
+	T_TASK(cmd)->t_tasks_failed = 1;
+
+	atomic_set(&task->task_timeout, 1);
+	task->task_error_status = PYX_TRANSPORT_TASK_TIMEOUT;
+	task->task_scsi_status = 1;
+
+	if (atomic_read(&task->task_stop)) {
+		DEBUG_TT("transport task: %p cmd: %p timeout task_stop"
+				" == 1\n", task, cmd);
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		complete(&task->task_stop_comp);
+		return;
+	}
+
+	if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_left))) {
+		DEBUG_TT("transport task: %p cmd: %p timeout non zero"
+				" t_task_cdbs_left\n", task, cmd);
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+	DEBUG_TT("transport task: %p cmd: %p timeout ZERO t_task_cdbs_left\n",
+			task, cmd);
+
+	cmd->t_state = TRANSPORT_COMPLETE_FAILURE;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_COMPLETE_FAILURE);
+}
+
+u32 transport_get_default_task_timeout(se_device_t *dev)
+{
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_DISK)
+		return TRANSPORT_TIMEOUT_TYPE_DISK;
+
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_ROM)
+		return TRANSPORT_TIMEOUT_TYPE_ROM;
+
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+		return TRANSPORT_TIMEOUT_TYPE_TAPE;
+
+	return TRANSPORT_TIMEOUT_TYPE_OTHER;
+}
+EXPORT_SYMBOL(transport_get_default_task_timeout);
+
+/*
+ * Called with T_TASK(cmd)->t_state_lock held.
+ */
+void transport_start_task_timer(se_task_t *task)
+{
+	int timeout;
+
+	if (task->task_flags & TF_RUNNING)
+		return;
+	/*
+	 * If the task_timeout is disabled, exit now.
+	 */
+	timeout = task->se_obj_api->get_task_timeout(task->se_obj_ptr);
+	if (!(timeout))
+		return;
+
+	init_timer(&task->task_timer);
+	task->task_timer.expires = (get_jiffies_64() + timeout * HZ);
+	task->task_timer.data = (unsigned long) task;
+	task->task_timer.function = transport_task_timeout_handler;
+
+	task->task_flags |= TF_RUNNING;
+	add_timer(&task->task_timer);
+#if 0
+	printk(KERN_INFO "Starting task timer for cmd: %p task: %p seconds:"
+		" %d\n", task->task_se_cmd, task, timeout);
+#endif
+}
+
+/*
+ * Called with spin_lock_irq(&T_TASK(cmd)->t_state_lock) held.
+ */
+void __transport_stop_task_timer(se_task_t *task, unsigned long *flags)
+{
+	se_cmd_t *cmd = TASK_CMD(task);
+
+	if (!(task->task_flags & TF_RUNNING))
+		return;
+
+	task->task_flags |= TF_STOP;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, *flags);
+
+	del_timer_sync(&task->task_timer);
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, *flags);
+	task->task_flags &= ~TF_RUNNING;
+	task->task_flags &= ~TF_STOP;
+}
+
+void transport_stop_task_timer(se_task_t *task)
+{
+	se_cmd_t *cmd = TASK_CMD(task);
+	unsigned long flags;
+#if 0
+	printk(KERN_INFO "Stopping task timer for cmd: %p task: %p\n",
+			cmd, task);
+#endif
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (!(task->task_flags & TF_RUNNING)) {
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return;
+	}
+	task->task_flags |= TF_STOP;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	del_timer_sync(&task->task_timer);
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	task->task_flags &= ~TF_RUNNING;
+	task->task_flags &= ~TF_STOP;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+void transport_stop_all_task_timers(se_cmd_t *cmd)
+{
+	se_task_t *task = NULL, *task_tmp;
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	list_for_each_entry_safe(task, task_tmp,
+				&T_TASK(cmd)->t_task_list, t_list)
+		__transport_stop_task_timer(task, &flags);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+static inline int transport_tcq_window_closed(se_device_t *dev)
+{
+	if (dev->dev_tcq_window_closed++ <
+			PYX_TRANSPORT_WINDOW_CLOSED_THRESHOLD) {
+		msleep(PYX_TRANSPORT_WINDOW_CLOSED_WAIT_SHORT);
+	} else
+		msleep(PYX_TRANSPORT_WINDOW_CLOSED_WAIT_LONG);
+
+	wake_up_interruptible(&dev->dev_queue_obj->thread_wq);
+	return 0;
+}
+
+/*
+ * Called from Fabric Module context from transport_execute_tasks()
+ *
+ * The return of this function determins if the tasks from se_cmd_t
+ * get added to the execution queue in transport_execute_tasks(),
+ * or are added to the delayed or ordered lists here.
+ */
+static inline int transport_execute_task_attr(se_cmd_t *cmd)
+{
+	if (SE_DEV(cmd)->dev_task_attr_type != SAM_TASK_ATTR_EMULATED)
+		return 1;
+	/*
+	 * Check for the existance of HEAD_OF_QUEUE, and if true return 1
+	 * to allow the passed se_cmd_t list of tasks to the front of the list.
+	 */
+	 if (cmd->sam_task_attr == TASK_ATTR_HOQ) {
+		atomic_inc(&SE_DEV(cmd)->dev_hoq_count);
+		smp_mb__after_atomic_inc();
+		DEBUG_STA("Added HEAD_OF_QUEUE for CDB:"
+			" 0x%02x, se_ordered_id: %u\n",
+			T_TASK(cmd)->t_task_cdb[0],
+			cmd->se_ordered_id);
+		return 1;
+	} else if (cmd->sam_task_attr == TASK_ATTR_ORDERED) {
+		spin_lock(&SE_DEV(cmd)->ordered_cmd_lock);
+		list_add_tail(&cmd->se_ordered_list,
+				&SE_DEV(cmd)->ordered_cmd_list);
+		spin_unlock(&SE_DEV(cmd)->ordered_cmd_lock);
+
+		atomic_inc(&SE_DEV(cmd)->dev_ordered_sync);
+		smp_mb__after_atomic_inc();
+
+		DEBUG_STA("Added ORDERED for CDB: 0x%02x to ordered"
+				" list, se_ordered_id: %u\n",
+				T_TASK(cmd)->t_task_cdb[0],
+				cmd->se_ordered_id);
+		/*
+		 * Add ORDERED command to tail of execution queue if
+		 * no other older commands exist that need to be
+		 * completed first.
+		 */
+		if (!(atomic_read(&SE_DEV(cmd)->simple_cmds)))
+			return 1;
+	} else {
+		/*
+		 * For SIMPLE and UNTAGGED Task Attribute commands
+		 */
+		atomic_inc(&SE_DEV(cmd)->simple_cmds);
+		smp_mb__after_atomic_inc();
+	}
+	/*
+	 * Otherwise if one or more outstanding ORDERED task attribute exist,
+	 * add the dormant task(s) built for the passed se_cmd_t to the
+	 * execution queue and become in Active state for this se_device_t.
+	 */
+	if (atomic_read(&SE_DEV(cmd)->dev_ordered_sync) != 0) {
+		/*
+		 * Otherwise, add cmd w/ tasks to delayed cmd queue that
+		 * will be drained upon competion of HEAD_OF_QUEUE task.
+		 */
+		spin_lock(&SE_DEV(cmd)->delayed_cmd_lock);
+		cmd->se_cmd_flags |= SCF_DELAYED_CMD_FROM_SAM_ATTR;
+		list_add_tail(&cmd->se_delayed_list,
+				&SE_DEV(cmd)->delayed_cmd_list);
+		spin_unlock(&SE_DEV(cmd)->delayed_cmd_lock);
+
+		DEBUG_STA("Added CDB: 0x%02x Task Attr: 0x%02x to"
+			" delayed CMD list, se_ordered_id: %u\n",
+			T_TASK(cmd)->t_task_cdb[0], cmd->sam_task_attr,
+			cmd->se_ordered_id);
+		/*
+		 * Return zero to let transport_execute_tasks() know
+		 * not to add the delayed tasks to the execution list.
+		 */
+		return 0;
+	}
+	/*
+	 * Otherwise, no ORDERED task attributes exist..
+	 */
+	return 1;
+}
+
+/*
+ * Called from fabric module context in transport_generic_new_cmd() and
+ * transport_generic_process_write()
+ */
+int transport_execute_tasks(se_cmd_t *cmd)
+{
+	int add_tasks;
+
+	if (!(cmd->se_cmd_flags & SCF_SE_DISABLE_ONLINE_CHECK)) {
+		if (CMD_ORIG_OBJ_API(cmd)->check_online(
+					cmd->se_orig_obj_ptr) != 0) {
+			cmd->transport_error_status =
+				PYX_TRANSPORT_LU_COMM_FAILURE;
+			transport_generic_request_failure(cmd, NULL, 0, 1);
+			return 0;
+		}
+	}
+	/*
+	 * Call transport_cmd_check_stop() to see if a fabric exception
+	 * has occured that prevents execution.
+	 */
+	if (!(transport_cmd_check_stop(cmd, 0, TRANSPORT_PROCESSING))) {
+		/*
+		 * Check for SAM Task Attribute emulation and HEAD_OF_QUEUE
+		 * attribute for the tasks of the received se_cmd_t CDB
+		 */
+		add_tasks = transport_execute_task_attr(cmd);
+		if (add_tasks == 0)
+			goto execute_tasks;
+		/*
+		 * This calls transport_add_tasks_from_cmd() to handle
+		 * HEAD_OF_QUEUE ordering for SAM Task Attribute emulation
+		 * (if enabled) in __transport_add_task_to_execute_queue() and
+		 * transport_add_task_check_sam_attr().
+		 */
+		CMD_ORIG_OBJ_API(cmd)->add_tasks(cmd->se_orig_obj_ptr, cmd);
+	}
+	/*
+	 * Kick the execution queue for the cmd associated se_device_t
+	 * storage object.
+	 */
+execute_tasks:
+	CMD_ORIG_OBJ_API(cmd)->execute_tasks(cmd->se_orig_obj_ptr);
+	return 0;
+}
+
+/*
+ * Called to check se_device_t tcq depth window, and once open pull se_task_t
+ * from se_device_t->execute_task_list and
+ *
+ * Called from transport_processing_thread()
+ */
+int __transport_execute_tasks(se_device_t *dev)
+{
+	int error;
+	se_cmd_t *cmd = NULL;
+	se_task_t *task;
+	unsigned long flags;
+
+	/*
+	 * Check if there is enough room in the device and HBA queue to send
+	 * se_transport_task_t's to the selected transport.
+	 */
+check_depth:
+	spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
+	if (!(atomic_read(&dev->depth_left)) ||
+	    !(atomic_read(&SE_HBA(dev)->left_queue_depth))) {
+		spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+		return transport_tcq_window_closed(dev);
+	}
+	dev->dev_tcq_window_closed = 0;
+
+	spin_lock(&dev->execute_task_lock);
+	task = transport_get_task_from_execute_queue(dev);
+	spin_unlock(&dev->execute_task_lock);
+
+	if (!task) {
+		spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+		return 0;
+	}
+
+	atomic_dec(&dev->depth_left);
+	atomic_dec(&SE_HBA(dev)->left_queue_depth);
+	spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+
+	cmd = TASK_CMD(task);
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	atomic_set(&task->task_active, 1);
+	atomic_set(&task->task_sent, 1);
+	atomic_inc(&T_TASK(cmd)->t_task_cdbs_sent);
+
+	if (atomic_read(&T_TASK(cmd)->t_task_cdbs_sent) ==
+	    T_TASK(cmd)->t_task_cdbs)
+		atomic_set(&cmd->transport_sent, 1);
+
+	transport_start_task_timer(task);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+	/*
+	 * The se_cmd_t->transport_emulate_cdb() function pointer is used
+	 * to grab REPORT_LUNS CDBs before they hit the
+	 * se_subsystem_api_t->do_task() caller below.
+	 */
+	if (cmd->transport_emulate_cdb) {
+		error = cmd->transport_emulate_cdb(cmd);
+		if (error != 0) {
+			cmd->transport_error_status = error;
+			atomic_set(&task->task_active, 0);
+			atomic_set(&cmd->transport_sent, 0);
+			transport_stop_tasks_for_cmd(cmd);
+			transport_generic_request_failure(cmd, dev, 0, 1);
+			goto check_depth;
+		}
+		/*
+		 * Handle the successful completion for transport_emulate_cdb()
+		 * usage.
+		 */
+		cmd->scsi_status = SAM_STAT_GOOD;
+		task->task_scsi_status = GOOD;
+		transport_complete_task(task, 1);
+	} else {
+		error = TRANSPORT(dev)->do_task(task);
+		if (error != 0) {
+			cmd->transport_error_status = error;
+			atomic_set(&task->task_active, 0);
+			atomic_set(&cmd->transport_sent, 0);
+			transport_stop_tasks_for_cmd(cmd);
+			transport_generic_request_failure(cmd, dev, 0, 1);
+		}
+	}
+
+	goto check_depth;
+
+	return 0;
+}
+
+/*	transport_new_cmd_failure():
+ *
+ *
+ */
+void transport_new_cmd_failure(se_cmd_t *se_cmd)
+{
+	unsigned long flags;
+	/*
+	 * Any unsolicited data will get dumped for failed command inside of
+	 * the fabric plugin
+	 */
+	spin_lock_irqsave(&T_TASK(se_cmd)->t_state_lock, flags);
+	se_cmd->se_cmd_flags |= SCF_SE_CMD_FAILED;
+	se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+	spin_unlock_irqrestore(&T_TASK(se_cmd)->t_state_lock, flags);
+
+	CMD_TFO(se_cmd)->new_cmd_failure(se_cmd);
+}
+
+static int transport_generic_map_buffers_to_tasks(se_cmd_t *);
+static void transport_nop_wait_for_tasks(se_cmd_t *, int, int);
+
+static inline u32 transport_get_sectors_6(
+	unsigned char *cdb,
+	se_cmd_t *cmd,
+	int *ret)
+{
+	se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+	/*
+	 * Assume TYPE_DISK for non se_device_t objects.
+	 * Use 8-bit sector value.
+	 */
+	if (!dev)
+		goto type_disk;
+
+	/*
+	 * Use 24-bit allocation length for TYPE_TAPE.
+	 */
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+		return (u32)(cdb[2] << 16) + (cdb[3] << 8) + cdb[4];
+
+	/*
+	 * Everything else assume TYPE_DISK Sector CDB location.
+	 * Use 8-bit sector value.
+	 */
+type_disk:
+	return (u32)cdb[4];
+}
+
+static inline u32 transport_get_sectors_10(
+	unsigned char *cdb,
+	se_cmd_t *cmd,
+	int *ret)
+{
+	se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+	/*
+	 * Assume TYPE_DISK for non se_device_t objects.
+	 * Use 16-bit sector value.
+	 */
+	if (!dev)
+		goto type_disk;
+
+	/*
+	 * XXX_10 is not defined in SSC, throw an exception
+	 */
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
+		*ret = -1;
+		return 0;
+	}
+
+	/*
+	 * Everything else assume TYPE_DISK Sector CDB location.
+	 * Use 16-bit sector value.
+	 */
+type_disk:
+	return (u32)(cdb[7] << 8) + cdb[8];
+}
+
+static inline u32 transport_get_sectors_12(
+	unsigned char *cdb,
+	se_cmd_t *cmd,
+	int *ret)
+{
+	se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+	/*
+	 * Assume TYPE_DISK for non se_device_t objects.
+	 * Use 32-bit sector value.
+	 */
+	if (!dev)
+		goto type_disk;
+
+	/*
+	 * XXX_12 is not defined in SSC, throw an exception
+	 */
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
+		*ret = -1;
+		return 0;
+	}
+
+	/*
+	 * Everything else assume TYPE_DISK Sector CDB location.
+	 * Use 32-bit sector value.
+	 */
+type_disk:
+	return (u32)(cdb[6] << 24) + (cdb[7] << 16) + (cdb[8] << 8) + cdb[9];
+}
+
+static inline u32 transport_get_sectors_16(
+	unsigned char *cdb,
+	se_cmd_t *cmd,
+	int *ret)
+{
+	se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+	/*
+	 * Assume TYPE_DISK for non se_device_t objects.
+	 * Use 32-bit sector value.
+	 */
+	if (!dev)
+		goto type_disk;
+
+	/*
+	 * Use 24-bit allocation length for TYPE_TAPE.
+	 */
+	if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+		return (u32)(cdb[12] << 16) + (cdb[13] << 8) + cdb[14];
+
+type_disk:
+	return (u32)(cdb[10] << 24) + (cdb[11] << 16) +
+		    (cdb[12] << 8) + cdb[13];
+}
+
+static inline u32 transport_get_size(
+	u32 sectors,
+	unsigned char *cdb,
+	se_cmd_t *cmd)
+{
+	return CMD_ORIG_OBJ_API(cmd)->get_cdb_size(cmd->se_orig_obj_ptr,
+		sectors, cdb);
+}
+
+static inline void transport_get_maps(se_cmd_t *cmd)
+{
+	cmd->transport_map_SG_segments = &transport_generic_map_SG_segments;
+	cmd->transport_unmap_SG_segments = &transport_generic_unmap_SG_segments;
+}
+
+static unsigned char asciihex_to_binaryhex(unsigned char val[2])
+{
+	unsigned char result = 0;
+	/*
+	 * MSB
+	 */
+	if ((val[0] >= 'a') && (val[0] <= 'f'))
+		result = ((val[0] - 'a' + 10) & 0xf) << 4;
+	else
+		if ((val[0] >= 'A') && (val[0] <= 'F'))
+			result = ((val[0] - 'A' + 10) & 0xf) << 4;
+		else /* digit */
+			result = ((val[0] - '0') & 0xf) << 4;
+	/*
+	 * LSB
+	 */
+	if ((val[1] >= 'a') && (val[1] <= 'f'))
+		result |= ((val[1] - 'a' + 10) & 0xf);
+	else
+		if ((val[1] >= 'A') && (val[1] <= 'F'))
+			result |= ((val[1] - 'A' + 10) & 0xf);
+		else /* digit */
+			result |= ((val[1] - '0') & 0xf);
+
+	return result;
+}
+
+extern int transport_generic_emulate_inquiry(
+	se_cmd_t *cmd,
+	unsigned char type,
+	unsigned char *prod,
+	unsigned char *version,
+	unsigned char *se_location)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	se_lun_t *lun = SE_LUN(cmd);
+	se_port_t *port = NULL;
+	se_portal_group_t *tpg = NULL;
+	t10_alua_lu_gp_member_t *lu_gp_mem;
+	t10_alua_tg_pt_gp_t *tg_pt_gp;
+	t10_alua_tg_pt_gp_member_t *tg_pt_gp_mem;
+	unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+	unsigned char *cdb = T_TASK(cmd)->t_task_cdb;
+	unsigned char *iqn_sn, binary, binary_new;
+	u32 prod_len, iqn_sn_len, se_location_len;
+	u32 unit_serial_len, off = 0;
+	int i;
+	u16 len = 0, id_len;
+
+	/*
+	 * Make sure we at least have 8 bytes of INQUIRY response payload
+	 * going back.
+	 */
+	if (cmd->data_length < 8) {
+		printk(KERN_ERR "SCSI Inquiry payload length: %u too small\n",
+				cmd->data_length);
+		return -1;
+	}
+	buf[0] = type;
+
+	if (!(cdb[1] & 0x1)) {
+		if (type == TYPE_TAPE)
+			buf[1] = 0x80;
+		buf[2]          = TRANSPORT(dev)->get_device_rev(dev);
+		/*
+		 * Enable SCCS and TPGS fields for Emulated ALUA
+		 */
+		if (T10_ALUA(dev->se_sub_dev)->alua_type ==
+				SPC3_ALUA_EMULATED) {
+			/*
+			 * Set SCCS for MAINTENANCE_IN +
+			 * REPORT_TARGET_PORT_GROUPS
+			 */
+			buf[5]	= 0x80;
+			/*
+			 * Set TPGS field for explict and/or implict ALUA
+			 * access type and opteration.
+			 *
+			 * See spc4r17 section 6.4.2 Table 135
+			 */
+			port = lun->lun_sep;
+			if (!(port))
+				goto after_tpgs;
+			tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
+			if (!(tg_pt_gp_mem))
+				goto after_tpgs;
+
+			spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+			tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
+			if (!(tg_pt_gp)) {
+				spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+				goto after_tpgs;
+			}
+			buf[5] |= tg_pt_gp->tg_pt_gp_alua_access_type;
+			spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+		}
+after_tpgs:
+		buf[7]		= 0x32; /* Sync=1 and CmdQue=1 */
+		/*
+		 * Do not include vendor, product, reversion info in INQUIRY
+		 * response payload for cdbs with a small allocation length.
+		 */
+		if (cmd->data_length < 36) {
+			buf[4] = 3; /* Set additional length to 3 */
+			return 0;
+		}
+
+		snprintf((unsigned char *)&buf[8], 8, "LIO-ORG");
+		snprintf((unsigned char *)&buf[16], 16, "%s", prod);
+		snprintf((unsigned char *)&buf[32], 4, "%s", version);
+		buf[4] = 31; /* Set additional length to 31 */
+		return 0;
+	}
+
+	switch (cdb[2]) {
+	case 0x00: /* supported vital product data pages */
+		buf[1] = 0x00;
+		buf[3] = 3;
+		buf[4] = 0x0;
+		buf[5] = 0x80;
+		buf[6] = 0x83;
+		len = 3;
+		break;
+	case 0x80: /* unit serial number */
+		buf[1] = 0x80;
+		if (dev->se_sub_dev->su_dev_flags &
+					SDF_EMULATED_VPD_UNIT_SERIAL) {
+			unit_serial_len =
+				strlen(&DEV_T10_WWN(dev)->unit_serial[0]);
+			unit_serial_len++; /* For NULL Terminator */
+
+			if (((len + 4) + unit_serial_len) > cmd->data_length) {
+				len += unit_serial_len;
+				goto set_len;
+			}
+			len += sprintf((unsigned char *)&buf[4], "%s",
+				&DEV_T10_WWN(dev)->unit_serial[0]);
+		 } else {
+			iqn_sn = transport_get_iqn_sn();
+			iqn_sn_len = strlen(iqn_sn);
+			iqn_sn_len++; /* For ":" */
+			se_location_len = strlen(se_location);
+			se_location_len++; /* For NULL Terminator */
+
+			if (((len + 4) + (iqn_sn_len + se_location_len)) >
+					cmd->data_length) {
+				len += (iqn_sn_len + se_location_len);
+				goto set_len;
+			}
+			len += sprintf((unsigned char *)&buf[4], "%s:%s",
+				iqn_sn, se_location);
+		}
+		len++; /* Extra Byte for NULL Terminator */
+		buf[3] = len;
+		break;
+	case 0x83:
+		/*
+		 * Device identification VPD, for a complete list of
+		 * DESIGNATOR TYPEs see spc4r17 Table 459.
+		 */
+		buf[1] = 0x83;
+		off = 4;
+		/*
+		 * NAA IEEE Registered Extended Assigned designator format,
+		 * see spc4r17 section 7.7.3.6.5
+		 *
+		 * We depend upon a target_core_mod/ConfigFS provided
+		 * /sys/kernel/config/target/core/$HBA/$DEV/wwn/vpd_unit_serial
+		 * value in order to return the NAA id.
+		 */
+		if (!(dev->se_sub_dev->su_dev_flags &
+					SDF_EMULATED_VPD_UNIT_SERIAL))
+			goto check_t10_vend_desc;
+		if ((off + 20) > cmd->data_length)
+			goto check_t10_vend_desc;
+		/* CODE SET == Binary */
+		buf[off++] = 0x1;
+		/* Set ASSOICATION == addressed logical unit: 0)b */
+		buf[off] = 0x00;
+		/* Identifier/Designator type == NAA identifier */
+		buf[off++] = 0x3;
+		off++;
+		/* Identifier/Designator length */
+		buf[off++] = 0x10;
+		/*
+		 * Start NAA IEEE Registered Extended Identifier/Designator
+		 */
+		buf[off++] = (0x6 << 4);
+		/*
+		 * Use OpenFabrics IEEE Company ID: 00 14 05
+		 */
+		buf[off++] = 0x01;
+		buf[off++] = 0x40;
+		buf[off] = (0x5 << 4);
+		/*
+		 * Return ConfigFS Unit Serial Number information for
+		 * VENDOR_SPECIFIC_IDENTIFIER and
+		 * VENDOR_SPECIFIC_IDENTIFIER_EXTENTION
+		 */
+		binary = asciihex_to_binaryhex(
+					&DEV_T10_WWN(dev)->unit_serial[0]);
+		buf[off++] |= (binary & 0xf0) >> 4;
+		for (i = 0; i < 24; i += 2) {
+			binary_new = asciihex_to_binaryhex(
+				&DEV_T10_WWN(dev)->unit_serial[i+2]);
+			buf[off] = (binary & 0x0f) << 4;
+			buf[off++] |= (binary_new & 0xf0) >> 4;
+			binary = binary_new;
+		}
+		len = 20;
+		off = (len + 4);
+check_t10_vend_desc:
+		/*
+		 * T10 Vendor Identifier Page, see spc4r17 section 7.7.3.4
+		 */
+		id_len = 8; /* For Vendor field */
+		prod_len = 4; /* For VPD Header */
+		prod_len += 8; /* For Vendor field */
+		prod_len += strlen(prod);
+		prod_len++; /* For : */
+
+		if (dev->se_sub_dev->su_dev_flags &
+					SDF_EMULATED_VPD_UNIT_SERIAL) {
+			unit_serial_len =
+				strlen(&DEV_T10_WWN(dev)->unit_serial[0]);
+			unit_serial_len++; /* For NULL Terminator */
+
+			if ((len + (id_len + 4) +
+			    (prod_len + unit_serial_len)) >
+					cmd->data_length) {
+				len += (prod_len + unit_serial_len);
+				goto check_port;
+			}
+			id_len += sprintf((unsigned char *)&buf[off+12],
+					"%s:%s", prod,
+					&DEV_T10_WWN(dev)->unit_serial[0]);
+		} else {
+			iqn_sn = transport_get_iqn_sn();
+			iqn_sn_len = strlen(iqn_sn);
+			iqn_sn_len++; /* For ":" */
+			se_location_len = strlen(se_location);
+			se_location_len++; /* For NULL Terminator */
+
+			if ((len + (id_len + 4) + (prod_len + iqn_sn_len +
+					se_location_len)) > cmd->data_length) {
+				len += (prod_len + iqn_sn_len +
+						se_location_len);
+				goto check_port;
+			}
+			id_len += sprintf((unsigned char *)&buf[off+12],
+				"%s:%s:%s", prod, iqn_sn, se_location);
+		}
+		buf[off] = 0x2; /* ASCII */
+		buf[off+1] = 0x1; /* T10 Vendor ID */
+		buf[off+2] = 0x0;
+		memcpy((unsigned char *)&buf[off+4], "LIO-ORG", 8);
+		/* Extra Byte for NULL Terminator */
+		id_len++;
+		/* Identifier Length */
+		buf[off+3] = id_len;
+		/* Header size for Designation descriptor */
+		len += (id_len + 4);
+		off += (id_len + 4);
+		/*
+		 * se_port_t is only set for INQUIRY VPD=1 through $FABRIC_MOD
+		 */
+check_port:
+		port = lun->lun_sep;
+		if (port) {
+			t10_alua_lu_gp_t *lu_gp;
+			u32 padding, scsi_name_len;
+			u16 lu_gp_id = 0;
+			u16 tg_pt_gp_id = 0;
+			u16 tpgt;
+
+			tpg = port->sep_tpg;
+			/*
+			 * Relative target port identifer, see spc4r17
+			 * section 7.7.3.7
+			 *
+			 * Get the PROTOCOL IDENTIFIER as defined by spc4r17
+			 * section 7.5.1 Table 362
+			 */
+			if (((len + 4) + 8) > cmd->data_length) {
+				len += 8;
+				goto check_tpgi;
+			}
+			buf[off] =
+				(TPG_TFO(tpg)->get_fabric_proto_ident() << 4);
+			buf[off++] |= 0x1; /* CODE SET == Binary */
+			buf[off] = 0x80; /* Set PIV=1 */
+			/* Set ASSOICATION == target port: 01b */
+			buf[off] |= 0x10;
+			/* DESIGNATOR TYPE == Relative target port identifer */
+			buf[off++] |= 0x4;
+			off++; /* Skip over Reserved */
+			buf[off++] = 4; /* DESIGNATOR LENGTH */
+			/* Skip over Obsolete field in RTPI payload
+			 * in Table 472 */
+			off += 2;
+			buf[off++] = ((port->sep_rtpi >> 8) & 0xff);
+			buf[off++] = (port->sep_rtpi & 0xff);
+			len += 8; /* Header size + Designation descriptor */
+			/*
+			 * Target port group identifier, see spc4r17
+			 * section 7.7.3.8
+			 *
+			 * Get the PROTOCOL IDENTIFIER as defined by spc4r17
+			 * section 7.5.1 Table 362
+			 */
+check_tpgi:
+			if (T10_ALUA(dev->se_sub_dev)->alua_type !=
+					SPC3_ALUA_EMULATED)
+				goto check_scsi_name;
+
+			if (((len + 4) + 8) > cmd->data_length) {
+				len += 8;
+				goto check_lu_gp;
+			}
+			tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
+			if (!(tg_pt_gp_mem))
+				goto check_lu_gp;
+
+			spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+			tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
+			if (!(tg_pt_gp)) {
+				spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+				goto check_lu_gp;
+			}
+			tg_pt_gp_id = tg_pt_gp->tg_pt_gp_id;
+			spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+
+			buf[off] =
+				(TPG_TFO(tpg)->get_fabric_proto_ident() << 4);
+			buf[off++] |= 0x1; /* CODE SET == Binary */
+			buf[off] = 0x80; /* Set PIV=1 */
+			/* Set ASSOICATION == target port: 01b */
+			buf[off] |= 0x10;
+			/* DESIGNATOR TYPE == Target port group identifier */
+			buf[off++] |= 0x5;
+			off++; /* Skip over Reserved */
+			buf[off++] = 4; /* DESIGNATOR LENGTH */
+			off += 2; /* Skip over Reserved Field */
+			buf[off++] = ((tg_pt_gp_id >> 8) & 0xff);
+			buf[off++] = (tg_pt_gp_id & 0xff);
+			len += 8; /* Header size + Designation descriptor */
+			/*
+			 * Logical Unit Group identifier, see spc4r17
+			 * section 7.7.3.8
+			 */
+check_lu_gp:
+			if (((len + 4) + 8) > cmd->data_length) {
+				len += 8;
+				goto check_scsi_name;
+			}
+			lu_gp_mem = dev->dev_alua_lu_gp_mem;
+			if (!(lu_gp_mem))
+				goto check_scsi_name;
+
+			spin_lock(&lu_gp_mem->lu_gp_mem_lock);
+			lu_gp = lu_gp_mem->lu_gp;
+			if (!(lu_gp)) {
+				spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
+				goto check_scsi_name;
+			}
+			lu_gp_id = lu_gp->lu_gp_id;
+			spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
+
+			buf[off++] |= 0x1; /* CODE SET == Binary */
+			/* DESIGNATOR TYPE == Logical Unit Group identifier */
+			buf[off++] |= 0x6;
+			off++; /* Skip over Reserved */
+			buf[off++] = 4; /* DESIGNATOR LENGTH */
+			off += 2; /* Skip over Reserved Field */
+			buf[off++] = ((lu_gp_id >> 8) & 0xff);
+			buf[off++] = (lu_gp_id & 0xff);
+			len += 8; /* Header size + Designation descriptor */
+			/*
+			 * SCSI name string designator, see spc4r17
+			 * section 7.7.3.11
+			 *
+			 * Get the PROTOCOL IDENTIFIER as defined by spc4r17
+			 * section 7.5.1 Table 362
+			 */
+check_scsi_name:
+			scsi_name_len = strlen(TPG_TFO(tpg)->tpg_get_wwn(tpg));
+			/* UTF-8 ",t,0x<16-bit TPGT>" + NULL Terminator */
+			scsi_name_len += 10;
+			/* Header size + Designation descriptor */
+			scsi_name_len += 4;
+			padding = ((-scsi_name_len) & 3);
+			if (padding != 0)
+				scsi_name_len += padding;
+
+			if ((len + scsi_name_len) > cmd->data_length) {
+				len += scsi_name_len;
+				goto set_len;
+			}
+			buf[off] =
+				(TPG_TFO(tpg)->get_fabric_proto_ident() << 4);
+			buf[off++] |= 0x3; /* CODE SET == UTF-8 */
+			buf[off] = 0x80; /* Set PIV=1 */
+			/* Set ASSOICATION == target port: 01b */
+			buf[off] |= 0x10;
+			/* DESIGNATOR TYPE == SCSI name string */
+			buf[off++] |= 0x8;
+			off += 2; /* Skip over Reserved and length */
+			/*
+			 * SCSI name string identifer containing, $FABRIC_MOD
+			 * dependent information.  For LIO-Target and iSCSI
+			 * Target Port, this means "<iSCSI name>,t,0x<TPGT> in
+			 * UTF-8 encoding.
+			 */
+			tpgt = TPG_TFO(tpg)->tpg_get_tag(tpg);
+			scsi_name_len = sprintf(&buf[off], "%s,t,0x%04x",
+					TPG_TFO(tpg)->tpg_get_wwn(tpg), tpgt);
+			scsi_name_len += 1 /* Include  NULL terminator */;
+			/*
+			 * The null-terminated, null-padded (see 4.4.2) SCSI
+			 * NAME STRING field contains a UTF-8 format string.
+			 * The number of bytes in the SCSI NAME STRING field
+			 * (i.e., the value in the DESIGNATOR LENGTH field)
+			 * shall be no larger than 256 and shall be a multiple
+			 * of four.
+			 */
+			if (padding)
+				scsi_name_len += padding;
+
+			buf[off-1] = scsi_name_len;
+			off += scsi_name_len;
+			/* Header size + Designation descriptor */
+			len += (scsi_name_len + 4);
+		}
+set_len:
+		buf[2] = ((len >> 8) & 0xff);
+		buf[3] = (len & 0xff); /* Page Length for VPD 0x83 */
+		break;
+	default:
+		printk(KERN_ERR "Unknown VPD Code: 0x%02x\n", cdb[2]);
+		return -1;
+	}
+
+	return 0;
+}
+
+int transport_generic_emulate_readcapacity(
+	se_cmd_t *cmd,
+	u32 blocks)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+
+	buf[0] = (blocks >> 24) & 0xff;
+	buf[1] = (blocks >> 16) & 0xff;
+	buf[2] = (blocks >> 8) & 0xff;
+	buf[3] = blocks & 0xff;
+	buf[4] = (DEV_ATTRIB(dev)->block_size >> 24) & 0xff;
+	buf[5] = (DEV_ATTRIB(dev)->block_size >> 16) & 0xff;
+	buf[6] = (DEV_ATTRIB(dev)->block_size >> 8) & 0xff;
+	buf[7] = DEV_ATTRIB(dev)->block_size & 0xff;
+
+	return 0;
+}
+
+int transport_generic_emulate_readcapacity_16(
+	se_cmd_t *cmd,
+	unsigned long long blocks)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+
+	buf[0] = (blocks >> 56) & 0xff;
+	buf[1] = (blocks >> 48) & 0xff;
+	buf[2] = (blocks >> 40) & 0xff;
+	buf[3] = (blocks >> 32) & 0xff;
+	buf[4] = (blocks >> 24) & 0xff;
+	buf[5] = (blocks >> 16) & 0xff;
+	buf[6] = (blocks >> 8) & 0xff;
+	buf[7] = blocks & 0xff;
+	buf[8] = (DEV_ATTRIB(dev)->block_size >> 24) & 0xff;
+	buf[9] = (DEV_ATTRIB(dev)->block_size >> 16) & 0xff;
+	buf[10] = (DEV_ATTRIB(dev)->block_size >> 8) & 0xff;
+	buf[11] = DEV_ATTRIB(dev)->block_size & 0xff;
+
+	return 0;
+}
+
+static int transport_modesense_rwrecovery(unsigned char *p)
+{
+	p[0] = 0x01;
+	p[1] = 0x0a;
+
+	return 12;
+}
+
+static int transport_modesense_control(se_device_t *dev, unsigned char *p)
+{
+	p[0] = 0x0a;
+	p[1] = 0x0a;
+	p[2] = 2;
+	/*
+	 * From spc4r17, section 7.4.6 Control mode Page
+	 *
+	 * Unit Attention interlocks control (UN_INTLCK_CTRL) to code 00b
+	 *
+	 * 00b: The logical unit shall clear any unit attention condition
+	 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
+	 * status and shall not establish a unit attention condition when a com-
+	 * mand is completed with BUSY, TASK SET FULL, or RESERVATION CONFLICT
+	 * status.
+	 *
+	 * 10b: The logical unit shall not clear any unit attention condition
+	 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
+	 * status and shall not establish a unit attention condition when
+	 * a command is completed with BUSY, TASK SET FULL, or RESERVATION
+	 * CONFLICT status.
+	 *
+	 * 11b a The logical unit shall not clear any unit attention condition
+	 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
+	 * status and shall establish a unit attention condition for the
+	 * initiator port associated with the I_T nexus on which the BUSY,
+	 * TASK SET FULL, or RESERVATION CONFLICT status is being returned.
+	 * Depending on the status, the additional sense code shall be set to
+	 * PREVIOUS BUSY STATUS, PREVIOUS TASK SET FULL STATUS, or PREVIOUS
+	 * RESERVATION CONFLICT STATUS. Until it is cleared by a REQUEST SENSE
+	 * command, a unit attention condition shall be established only once
+	 * for a BUSY, TASK SET FULL, or RESERVATION CONFLICT status regardless
+	 * to the number of commands completed with one of those status codes.
+	 */
+	p[4] = (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl == 2) ? 0x30 :
+	       (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl == 1) ? 0x20 : 0x00;
+	/*
+	 * From spc4r17, section 7.4.6 Control mode Page
+	 *
+	 * Task Aborted Status (TAS) bit set to zero.
+	 *
+	 * A task aborted status (TAS) bit set to zero specifies that aborted
+	 * tasks shall be terminated by the device server without any response
+	 * to the application client. A TAS bit set to one specifies that tasks
+	 * aborted by the actions of an I_T nexus other than the I_T nexus on
+	 * which the command was received shall be completed with TASK ABORTED
+	 * status (see SAM-4).
+	 */
+	p[5] = (DEV_ATTRIB(dev)->emulate_tas) ? 0x40 : 0x00;
+	p[8] = 0xff;
+	p[9] = 0xff;
+	p[11] = 30;
+
+	return 12;
+}
+
+static int transport_modesense_caching(unsigned char *p)
+{
+	p[0] = 0x08;
+	p[1] = 0x12;
+#if 0
+	p[2] = 0x04; /* Write Cache Enable */
+#endif
+	p[12] = 0x20; /* Disabled Read Ahead */
+
+	return 20;
+}
+
+#if 0
+static int transport_modesense_devicecaps(unsigned char *p)
+{
+	p[0] = 0x2a;
+	p[1] = 0x0a;
+
+	return 12;
+}
+#endif
+
+static void transport_modesense_write_protect(
+	unsigned char *buf,
+	int type)
+{
+	/*
+	 * I believe that the WP bit (bit 7) in the mode header is the same for
+	 * all device types..
+	 */
+	switch (type) {
+	case TYPE_DISK:
+	case TYPE_TAPE:
+	default:
+		buf[0] |= 0x80; /* WP bit */
+		break;
+	}
+}
+
+int transport_generic_emulate_modesense(
+	se_cmd_t *cmd,
+	unsigned char *cdb,
+	unsigned char *rbuf,
+	int ten,
+	int type)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	int offset = (ten) ? 8 : 4;
+	int length = 0;
+	unsigned char buf[SE_MODE_PAGE_BUF];
+
+	memset(buf, 0, SE_MODE_PAGE_BUF);
+
+	switch (cdb[2] & 0x3f) {
+	case 0x01:
+		length = transport_modesense_rwrecovery(&buf[offset]);
+		break;
+	case 0x08:
+		length = transport_modesense_caching(&buf[offset]);
+		break;
+	case 0x0a:
+		length = transport_modesense_control(dev, &buf[offset]);
+		break;
+#if 0
+	case 0x2a:
+		length = transport_modesense_devicecaps(&buf[offset]);
+		break;
+#endif
+	case 0x3f:
+		length = transport_modesense_rwrecovery(&buf[offset]);
+		length += transport_modesense_caching(&buf[offset+length]);
+		length += transport_modesense_control(dev, &buf[offset+length]);
+#if 0
+		length += transport_modesense_devicecaps(&buf[offset+length]);
+#endif
+		break;
+	default:
+		printk(KERN_ERR "Got Unknown Mode Page: 0x%02x\n",
+				cdb[2] & 0x3f);
+		return PYX_TRANSPORT_UNKNOWN_MODE_PAGE;
+	}
+	offset += length;
+
+	if (ten) {
+		offset -= 2;
+		buf[0] = (offset >> 8) & 0xff;
+		buf[1] = offset & 0xff;
+
+		if ((SE_LUN(cmd)->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
+		    (cmd->se_deve &&
+		    (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
+			transport_modesense_write_protect(&buf[3], type);
+
+		if ((offset + 2) > cmd->data_length)
+			offset = cmd->data_length;
+
+	} else {
+		offset -= 1;
+		buf[0] = offset & 0xff;
+
+		if ((SE_LUN(cmd)->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
+		    (cmd->se_deve &&
+		    (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
+			transport_modesense_write_protect(&buf[2], type);
+
+		if ((offset + 1) > cmd->data_length)
+			offset = cmd->data_length;
+	}
+	memcpy(rbuf, buf, offset);
+
+	return 0;
+}
+
+int transport_generic_emulate_request_sense(
+	se_cmd_t *cmd,
+	unsigned char *cdb)
+{
+	unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+	u8 ua_asc = 0, ua_ascq = 0;
+
+	if (cdb[1] & 0x01) {
+		printk(KERN_ERR "REQUEST_SENSE description emulation not"
+			" supported\n");
+		return PYX_TRANSPORT_INVALID_CDB_FIELD;
+	}
+	if (!(core_scsi3_ua_clear_for_request_sense(cmd, &ua_asc, &ua_ascq))) {
+		/*
+		 * CURRENT ERROR, UNIT ATTENTION
+		 */
+		buf[0] = 0x70;
+		buf[SPC_SENSE_KEY_OFFSET] = UNIT_ATTENTION;
+		/*
+		 * Make sure request data length is enough for additional
+		 * sense data.
+		 */
+		if (cmd->data_length <= 18) {
+			buf[7] = 0x00;
+			return 0;
+		}
+		/*
+		 * The Additional Sense Code (ASC) from the UNIT ATTENTION
+		 */
+		buf[SPC_ASC_KEY_OFFSET] = ua_asc;
+		buf[SPC_ASCQ_KEY_OFFSET] = ua_ascq;
+		buf[7] = 0x0A;
+	} else {
+		/*
+		 * CURRENT ERROR, NO SENSE
+		 */
+		buf[0] = 0x70;
+		buf[SPC_SENSE_KEY_OFFSET] = NO_SENSE;
+		/*
+		 * Make sure request data length is enough for additional
+		 * sense data.
+		 */
+		if (cmd->data_length <= 18) {
+			buf[7] = 0x00;
+			return 0;
+		}
+		/*
+		 * NO ADDITIONAL SENSE INFORMATION
+		 */
+		buf[SPC_ASC_KEY_OFFSET] = 0x00;
+		buf[7] = 0x0A;
+	}
+
+	return 0;
+}
+
+/*
+ * Used to obtain Sense Data from underlying Linux/SCSI struct scsi_cmnd
+ */
+int transport_get_sense_data(se_cmd_t *cmd)
+{
+	unsigned char *buffer = cmd->sense_buffer, *sense_buffer = NULL;
+	se_device_t *dev;
+	se_task_t *task = NULL, *task_tmp;
+	unsigned long flags;
+	u32 offset = 0;
+
+	if (!SE_LUN(cmd)) {
+		printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
+		return -1;
+	}
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (cmd->se_cmd_flags & SCF_SENT_CHECK_CONDITION) {
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return 0;
+	}
+
+	list_for_each_entry_safe(task, task_tmp,
+				&T_TASK(cmd)->t_task_list, t_list) {
+
+		if (!task->task_sense)
+			continue;
+
+		dev = task->se_dev;
+		if (!(dev))
+			continue;
+
+		if (!TRANSPORT(dev)->get_sense_buffer) {
+			printk(KERN_ERR "TRANSPORT(dev)->get_sense_buffer"
+					" is NULL\n");
+			continue;
+		}
+
+		sense_buffer = TRANSPORT(dev)->get_sense_buffer(task);
+		if (!(sense_buffer)) {
+			printk(KERN_ERR "ITT[0x%08x]_TASK[%d]: Unable to locate"
+				" sense buffer for task with sense\n",
+				CMD_TFO(cmd)->get_task_tag(cmd), task->task_no);
+			continue;
+		}
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+		offset = CMD_TFO(cmd)->set_fabric_sense_len(cmd,
+				TRANSPORT_SENSE_BUFFER);
+
+		memcpy((void *)&buffer[offset], (void *)sense_buffer,
+				TRANSPORT_SENSE_BUFFER);
+		cmd->scsi_status = task->task_scsi_status;
+		/* Automatically padded */
+		cmd->scsi_sense_length =
+				(TRANSPORT_SENSE_BUFFER + offset);
+
+		printk(KERN_INFO "HBA_[%u]_PLUG[%s]: Set SAM STATUS: 0x%02x"
+				" and sense\n",
+			dev->se_hba->hba_id, TRANSPORT(dev)->name,
+				cmd->scsi_status);
+		return 0;
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	return -1;
+}
+
+/*
+ * Generic function pointers for target_core_mod/ConfigFS
+ */
+#define SET_GENERIC_TRANSPORT_FUNCTIONS(cmd)				\
+do {									\
+	cmd->transport_allocate_iovecs =				\
+			&transport_generic_allocate_iovecs;		\
+	cmd->transport_get_task = &transport_generic_get_task;		\
+	cmd->transport_map_buffers_to_tasks =				\
+			&transport_generic_map_buffers_to_tasks;	\
+	cmd->transport_set_iovec_ptrs =					\
+			&transport_generic_set_iovec_ptrs;		\
+} while (0)
+
+/*	transport_generic_cmd_sequencer():
+ *
+ *	Generic Command Sequencer that should work for most DAS transport
+ *	drivers.
+ *
+ *	Called from transport_generic_allocate_tasks() in the $FABRIC_MOD
+ *	RX Thread.
+ *
+ *	FIXME: Need to support other SCSI OPCODES where as well.
+ */
+static int transport_generic_cmd_sequencer(
+	se_cmd_t *cmd,
+	unsigned char *cdb)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	se_subsystem_dev_t *su_dev = dev->se_sub_dev;
+	int ret = 0, sector_ret = 0;
+	u32 sectors = 0, size = 0, pr_reg_type = 0;
+	u8 alua_ascq = 0;
+	/*
+	 * Check for an existing UNIT ATTENTION condition
+	 */
+	if (core_scsi3_ua_check(cmd, cdb) < 0) {
+		cmd->transport_wait_for_tasks =
+				&transport_nop_wait_for_tasks;
+		transport_get_maps(cmd);
+		return 8; /* UNIT ATTENTION */
+	}
+	/*
+	 * Check status of Asymmetric Logical Unit Assignment port
+	 */
+	ret = T10_ALUA(su_dev)->alua_state_check(cmd, cdb, &alua_ascq);
+	if (ret != 0) {
+		cmd->transport_wait_for_tasks = &transport_nop_wait_for_tasks;
+		transport_get_maps(cmd);
+		/*
+		 * Set SCSI additional sense code (ASC) to 'LUN Not Accessable';
+		 * The ALUA additional sense code qualifier (ASCQ) is determined
+		 * by the ALUA primary or secondary access state..
+		 */
+		if (ret > 0) {
+#if 0
+			printk(KERN_INFO "[%s]: ALUA TG Port not available,"
+				" SenseKey: NOT_READY, ASC/ASCQ: 0x04/0x%02x\n",
+				CMD_TFO(cmd)->get_fabric_name(), alua_ascq);
+#endif
+			transport_set_sense_codes(cmd, 0x04, alua_ascq);
+			return 9; /* NOT READY */
+		}
+		return 6; /* INVALID_CDB_FIELD */
+	}
+	/*
+	 * Check status for SPC-3 Persistent Reservations
+	 */
+	if (T10_RES(su_dev)->t10_reservation_check(cmd, &pr_reg_type) != 0) {
+		if (T10_RES(su_dev)->t10_seq_non_holder(
+					cmd, cdb, pr_reg_type) != 0) {
+			cmd->transport_wait_for_tasks =
+					&transport_nop_wait_for_tasks;
+			transport_get_maps(cmd);
+			return 5; /* RESERVATION CONFLIT */
+		}
+		/*
+		 * This means the CDB is allowed for the SCSI Initiator port
+		 * when said port is *NOT* holding the legacy SPC-2 or
+		 * SPC-3 Persistent Reservation.
+		 */
+	}
+
+	switch (cdb[0]) {
+	case READ_6:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_6(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_6;
+		cmd->transport_get_lba = &transport_lba_21;
+		break;
+	case READ_10:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_10(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_10;
+		cmd->transport_get_lba = &transport_lba_32;
+		break;
+	case READ_12:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_12(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_12;
+		cmd->transport_get_lba = &transport_lba_32;
+		break;
+	case READ_16:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_16(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_16;
+		cmd->transport_get_long_lba = &transport_lba_64;
+		break;
+	case WRITE_6:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_6(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_6;
+		cmd->transport_get_lba = &transport_lba_21;
+		break;
+	case WRITE_10:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_10(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_10;
+		cmd->transport_get_lba = &transport_lba_32;
+		break;
+	case WRITE_12:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_12(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_12;
+		cmd->transport_get_lba = &transport_lba_32;
+		break;
+	case WRITE_16:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = transport_get_sectors_16(cdb, cmd, &sector_ret);
+		if (sector_ret)
+			return 4;
+		size = transport_get_size(sectors, cdb, cmd);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		cmd->transport_split_cdb = &split_cdb_XX_16;
+		cmd->transport_get_long_lba = &transport_lba_64;
+		break;
+	case 0xa3:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
+			/* MAINTENANCE_IN from SCC-2 */
+			/*
+			 * Check for emulated MI_REPORT_TARGET_PGS.
+			 */
+			if (cdb[1] == MI_REPORT_TARGET_PGS) {
+				cmd->transport_emulate_cdb =
+				(T10_ALUA(su_dev)->alua_type ==
+				 SPC3_ALUA_EMULATED) ?
+				&core_scsi3_emulate_report_target_port_groups :
+				NULL;
+			}
+			size = (cdb[6] << 24) | (cdb[7] << 16) |
+			       (cdb[8] << 8) | cdb[9];
+		} else {
+			/* GPCMD_SEND_KEY from multi media commands */
+			size = (cdb[8] << 8) + cdb[9];
+		}
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case MODE_SELECT:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = cdb[4];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 1;
+		break;
+	case MODE_SELECT_10:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[7] << 8) + cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 1;
+		break;
+	case MODE_SENSE:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = cdb[4];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case MODE_SENSE_10:
+	case GPCMD_READ_BUFFER_CAPACITY:
+	case GPCMD_SEND_OPC:
+	case LOG_SELECT:
+	case LOG_SENSE:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[7] << 8) + cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case READ_BLOCK_LIMITS:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = READ_BLOCK_LEN;
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case GPCMD_GET_CONFIGURATION:
+	case GPCMD_READ_FORMAT_CAPACITIES:
+	case GPCMD_READ_DISC_INFO:
+	case GPCMD_READ_TRACK_RZONE_INFO:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[7] << 8) + cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 1;
+		break;
+	case PERSISTENT_RESERVE_IN:
+	case PERSISTENT_RESERVE_OUT:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		cmd->transport_emulate_cdb =
+			(T10_RES(su_dev)->res_type ==
+			 SPC3_PERSISTENT_RESERVATIONS) ?
+			&core_scsi3_emulate_pr : NULL;
+		size = (cdb[7] << 8) + cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case GPCMD_MECHANISM_STATUS:
+	case GPCMD_READ_DVD_STRUCTURE:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[8] << 8) + cdb[9];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 1;
+		break;
+	case READ_POSITION:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = READ_POSITION_LEN;
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case 0xa4:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
+			/* MAINTENANCE_OUT from SCC-2
+			 *
+			 * Check for emulated MO_SET_TARGET_PGS.
+			 */
+			if (cdb[1] == MO_SET_TARGET_PGS) {
+				cmd->transport_emulate_cdb =
+				(T10_ALUA(su_dev)->alua_type ==
+					SPC3_ALUA_EMULATED) ?
+				&core_scsi3_emulate_set_target_port_groups :
+				NULL;
+			}
+
+			size = (cdb[6] << 24) | (cdb[7] << 16) |
+			       (cdb[8] << 8) | cdb[9];
+		} else  {
+			/* GPCMD_REPORT_KEY from multi media commands */
+			size = (cdb[8] << 8) + cdb[9];
+		}
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case INQUIRY:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[3] << 8) + cdb[4];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		/*
+		 * Do implict HEAD_OF_QUEUE processing for INQUIRY.
+		 * See spc4r17 section 5.3
+		 */
+		if (SE_DEV(cmd)->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+			cmd->sam_task_attr = TASK_ATTR_HOQ;
+		ret = 2;
+		break;
+	case READ_BUFFER:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case READ_CAPACITY:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = READ_CAP_LEN;
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case READ_MEDIA_SERIAL_NUMBER:
+	case SECURITY_PROTOCOL_IN:
+	case SECURITY_PROTOCOL_OUT:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case SERVICE_ACTION_IN:
+	case ACCESS_CONTROL_IN:
+	case ACCESS_CONTROL_OUT:
+	case EXTENDED_COPY:
+	case READ_ATTRIBUTE:
+	case RECEIVE_COPY_RESULTS:
+	case WRITE_ATTRIBUTE:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[10] << 24) | (cdb[11] << 16) |
+		       (cdb[12] << 8) | cdb[13];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case VARIABLE_LENGTH_CMD:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[10] << 8) | cdb[11];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case RECEIVE_DIAGNOSTIC:
+	case SEND_DIAGNOSTIC:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[3] << 8) | cdb[4];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+/* #warning FIXME: Figure out correct GPCMD_READ_CD blocksize. */
+#if 0
+	case GPCMD_READ_CD:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		sectors = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
+		size = (2336 * sectors);
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+#endif
+	case READ_TOC:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case REQUEST_SENSE:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = cdb[4];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case READ_ELEMENT_STATUS:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = 65536 * cdb[7] + 256 * cdb[8] + cdb[9];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case WRITE_BUFFER:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		ret = 2;
+		break;
+	case RESERVE:
+	case RESERVE_10:
+		/*
+		 * The SPC-2 RESERVE does not contain a size in the SCSI CDB.
+		 * Assume the passthrough or $FABRIC_MOD will tell us about it.
+		 */
+		if (cdb[0] == RESERVE_10)
+			size = (cdb[7] << 8) | cdb[8];
+		else
+			size = cmd->data_length;
+
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		cmd->transport_allocate_resources =
+				&transport_generic_allocate_none;
+		transport_get_maps(cmd);
+
+		ret = T10_RES(su_dev)->t10_reserve(cmd);
+		if (ret != 0) {
+			if (ret > 0)
+				return 5; /* RESERVATION_CONFILIT */
+			else
+				return 7; /* ILLEGAL_REQUEST */
+		}
+		ret = 3;
+		break;
+	case RELEASE:
+	case RELEASE_10:
+		/*
+		 * The SPC-2 RELEASE does not contain a size in the SCSI CDB.
+		 * Assume the passthrough or $FABRIC_MOD will tell us about it.
+		*/
+		if (cdb[0] == RELEASE_10)
+			size = (cdb[7] << 8) | cdb[8];
+		else
+			size = cmd->data_length;
+
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		cmd->transport_allocate_resources =
+				&transport_generic_allocate_none;
+		transport_get_maps(cmd);
+
+		T10_RES(su_dev)->t10_release(cmd);
+		ret = 3;
+		break;
+	case ALLOW_MEDIUM_REMOVAL:
+	case GPCMD_CLOSE_TRACK:
+	case ERASE:
+	case INITIALIZE_ELEMENT_STATUS:
+	case GPCMD_LOAD_UNLOAD:
+	case REZERO_UNIT:
+	case SEEK_10:
+	case GPCMD_SET_SPEED:
+	case SPACE:
+	case START_STOP:
+	case SYNCHRONIZE_CACHE:
+	case TEST_UNIT_READY:
+	case VERIFY:
+	case WRITE_FILEMARKS:
+	case MOVE_MEDIUM:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		cmd->transport_allocate_resources =
+				&transport_generic_allocate_none;
+		transport_get_maps(cmd);
+		ret = 3;
+		break;
+	case REPORT_LUNS:
+		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+		cmd->transport_emulate_cdb =
+				&transport_core_report_lun_response;
+		size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
+		CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+		transport_get_maps(cmd);
+		/*
+		 * Do implict HEAD_OF_QUEUE processing for REPORT_LUNS
+		 * See spc4r17 section 5.3
+		 */
+		if (SE_DEV(cmd)->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+			cmd->sam_task_attr = TASK_ATTR_HOQ;
+		ret = 2;
+		break;
+	default:
+		printk(KERN_WARNING "TARGET_CORE[%s]: Unsupported SCSI Opcode"
+			" 0x%02x, sending CHECK_CONDITION.\n",
+			CMD_TFO(cmd)->get_fabric_name(), cdb[0]);
+		cmd->transport_wait_for_tasks = &transport_nop_wait_for_tasks;
+		transport_get_maps(cmd);
+		return 4;
+	}
+
+	if (size != cmd->data_length) {
+		printk(KERN_WARNING "TARGET_CORE[%s]: Expected Transfer Length:"
+			" %u does not match SCSI CDB Length: %u for SAM Opcode:"
+			" 0x%02x\n", CMD_TFO(cmd)->get_fabric_name(),
+				cmd->data_length, size, cdb[0]);
+
+		cmd->cmd_spdtl = size;
+
+		if (cmd->data_direction == SE_DIRECTION_WRITE) {
+			printk(KERN_ERR "Rejecting underflow/overflow"
+					" WRITE data\n");
+			return 6;
+		}
+		/*
+		 * Reject READ_* or WRITE_* with overflow/underflow for
+		 * type SCF_SCSI_DATA_SG_IO_CDB.
+		 */
+		if (!(ret) && (DEV_ATTRIB(dev)->block_size != 512))  {
+			printk(KERN_ERR "Failing OVERFLOW/UNDERFLOW for LBA op"
+				" CDB on non 512-byte sector setup subsystem"
+				" plugin: %s\n", TRANSPORT(dev)->name);
+			/* Returns CHECK_CONDITION + INVALID_CDB_FIELD */
+			return 6;
+		}
+
+		if (size > cmd->data_length) {
+			cmd->se_cmd_flags |= SCF_OVERFLOW_BIT;
+			cmd->residual_count = (size - cmd->data_length);
+		} else {
+			cmd->se_cmd_flags |= SCF_UNDERFLOW_BIT;
+			cmd->residual_count = (cmd->data_length - size);
+		}
+		cmd->data_length = size;
+	}
+
+	transport_set_supported_SAM_opcode(cmd);
+	return ret;
+}
+
+static inline se_cmd_t *transport_alloc_passthrough_cmd(
+	u32 data_length,
+	int data_direction)
+{
+	return __transport_alloc_se_cmd(&passthrough_fabric_ops, NULL, NULL,
+			data_length, data_direction, TASK_ATTR_SIMPLE);
+}
+
+se_cmd_t *transport_allocate_passthrough(
+	unsigned char *cdb,
+	int data_direction,
+	u32 se_cmd_flags,
+	void *mem,
+	u32 se_mem_num,
+	u32 length,
+	se_obj_lun_type_t *obj_api,
+	void *type_ptr)
+{
+	se_cmd_t *cmd;
+	se_transform_info_t ti;
+
+	cmd = transport_alloc_passthrough_cmd(length, data_direction);
+	if (!(cmd))
+		return NULL;
+	/*
+	 * Simulate an SE LUN entry need for passing SCSI CDBs into
+	 * se_cmd_t.
+	 */
+	cmd->se_lun = kzalloc(sizeof(se_lun_t), GFP_KERNEL);
+	if (!(cmd->se_lun)) {
+		printk(KERN_ERR "Unable to allocate cmd->se_lun\n");
+		goto fail;
+	}
+
+	spin_lock_init(&cmd->se_lun->lun_sep_lock);
+	SE_LUN(cmd)->lun_type = obj_api->se_obj_type;
+	SE_LUN(cmd)->lun_type_ptr = type_ptr;
+	SE_LUN(cmd)->lun_obj_api = obj_api;
+
+	cmd->se_orig_obj_api = obj_api;
+	cmd->se_orig_obj_ptr = type_ptr;
+	cmd->se_cmd_flags = se_cmd_flags;
+	SE_LUN(cmd)->se_dev = (se_device_t *) type_ptr;
+
+	/*
+	 * Double check that the passed object is currently accepting CDBs
+	 */
+	if (obj_api->check_online(type_ptr) != 0) {
+		DEBUG_SO("obj_api->check_online() failed!\n");
+		goto fail;
+	}
+
+	cmd->data_length = length;
+	cmd->data_direction = data_direction;
+	cmd->se_cmd_flags |= SCF_CMD_PASSTHROUGH;
+
+	if (transport_generic_allocate_tasks(cmd, cdb) < 0)
+		goto fail;
+
+	memset(&ti, 0, sizeof(se_transform_info_t));
+	ti.ti_data_length = cmd->data_length;
+	ti.ti_dev = SE_LUN(cmd)->se_dev;
+	ti.ti_se_cmd = cmd;
+	ti.se_obj_ptr = type_ptr;
+	ti.se_obj_api = SE_LUN(cmd)->lun_obj_api;
+
+	DEBUG_SO("ti.se_obj_ptr: %p\n", ti.se_obj_ptr);
+	DEBUG_SO("ti.se_obj_api: %p\n", ti.se_obj_api);
+	DEBUG_SO("Plugin: %s\n", ti.se_obj_api->obj_plugin->plugin_name);
+
+	if (!mem) {
+		if (cmd->transport_allocate_resources(cmd, cmd->data_length,
+					PAGE_SIZE) < 0)
+			goto fail;
+	} else {
+		/*
+		 * Passed *mem will contain a list_head containing preformatted
+		 * se_mem_t elements...
+		 */
+		T_TASK(cmd)->t_mem_list = (struct list_head *)mem;
+		T_TASK(cmd)->t_task_se_num = se_mem_num;
+		cmd->se_cmd_flags |= SCF_CMD_PASSTHROUGH_NOALLOC;
+
+#ifdef DEBUG_PASSTHROUGH
+		{
+		u32 total_se_length = 0;
+		se_mem_t *se_mem, *se_mem_tmp;
+
+		DEBUG_PT("Preallocated se_mem_list: %p se_mem_num: %d\n",
+				mem, se_mem_num);
+
+		list_for_each_entry_safe(se_mem, se_mem_tmp,
+				T_TASK(cmd)->t_mem_list, se_list) {
+			total_se_length += se_mem->se_len;
+			DEBUG_PT("se_mem: %p se_mem->se_page: %p %d:%d\n",
+				se_mem, se_mem->se_page, se_mem->se_len,
+				se_mem->se_off);
+		}
+		DEBUG_PT("Total calculated total_se_length: %u\n",
+				total_se_length);
+
+		if (total_se_length != length) {
+			printk(KERN_ERR "Passed length: %u does not equal"
+				" total_se_length: %u\n", length,
+					total_se_length);
+			BUG();
+		}
+		}
+#endif
+	}
+
+	if (transport_get_sectors(cmd, SE_LUN(cmd)->lun_obj_api, type_ptr) < 0)
+		goto fail;
+
+	if (transport_new_cmd_obj(cmd, &ti, SE_LUN(cmd)->lun_obj_api,
+			type_ptr, 0) < 0)
+		goto fail;
+
+	return cmd;
+
+fail:
+	if (T_TASK(cmd))
+		transport_release_tasks(cmd);
+	kfree(T_TASK(cmd));
+	kfree(cmd->se_lun);
+	transport_free_se_cmd(cmd);
+
+	return NULL;
+}
+
+void transport_passthrough_release(
+	se_cmd_t *cmd)
+{
+	if (!cmd) {
+		printk(KERN_ERR "transport_passthrough_release passed"
+			" NULL se_cmd_t\n");
+		return;
+	}
+
+	if (cmd->transport_wait_for_tasks)
+		cmd->transport_wait_for_tasks(cmd, 0, 0);
+
+	transport_generic_remove(cmd, 0, 0);
+}
+
+int transport_passthrough_complete(
+	se_cmd_t *cmd)
+{
+	if (cmd->se_orig_obj_api->check_shutdown(cmd->se_orig_obj_ptr) != 0)
+		return -2;
+
+	switch (cmd->scsi_status) {
+	case 0x00: /* GOOD */
+		DEBUG_PT("SCSI Status: GOOD\n");
+		return 0;
+	case 0x02: /* CHECK_CONDITION */
+		DEBUG_PT("SCSI Status: CHECK_CONDITION\n");
+/* #warning FIXME: Do some basic return values for Sense Data */
+		return -1;
+	default:
+		DEBUG_PT("SCSI Status: 0x%02x\n", cmd->scsi_status);
+		return -1;
+	}
+
+	return 0;
+}
+
+/*     transport_generic_passthrough():
+ *
+ *
+ */
+int transport_generic_passthrough_async(
+	se_cmd_t *cmd,
+	void (*callback)(se_cmd_t *cmd,
+		void *callback_arg, int complete_status),
+	void *callback_arg)
+{
+	int write = (cmd->data_direction == SE_DIRECTION_WRITE);
+	int no_alloc = (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC);
+	int pt_done = (cmd->transport_passthrough_done != NULL);
+
+	if (callback) {
+		cmd->callback = callback;
+		cmd->callback_arg = callback_arg;
+	}
+
+	if (transport_generic_handle_cdb(cmd) < 0)
+		return -1;
+
+	if (write && !no_alloc) {
+		if (wait_for_completion_interruptible(
+			&T_TASK(cmd)->t_transport_passthrough_wcomp) != 0)
+			return -1;
+
+		transport_generic_process_write(cmd);
+	}
+
+	if (callback || pt_done)
+		return 0;
+
+	wait_for_completion(&T_TASK(cmd)->t_transport_passthrough_comp);
+
+	return transport_passthrough_complete(cmd);
+}
+
+int transport_generic_passthrough(se_cmd_t *cmd)
+{
+	return transport_generic_passthrough_async(cmd, NULL, NULL);
+}
+
+/*
+ * Called from transport_generic_complete_ok() and
+ * transport_generic_request_failure() to determine which dormant/delayed
+ * and ordered cmds need to have their tasks added to the execution queue.
+ */
+void transport_complete_task_attr(se_cmd_t *cmd)
+{
+	se_device_t *dev = SE_DEV(cmd);
+	se_cmd_t *cmd_p, *cmd_tmp;
+	int new_active_tasks = 0;
+
+	if (cmd->sam_task_attr == TASK_ATTR_SIMPLE) {
+		atomic_dec(&dev->simple_cmds);
+		smp_mb__after_atomic_dec();
+		dev->dev_cur_ordered_id++;
+		DEBUG_STA("Incremented dev->dev_cur_ordered_id: %u for"
+			" SIMPLE: %u\n", dev->dev_cur_ordered_id,
+			cmd->se_ordered_id);
+	} else if (cmd->sam_task_attr == TASK_ATTR_HOQ) {
+		atomic_dec(&dev->dev_hoq_count);
+		smp_mb__after_atomic_dec();
+		dev->dev_cur_ordered_id++;
+		DEBUG_STA("Incremented dev_cur_ordered_id: %u for"
+			" HEAD_OF_QUEUE: %u\n", dev->dev_cur_ordered_id,
+			cmd->se_ordered_id);
+	} else if (cmd->sam_task_attr == TASK_ATTR_ORDERED) {
+		spin_lock(&dev->ordered_cmd_lock);
+		list_del(&cmd->se_ordered_list);
+		atomic_dec(&dev->dev_ordered_sync);
+		smp_mb__after_atomic_dec();
+		spin_unlock(&dev->ordered_cmd_lock);
+
+		dev->dev_cur_ordered_id++;
+		DEBUG_STA("Incremented dev_cur_ordered_id: %u for ORDERED:"
+			" %u\n", dev->dev_cur_ordered_id, cmd->se_ordered_id);
+	}
+	/*
+	 * Process all commands up to the last received
+	 * ORDERED task attribute which requires another blocking
+	 * boundary
+	 */
+	spin_lock(&dev->delayed_cmd_lock);
+	list_for_each_entry_safe(cmd_p, cmd_tmp,
+			&dev->delayed_cmd_list, se_delayed_list) {
+
+		list_del(&cmd_p->se_delayed_list);
+		spin_unlock(&dev->delayed_cmd_lock);
+
+		DEBUG_STA("Calling add_tasks() for"
+			" cmd_p: 0x%02x Task Attr: 0x%02x"
+			" Dormant -> Active, se_ordered_id: %u\n",
+			T_TASK(cmd_p)->t_task_cdb[0],
+			cmd_p->sam_task_attr, cmd_p->se_ordered_id);
+
+		CMD_ORIG_OBJ_API(cmd_p)->add_tasks(
+				cmd_p->se_orig_obj_ptr, cmd_p);
+		new_active_tasks++;
+
+		spin_lock(&dev->delayed_cmd_lock);
+		if (cmd_p->sam_task_attr == TASK_ATTR_ORDERED)
+			break;
+	}
+	spin_unlock(&dev->delayed_cmd_lock);
+	/*
+	 * If new tasks have become active, wake up the transport thread
+	 * to do the processing of the Active tasks.
+	 */
+	if (new_active_tasks != 0)
+		wake_up_interruptible(&dev->dev_queue_obj->thread_wq);
+}
+
+/*	transport_generic_complete_ok():
+ *
+ *
+ */
+void transport_generic_complete_ok(se_cmd_t *cmd)
+{
+	int reason = 0;
+	/*
+	 * Check if we need to move delayed/dormant tasks from cmds on the
+	 * delayed execution list after a HEAD_OF_QUEUE or ORDERED Task
+	 * Attribute.
+	 */
+	if (SE_DEV(cmd)->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+		transport_complete_task_attr(cmd);
+	/*
+	 * Check if we need to retrieve a sense buffer from
+	 * the se_cmd_t in question.
+	 */
+	if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH) {
+		transport_lun_remove_cmd(cmd);
+		if (!(transport_cmd_check_stop(cmd, 2, 0)))
+			transport_passthrough_check_stop(cmd);
+		return;
+	} else if (cmd->se_cmd_flags & SCF_TRANSPORT_TASK_SENSE) {
+		if (transport_get_sense_data(cmd) < 0)
+			reason = NON_EXISTENT_LUN;
+
+		/*
+		 * Only set when an se_task_t->task_scsi_status returned
+		 * a non GOOD status.
+		 */
+		if (cmd->scsi_status) {
+			transport_send_check_condition_and_sense(
+					cmd, reason, 1);
+			transport_lun_remove_cmd(cmd);
+			transport_cmd_check_stop(cmd, 2, 0);
+			return;
+		}
+	}
+
+	switch (cmd->data_direction) {
+	case SE_DIRECTION_READ:
+#ifdef SNMP_SUPPORT
+		spin_lock(&cmd->se_lun->lun_sep_lock);
+		if (SE_LUN(cmd)->lun_sep) {
+			SE_LUN(cmd)->lun_sep->sep_stats.tx_data_octets +=
+					cmd->data_length;
+		}
+		spin_unlock(&cmd->se_lun->lun_sep_lock);
+#endif
+		CMD_TFO(cmd)->queue_data_in(cmd);
+		break;
+	case SE_DIRECTION_WRITE:
+#ifdef SNMP_SUPPORT
+		spin_lock(&cmd->se_lun->lun_sep_lock);
+		if (SE_LUN(cmd)->lun_sep) {
+			SE_LUN(cmd)->lun_sep->sep_stats.rx_data_octets +=
+				cmd->data_length;
+		}
+		spin_unlock(&cmd->se_lun->lun_sep_lock);
+#endif
+		/* Fall through for SE_DIRECTION_WRITE */
+	case SE_DIRECTION_NONE:
+		CMD_TFO(cmd)->queue_status(cmd);
+		break;
+	default:
+		break;
+	}
+
+	transport_lun_remove_cmd(cmd);
+	transport_cmd_check_stop(cmd, 2, 0);
+}
+
+void transport_free_dev_tasks(se_cmd_t *cmd)
+{
+	se_task_t *task, *task_tmp;
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	list_for_each_entry_safe(task, task_tmp,
+				&T_TASK(cmd)->t_task_list, t_list) {
+		if (atomic_read(&task->task_active))
+			continue;
+
+		if (!task->transport_req)
+			continue;
+
+		kfree(task->task_sg);
+
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		if (task->se_dev)
+			TRANSPORT(task->se_dev)->free_task(task);
+		else
+			printk(KERN_ERR "task[%u] - task->se_dev is NULL\n",
+				task->task_no);
+		spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+
+		list_del(&task->t_list);
+		kmem_cache_free(se_task_cache, task);
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+static inline void transport_free_pages(se_cmd_t *cmd)
+{
+	se_mem_t *se_mem, *se_mem_tmp;
+
+	if (T_TASK(cmd)->t_task_buf) {
+		kfree(T_TASK(cmd)->t_task_buf);
+		T_TASK(cmd)->t_task_buf = NULL;
+		return;
+	}
+
+	if (cmd->transport_free_resources) {
+		cmd->transport_free_resources(cmd);
+		return;
+	}
+	/*
+	 * Caller will handle releasing of se_mem_t.
+	 */
+	if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC)
+		return;
+
+	if (!(T_TASK(cmd)->t_task_se_num))
+		return;
+
+	list_for_each_entry_safe(se_mem, se_mem_tmp,
+			T_TASK(cmd)->t_mem_list, se_list) {
+		__free_page(se_mem->se_page);
+
+		list_del(&se_mem->se_list);
+		kfree(se_mem);
+	}
+
+	kfree(T_TASK(cmd)->t_mem_list);
+	T_TASK(cmd)->t_mem_list = NULL;
+	T_TASK(cmd)->t_task_se_num = 0;
+}
+
+void transport_release_tasks(se_cmd_t *cmd)
+{
+	CMD_ORIG_OBJ_API(cmd)->free_tasks(cmd->se_orig_obj_ptr, cmd);
+}
+
+static inline int transport_dec_and_check(se_cmd_t *cmd)
+{
+	unsigned long flags;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+		if (!(atomic_dec_and_test(&T_TASK(cmd)->t_fe_count))) {
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+					flags);
+			return 1;
+		}
+	}
+
+	if (atomic_read(&T_TASK(cmd)->t_se_count)) {
+		if (!(atomic_dec_and_test(&T_TASK(cmd)->t_se_count))) {
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+					flags);
+			return 1;
+		}
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	return 0;
+}
+
+void transport_release_fe_cmd(se_cmd_t *cmd)
+{
+	unsigned long flags;
+
+	if (transport_dec_and_check(cmd))
+		return;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	transport_all_task_dev_remove_state(cmd);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	transport_release_tasks(cmd);
+	transport_free_pages(cmd);
+
+	if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+		kfree(cmd->se_lun);
+
+	CMD_TFO(cmd)->release_cmd_direct(cmd);
+	transport_free_se_cmd(cmd);
+}
+
+/*	transport_generic_remove():
+ *
+ *
+ */
+int transport_generic_remove(
+	se_cmd_t *cmd,
+	int release_to_pool,
+	int session_reinstatement)
+{
+	unsigned long flags;
+
+	if (!(T_TASK(cmd)))
+		goto release_cmd;
+
+	if (transport_dec_and_check(cmd)) {
+		if (session_reinstatement) {
+			spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+			transport_all_task_dev_remove_state(cmd);
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+					flags);
+		}
+		return 1;
+	}
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	transport_all_task_dev_remove_state(cmd);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	transport_release_tasks(cmd);
+	transport_free_pages(cmd);
+
+release_cmd:
+	if (release_to_pool && !(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+		transport_release_cmd_to_pool(cmd);
+	else {
+		if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+			kfree(cmd->se_lun);
+
+		CMD_TFO(cmd)->release_cmd_direct(cmd);
+		transport_free_se_cmd(cmd);
+	}
+
+	return 0;
+}
+
+/*	transport_generic_map_buffers_to_tasks():
+ *
+ *	Called from transport_generic_new_cmd() in Transport Processing Thread.
+ */
+static int transport_generic_map_buffers_to_tasks(se_cmd_t *cmd)
+{
+	se_task_t *task = NULL;
+	int ret;
+
+	/*
+	 * Deal with non [READ,WRITE]_XX CDBs here.
+	 */
+	if (cmd->se_cmd_flags & SCF_SCSI_NON_DATA_CDB)
+		goto non_scsi_data;
+	else if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_NONSG_IO_CDB) {
+		list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+			if (atomic_read(&task->task_sent))
+				continue;
+
+			ret = task->transport_map_task(task, task->task_size);
+			if (ret < 0)
+				return ret;
+
+			DEBUG_CMD_M("Mapping SCF_SCSI_CONTROL_NONSG_IO_CDB"
+				" task_size: %u\n", task->task_size);
+		}
+		return 0;
+	}
+
+	/*
+	 * Determine the scatterlist offset for each se_task_t,
+	 * and segment and set pointers to storage transport buffers
+	 * via task->transport_map_task().
+	 */
+	list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+		if (atomic_read(&task->task_sent))
+			continue;
+
+		ret = task->transport_map_task(task, task->task_size);
+		if (ret < 0)
+			return ret;
+
+		DEBUG_CMD_M("Mapping task[%d]_se_obj_ptr[%p] %s_IO task_lba:"
+			" %llu task_size: %u task_sg_num: %d\n",
+			task->task_no, task->se_obj_ptr,
+			(cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB) ?
+			"CONTROL" : "DATA", task->task_lba, task->task_size,
+			task->task_sg_num);
+	}
+
+	return 0;
+
+non_scsi_data:
+	list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+		if (atomic_read(&task->task_sent))
+			continue;
+
+		ret = task->transport_map_task(task, task->task_size);
+		if (ret < 0)
+			return ret;
+
+		DEBUG_CMD_M("Mapping SCF_SCSI_NON_DATA_CDB task_size: %u"
+			" task->task_sg_num: %d\n", task->task_size,
+				task->task_sg_num);
+	}
+
+	return 0;
+}
+
+/*	transport_generic_do_transform():
+ *
+ *
+ */
+int transport_generic_do_transform(se_cmd_t *cmd, se_transform_info_t *ti)
+{
+	if (cmd->transport_cdb_transform(cmd, ti) < 0)
+		return -1;
+
+	return 0;
+}
+
+int transport_get_sectors(
+	se_cmd_t *cmd,
+	se_obj_lun_type_t *obj_api,
+	void *obj_ptr)
+{
+	if (!(cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB))
+		return 0;
+
+	T_TASK(cmd)->t_task_sectors =
+		(cmd->data_length / obj_api->blocksize(obj_ptr));
+	if (!(T_TASK(cmd)->t_task_sectors))
+		T_TASK(cmd)->t_task_sectors = 1;
+
+	if (obj_api->get_device_type(obj_ptr) != TYPE_DISK)
+		return 0;
+
+	if ((T_TASK(cmd)->t_task_lba + T_TASK(cmd)->t_task_sectors) >
+	     obj_api->total_sectors(obj_ptr)) {
+		printk(KERN_ERR "LBA: %llu Sectors: %u exceeds"
+			" obj_api->total_sectors(): %llu\n",
+			T_TASK(cmd)->t_task_lba, T_TASK(cmd)->t_task_sectors,
+			obj_api->total_sectors(obj_ptr));
+		cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+		cmd->scsi_sense_reason = SECTOR_COUNT_TOO_MANY;
+		return PYX_TRANSPORT_REQ_TOO_MANY_SECTORS;
+	}
+
+	return 0;
+}
+
+int transport_new_cmd_obj(
+	se_cmd_t *cmd,
+	se_transform_info_t *ti,
+	se_obj_lun_type_t *obj_api,
+	void *obj_ptr,
+	int post_execute)
+{
+	u32 task_cdbs = 0, task_offset = 0;
+	se_mem_t *se_mem_out = NULL;
+
+	if (!(cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)) {
+		task_cdbs++;
+		T_TASK(cmd)->t_task_cdbs++;
+	} else {
+		ti->ti_set_counts = 1;
+
+		task_cdbs = obj_api->get_cdb_count(obj_ptr, ti,
+				T_TASK(cmd)->t_task_lba,
+				T_TASK(cmd)->t_task_sectors,
+				NULL, &se_mem_out, &task_offset);
+		if (!(task_cdbs)) {
+			cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+			cmd->scsi_sense_reason =
+					LOGICAL_UNIT_COMMUNICATION_FAILURE;
+			return PYX_TRANSPORT_LU_COMM_FAILURE;
+		}
+		T_TASK(cmd)->t_task_cdbs += task_cdbs;
+
+		cmd->transport_cdb_transform =
+				&transport_process_data_sg_transform;
+#if 0
+		printk(KERN_INFO "[%s]: api: %p ptr: %p data_length: %u, LBA:"
+			" %llu t_task_sectors: %u, t_task_cdbs: %u\n",
+			obj_api->obj_plugin->plugin_name, obj_api, obj_ptr,
+			cmd->data_length, T_TASK(cmd)->t_task_lba,
+			T_TASK(cmd)->t_task_sectors, T_TASK(cmd)->t_task_cdbs);
+#endif
+	}
+
+	cmd->transport_do_transform = &transport_generic_do_transform;
+	if (!post_execute) {
+		atomic_set(&T_TASK(cmd)->t_task_cdbs_left, task_cdbs);
+		atomic_set(&T_TASK(cmd)->t_task_cdbs_ex_left, task_cdbs);
+		atomic_set(&T_TASK(cmd)->t_task_cdbs_timeout_left, task_cdbs);
+	} else {
+		atomic_add(task_cdbs, &T_TASK(cmd)->t_task_cdbs_left);
+		atomic_add(task_cdbs, &T_TASK(cmd)->t_task_cdbs_ex_left);
+		atomic_add(task_cdbs, &T_TASK(cmd)->t_task_cdbs_timeout_left);
+	}
+
+	return 0;
+}
+
+unsigned char *transport_get_vaddr(se_mem_t *se_mem)
+{
+	return page_address(se_mem->se_page) + se_mem->se_off;
+}
+
+struct list_head *transport_init_se_mem_list(void)
+{
+	struct list_head *se_mem_list;
+
+	se_mem_list = kzalloc(sizeof(struct list_head), GFP_KERNEL);
+	if (!(se_mem_list)) {
+		printk(KERN_ERR "Unable to allocate memory for se_mem_list\n");
+		return NULL;
+	}
+	INIT_LIST_HEAD(se_mem_list);
+
+	return se_mem_list;
+}
+
+void transport_free_se_mem_list(struct list_head *se_mem_list)
+{
+	se_mem_t *se_mem, *se_mem_tmp;
+
+	if (!se_mem_list)
+		return;
+
+	list_for_each_entry_safe(se_mem, se_mem_tmp, se_mem_list, se_list) {
+		list_del(&se_mem->se_list);
+		kfree(se_mem);
+	}
+	kfree(se_mem_list);
+}
+
+int transport_generic_get_mem(se_cmd_t *cmd, u32 length, u32 dma_size)
+{
+	unsigned char *buf;
+	se_mem_t *se_mem;
+
+	T_TASK(cmd)->t_mem_list = transport_init_se_mem_list();
+	if (!(T_TASK(cmd)->t_mem_list))
+		return -1;
+
+	while (length) {
+		se_mem = kzalloc(sizeof(se_mem_t), GFP_KERNEL);
+		if (!(se_mem)) {
+			printk(KERN_ERR "Unable to allocate se_mem_t\n");
+			goto out;
+		}
+		INIT_LIST_HEAD(&se_mem->se_list);
+		se_mem->se_len = (length > dma_size) ? dma_size : length;
+
+/* #warning FIXME Allocate contigous pages for se_mem_t elements */
+		se_mem->se_page = (struct page *) alloc_pages(GFP_KERNEL, 0);
+		if (!(se_mem->se_page)) {
+			printk(KERN_ERR "alloc_pages() failed\n");
+			goto out;
+		}
+
+		buf = kmap_atomic(se_mem->se_page, KM_IRQ0);
+		if (!(buf)) {
+			printk(KERN_ERR "kmap_atomic() failed\n");
+			goto out;
+		}
+		memset(buf, 0, se_mem->se_len);
+		kunmap_atomic(buf, KM_IRQ0);
+
+		list_add_tail(&se_mem->se_list, T_TASK(cmd)->t_mem_list);
+		T_TASK(cmd)->t_task_se_num++;
+
+		DEBUG_MEM("Allocated se_mem_t page(%p) Length(%u)"
+			" Offset(%u)\n", se_mem->se_page, se_mem->se_len,
+			se_mem->se_off);
+
+		length -= se_mem->se_len;
+	}
+
+	DEBUG_MEM("Allocated total se_mem_t elements(%u)\n",
+			T_TASK(cmd)->t_task_se_num);
+
+	return 0;
+out:
+	return -1;
+}
+
+extern u32 transport_calc_sg_num(
+	se_task_t *task,
+	se_mem_t *in_se_mem,
+	u32 task_offset)
+{
+	se_mem_t *se_mem = in_se_mem;
+	u32 sg_length, sg_offset, task_size = task->task_size;
+	u32 saved_task_offset = 0;
+
+	while (task_size) {
+		DEBUG_SC("se_mem->se_page(%p) se_mem->se_len(%u)"
+			" se_mem->se_off(%u) task_offset(%u)\n",
+			se_mem->se_page, se_mem->se_len,
+			se_mem->se_off, task_offset);
+
+		if (task_offset == 0) {
+			if (task_size > se_mem->se_len)
+				sg_length = se_mem->se_len;
+			else
+				sg_length = task_size;
+
+			DEBUG_SC("sg_length(%u) task_size(%u)\n",
+					sg_length, task_size);
+
+			if (saved_task_offset)
+				task_offset = saved_task_offset;
+		} else {
+			sg_offset = task_offset;
+
+			if ((se_mem->se_len - task_offset) > task_size)
+				sg_length = task_size;
+			else
+				sg_length = (se_mem->se_len - task_offset);
+
+			DEBUG_SC("sg_length(%u) task_size(%u)\n",
+					sg_length, task_size);
+
+			saved_task_offset = task_offset;
+			task_offset = 0;
+		}
+		task_size -= sg_length;
+
+		DEBUG_SC("task[%u] - Reducing task_size to(%u)\n",
+			task->task_no, task_size);
+
+		task->task_sg_num++;
+
+		list_for_each_entry_continue(se_mem,
+				task->task_se_cmd->t_task->t_mem_list, se_list)
+			break;
+
+		if (!se_mem)
+			break;
+	}
+
+	task->task_sg = kzalloc(task->task_sg_num *
+			sizeof(struct scatterlist), GFP_KERNEL);
+	if (!(task->task_sg)) {
+		printk(KERN_ERR "Unable to allocate memory for"
+				" task->task_sg\n");
+		return 0;
+	}
+
+	sg_init_table(&task->task_sg[0], task->task_sg_num);
+
+	DEBUG_SC("Successfully allocated task->task_sg_num(%u)\n",
+			task->task_sg_num);
+
+	return task->task_sg_num;
+}
+
+static inline int transport_set_task_sectors_disk(
+	se_task_t *task,
+	se_obj_lun_type_t *obj_api,
+	void *obj_ptr,
+	unsigned long long lba,
+	u32 sectors,
+	int *max_sectors_set)
+{
+	if ((lba + sectors) > obj_api->end_lba(obj_ptr, 1)) {
+		task->task_sectors = ((obj_api->end_lba(obj_ptr, 1) - lba) + 1);
+
+		if (task->task_sectors > obj_api->max_sectors(obj_ptr)) {
+			task->task_sectors = obj_api->max_sectors(obj_ptr);
+			*max_sectors_set = 1;
+		}
+	} else {
+		if (sectors > obj_api->max_sectors(obj_ptr)) {
+			task->task_sectors = obj_api->max_sectors(obj_ptr);
+			*max_sectors_set = 1;
+		} else
+			task->task_sectors = sectors;
+	}
+
+	return 0;
+}
+
+static inline int transport_set_task_sectors_non_disk(
+	se_task_t *task,
+	se_obj_lun_type_t *obj_api,
+	void *obj_ptr,
+	unsigned long long lba,
+	u32 sectors,
+	int *max_sectors_set)
+{
+	if (sectors > obj_api->max_sectors(obj_ptr)) {
+		task->task_sectors = obj_api->max_sectors(obj_ptr);
+		*max_sectors_set = 1;
+	} else
+		task->task_sectors = sectors;
+
+	return 0;
+}
+
+static inline int transport_set_task_sectors(
+	se_task_t *task,
+	se_obj_lun_type_t *obj_api,
+	void *obj_ptr,
+	unsigned long long lba,
+	u32 sectors,
+	int *max_sectors_set)
+{
+	return (obj_api->get_device_type(obj_ptr) == TYPE_DISK)	?
+		transport_set_task_sectors_disk(task, obj_api, obj_ptr,
+				lba, sectors, max_sectors_set) :
+		transport_set_task_sectors_non_disk(task, obj_api, obj_ptr,
+				lba, sectors, max_sectors_set);
+}
+
+int transport_map_sg_to_mem(
+	se_task_t *task,
+	struct list_head *se_mem_list,
+	void *in_mem,
+	se_mem_t *in_se_mem,
+	se_mem_t **out_se_mem,
+	u32 *se_mem_cnt,
+	u32 *task_offset)
+{
+	se_mem_t *se_mem;
+	struct scatterlist *sg_s;
+	u32 j = 0, saved_task_offset = 0, task_size = task->task_size;
+
+	if (!in_mem) {
+		printk(KERN_ERR "No source scatterlist\n");
+		return -1;
+	}
+	sg_s = (struct scatterlist *)in_mem;
+
+	while (task_size) {
+		se_mem = kzalloc(sizeof(se_mem_t), GFP_KERNEL);
+		if (!(se_mem)) {
+			printk(KERN_ERR "Unable to allocate se_mem_t\n");
+			return -1;
+		}
+		INIT_LIST_HEAD(&se_mem->se_list);
+
+		if (*task_offset == 0) {
+			se_mem->se_page = sg_page(&sg_s[j]);
+			se_mem->se_off = sg_s[j].offset;
+
+			if (task_size >= sg_s[j].length)
+				se_mem->se_len =  sg_s[j++].length;
+			else {
+				se_mem->se_len = task_size;
+
+				task_size -= se_mem->se_len;
+				if (!(task_size)) {
+					*task_offset = (se_mem->se_len +
+							saved_task_offset);
+					goto next;
+				}
+			}
+
+			if (saved_task_offset)
+				*task_offset = saved_task_offset;
+		} else {
+			se_mem->se_page = sg_page(&sg_s[j]);
+			se_mem->se_off = (*task_offset + sg_s[j].offset);
+
+			if ((sg_s[j].length - *task_offset) > task_size) {
+				se_mem->se_len = task_size;
+
+				task_size -= se_mem->se_len;
+				if (!(task_size)) {
+					*task_offset += se_mem->se_len;
+					goto next;
+				}
+			} else
+				se_mem->se_len = (sg_s[j++].length -
+						*task_offset);
+
+			saved_task_offset = *task_offset;
+			*task_offset = 0;
+		}
+		task_size -= se_mem->se_len;
+next:
+		list_add_tail(&se_mem->se_list, se_mem_list);
+		(*se_mem_cnt)++;
+	}
+
+	DEBUG_MEM("task[%u] - Mapped(%u) struct scatterlist segments to(%u)"
+		" se_mem_t\n", task->task_no, j, *se_mem_cnt);
+
+	return 0;
+}
+
+int transport_map_mem_to_mem(
+	se_task_t *task,
+	struct list_head *se_mem_list,
+	void *in_mem,
+	se_mem_t *in_se_mem,
+	se_mem_t **out_se_mem,
+	u32 *se_mem_cnt,
+	u32 *task_offset)
+{
+	se_mem_t *se_mem = in_se_mem, *se_mem_new;
+	u32 saved_task_offset = 0, task_size = task->task_size;
+
+	if (!se_mem) {
+		printk(KERN_ERR "Invalid se_mem_t pointer\n");
+		return -1;
+	}
+
+	while (task_size) {
+		se_mem_new = kzalloc(sizeof(se_mem_t), GFP_KERNEL);
+		if (!(se_mem_new)) {
+			printk(KERN_ERR "Unable to allocate se_mem_t\n");
+			return -1;
+		}
+		INIT_LIST_HEAD(&se_mem_new->se_list);
+
+		if (*task_offset == 0) {
+			se_mem_new->se_page = se_mem->se_page;
+			se_mem_new->se_off = se_mem->se_off;
+
+			if (task_size >= se_mem->se_len) {
+				se_mem_new->se_len = se_mem->se_len;
+
+				se_mem = list_entry(se_mem->se_list.next,
+							se_mem_t, se_list);
+				if (!(se_mem)) {
+					printk(KERN_ERR "Unable to locate next"
+							" se_mem_t\n");
+					return -1;
+				}
+			} else {
+				se_mem_new->se_len = task_size;
+
+				task_size -= se_mem_new->se_len;
+				if (!(task_size)) {
+					*task_offset = (se_mem_new->se_len +
+							saved_task_offset);
+					goto next;
+				}
+			}
+
+			if (saved_task_offset)
+				*task_offset = saved_task_offset;
+		} else {
+			se_mem_new->se_page = se_mem->se_page;
+			se_mem_new->se_off = (*task_offset + se_mem->se_off);
+
+			if ((se_mem->se_len - *task_offset) > task_size) {
+				se_mem_new->se_len = task_size;
+
+				task_size -= se_mem_new->se_len;
+				if (!(task_size)) {
+					*task_offset += se_mem_new->se_len;
+					goto next;
+				}
+			} else {
+				se_mem_new->se_len = (se_mem->se_len -
+							*task_offset);
+
+				se_mem = list_entry(se_mem->se_list.next,
+							se_mem_t, se_list);
+				if (!(se_mem)) {
+					printk(KERN_ERR "Unable to locate next"
+							" se_mem_t\n");
+					return -1;
+				}
+			}
+
+			saved_task_offset = *task_offset;
+			*task_offset = 0;
+		}
+		task_size -= se_mem_new->se_len;
+next:
+		list_add_tail(&se_mem_new->se_list, se_mem_list);
+		(*se_mem_cnt)++;
+
+		DEBUG_MEM2("task[%u] - se_mem_cnt(%u) se_page(%p) se_off(%u)"
+			" se_len(%u)\n", task->task_no, *se_mem_cnt,
+			se_mem_new->se_page, se_mem_new->se_off,
+			se_mem->se_len);
+		DEBUG_MEM2("task[%u] - Reducing task_size to(%u)\n",
+			task->task_no, task_size);
+	}
+	*out_se_mem = se_mem;
+
+	return 0;
+}
+
+/*	transport_map_mem_to_sg():
+ *
+ *
+ */
+int transport_map_mem_to_sg(
+	se_task_t *task,
+	struct list_head *se_mem_list,
+	void *in_mem,
+	se_mem_t *in_se_mem,
+	se_mem_t **out_se_mem,
+	u32 *se_mem_cnt,
+	u32 *task_offset)
+{
+	se_mem_t *se_mem = in_se_mem;
+	struct scatterlist *sg = (struct scatterlist *)in_mem;
+	u32 saved_task_offset = 0, sg_no = 0;
+	u32 task_size = task->task_size;
+
+	if (!sg) {
+		printk(KERN_ERR "Unable to locate valid struct"
+				" scatterlist pointer\n");
+		return -1;
+	}
+
+	while (task_size) {
+		if (*task_offset == 0) {
+			sg_assign_page(&sg[sg_no], se_mem->se_page);
+			sg[sg_no].offset = se_mem->se_off;
+
+			if (task_size >= se_mem->se_len) {
+				sg[sg_no].length = se_mem->se_len;
+
+				se_mem = list_entry(se_mem->se_list.next,
+							se_mem_t, se_list);
+				if (!(se_mem)) {
+					printk(KERN_ERR "Unable to locate"
+						" next se_mem_t\n");
+					return -1;
+				}
+				(*se_mem_cnt)++;
+			} else {
+				sg[sg_no].length = task_size;
+				/*
+				 * Determine if we need to calculate an offset
+				 * into the se_mem_t on the next go around..
+				 */
+				task_size -= sg[sg_no].length;
+				if (!(task_size)) {
+					*task_offset = (sg[sg_no].length +
+							saved_task_offset);
+					goto next;
+				}
+			}
+
+			if (saved_task_offset)
+				*task_offset = saved_task_offset;
+		} else {
+			sg_assign_page(&sg[sg_no], se_mem->se_page);
+			sg[sg_no].offset = (*task_offset + se_mem->se_off);
+
+			if ((se_mem->se_len - *task_offset) > task_size) {
+				sg[sg_no].length = task_size;
+				/*
+				 * Determine if we need to calculate an offset
+				 * into the se_mem_t on the next go around..
+				 */
+				task_size -= sg[sg_no].length;
+				if (!(task_size)) {
+					*task_offset += sg[sg_no].length;
+					goto next;
+				}
+			} else {
+				sg[sg_no].length = (se_mem->se_len -
+						*task_offset);
+
+				se_mem = list_entry(se_mem->se_list.next,
+						se_mem_t, se_list);
+				if (!(se_mem)) {
+					printk(KERN_ERR "Unable to locate"
+						" next se_mem_t\n");
+					return -1;
+				}
+				(*se_mem_cnt)++;
+			}
+
+			saved_task_offset = *task_offset;
+			*task_offset = 0;
+		}
+		task_size -= sg[sg_no].length;
+next:
+		DEBUG_MEM("task[%u] - sg[%u](%p)(%u)(%u) - Reducing task_size"
+			" to(%u)\n", task->task_no, sg_no,
+			sg_page(&sg[sg_no]), sg[sg_no].length,
+			sg[sg_no].offset, task_size);
+
+		sg_no++;
+	}
+	*out_se_mem = se_mem;
+	task->task_sg_num = sg_no;
+
+	DEBUG_MEM("task[%u] - Mapped(%u) se_mem_t segments to total(%u) SGs"
+		" saved task_offset(%u)\n", task->task_no, *se_mem_cnt,
+			sg_no, *task_offset);
+
+	return 0;
+}
+
+u32 transport_generic_get_cdb_count(
+	se_cmd_t *cmd,
+	se_transform_info_t *ti,
+	se_obj_lun_type_t *head_obj_api,
+	void *head_obj_ptr,
+	unsigned long long starting_lba,
+	u32 sectors,
+	se_mem_t *se_mem_in,
+	se_mem_t **se_mem_out,
+	u32 *task_offset_in)
+{
+	unsigned char *cdb = NULL;
+	void *obj_ptr, *next_obj_ptr = NULL;
+	se_task_t *task;
+	se_mem_t *se_mem, *se_mem_lout = NULL;
+	se_obj_lun_type_t *obj_api;
+	int max_sectors_set = 0, ret;
+	u32 se_mem_cnt = 0, task_cdbs = 0;
+	unsigned long long lba;
+
+	if (!se_mem_in) {
+		list_for_each_entry(se_mem_in, T_TASK(cmd)->t_mem_list, se_list)
+			break;
+
+		if (!se_mem_in) {
+			printk(KERN_ERR "se_mem_in is NULL\n");
+			return 0;
+		}
+	}
+	se_mem = se_mem_in;
+
+	/*
+	 * Locate the start volume segment in which the received LBA will be
+	 * executed upon.
+	 */
+	head_obj_api->obtain_obj_lock(head_obj_ptr);
+	if (head_obj_api->obj_start(head_obj_ptr, ti, starting_lba) < 0) {
+		head_obj_api->release_obj_lock(head_obj_ptr);
+		return 0;
+	}
+
+	/*
+	 * Locate starting object from original starting_lba.
+	 */
+	lba = ti->ti_lba;
+	obj_api = ti->ti_obj_api;
+	obj_ptr = ti->ti_obj_ptr;
+	DEBUG_VOL("Starting Physical LBA(%llu) for head_obj_api->(%p)\n",
+			lba, head_obj_api);
+
+	while (sectors) {
+		if (!obj_api) {
+			head_obj_api->release_obj_lock(head_obj_ptr);
+			printk(KERN_ERR "obj_api is NULL LBA(%llu)->Sectors"
+				"(%u)\n", lba, sectors);
+			return 0;
+		}
+
+		DEBUG_VOL("ITT[0x%08x] LBA(%llu) SectorsLeft(%u) EOBJ(%llu)\n",
+			CMD_TFO(cmd)->get_task_tag(cmd), lba, sectors,
+			obj_api->end_lba(obj_ptr, 1));
+
+		head_obj_api->release_obj_lock(head_obj_ptr);
+
+		task = cmd->transport_get_task(ti, cmd, obj_ptr, obj_api);
+		if (!(task))
+			goto out;
+
+		transport_set_task_sectors(task, obj_api, obj_ptr, lba,
+				sectors, &max_sectors_set);
+
+		task->task_lba = lba;
+		lba += task->task_sectors;
+		sectors -= task->task_sectors;
+		task->task_size = (task->task_sectors *
+					obj_api->blocksize(obj_ptr));
+		task->transport_map_task =
+			obj_api->get_map_SG(obj_ptr, cmd->data_direction);
+
+		cdb = obj_api->get_cdb(obj_ptr, task);
+		if ((cdb)) {
+			memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+			cmd->transport_split_cdb(task->task_lba,
+					&task->task_sectors, cdb);
+		}
+
+		/*
+		 * Perform the SE OBJ plugin and/or Transport plugin specific
+		 * mapping for T_TASK(cmd)->t_mem_list.
+		 */
+		ret = obj_api->do_se_mem_map(obj_ptr, task,
+				T_TASK(cmd)->t_mem_list, NULL, se_mem,
+				&se_mem_lout, &se_mem_cnt, task_offset_in);
+		if (ret < 0)
+			goto out;
+
+		head_obj_api->obtain_obj_lock(head_obj_ptr);
+
+		se_mem = se_mem_lout;
+		*se_mem_out = se_mem_lout;
+		task_cdbs++;
+
+		DEBUG_VOL("Incremented task_cdbs(%u) task->task_sg_num(%u)\n",
+				task_cdbs, task->task_sg_num);
+
+		if (max_sectors_set) {
+			max_sectors_set = 0;
+			continue;
+		}
+
+		if (!sectors)
+			break;
+
+		obj_api = obj_api->get_next_obj_api(obj_ptr, &next_obj_ptr);
+		if (obj_api) {
+			obj_ptr = next_obj_ptr;
+			lba = obj_api->get_next_lba(obj_ptr, lba);
+		}
+	}
+	head_obj_api->release_obj_lock(head_obj_ptr);
+
+	if (ti->ti_set_counts) {
+		atomic_inc(&T_TASK(cmd)->t_fe_count);
+		atomic_inc(&T_TASK(cmd)->t_se_count);
+	}
+
+	DEBUG_VOL("ITT[0x%08x] total cdbs(%u)\n",
+		CMD_TFO(cmd)->get_task_tag(cmd), task_cdbs);
+
+	return task_cdbs;
+out:
+	return 0;
+}
+
+/*	 transport_generic_new_cmd(): Called from transport_processing_thread()
+ *
+ *	 Allocate storage transport resources from a set of values predefined
+ *	 by transport_generic_cmd_sequencer() from the iSCSI Target RX process.
+ *	 Any non zero return here is treated as an "out of resource' op here.
+ */
+int transport_generic_new_cmd(se_cmd_t *cmd)
+{
+	int ret = 0;
+	se_transform_info_t ti;
+	/*
+	 * Generate se_task_t(s) and/or their payloads for this CDB.
+	 */
+	memset((void *)&ti, 0, sizeof(se_transform_info_t));
+	ti.ti_se_cmd = cmd;
+	ti.se_obj_ptr = SE_LUN(cmd)->lun_type_ptr;
+	ti.se_obj_api = SE_LUN(cmd)->lun_obj_api;
+
+	if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)) {
+		/* #warning FIXME v3.2: Enable > PAGE_SIZE usage */
+		ret = cmd->transport_allocate_resources(cmd, cmd->data_length,
+				PAGE_SIZE);
+		if (ret < 0)
+			goto failure;
+
+		ret = transport_get_sectors(cmd, SE_LUN(cmd)->lun_obj_api,
+					SE_LUN(cmd)->lun_type_ptr);
+		if (ret < 0)
+			goto failure;
+
+		ret = transport_new_cmd_obj(cmd, &ti, SE_LUN(cmd)->lun_obj_api,
+					SE_LUN(cmd)->lun_type_ptr, 0);
+		if (ret < 0)
+			goto failure;
+
+		/*
+		 * Allocate iovecs for frontend mappings.  This currently
+		 * assumes traditional iSCSI going to sockets.
+		 *
+		 * FIXME: This should be specific to frontend protocol/hardware.
+		 */
+		ret = cmd->transport_allocate_iovecs(cmd);
+		if (ret < 0) {
+			ret = PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
+			goto failure;
+		}
+	}
+	/*
+	 * This is dependent upon the storage processing algorithm.
+	 */
+	if (cmd->transport_do_transform(cmd, &ti) < 0) {
+		ret = PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
+		goto failure;
+	}
+	/*
+	 * Set the correct (usually DMAable) buffer pointers from the master
+	 * buffer list in se_cmd_t to the transport task's native
+	 * buffers format.
+	 */
+	ret = cmd->transport_map_buffers_to_tasks(cmd);
+	if (ret < 0)
+		goto failure;
+	/*
+	 * For WRITEs, let the iSCSI Target RX Thread know its buffer is ready..
+	 * This WRITE se_cmd_t (and all of its associated se_task_t's)
+	 * will be added to the se_device_t execution queue after its WRITE
+	 * data has arrived. (ie: It gets handled by the transport processing
+	 * thread a second time)
+	 */
+	if (cmd->data_direction == SE_DIRECTION_WRITE) {
+		transport_add_tasks_to_state_queue(cmd);
+		return transport_generic_write_pending(cmd);
+	}
+	/*
+	 * Everything else but a WRITE, add the se_cmd_t's se_task_t's
+	 * to the execution queue.
+	 */
+	transport_execute_tasks(cmd);
+	return 0;
+
+failure:
+	return ret;
+}
+
+/*	transport_generic_process_write():
+ *
+ *
+ */
+void transport_generic_process_write(se_cmd_t *cmd)
+{
+#if 0
+	/*
+	 * Copy SCSI Presented DTL sector(s) from received buffers allocated to
+	 * original EDTL
+	 */
+	if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT) {
+		if (!T_TASK(cmd)->t_task_se_num) {
+			unsigned char *dst, *buf =
+				(unsigned char *)T_TASK(cmd)->t_task_buf;
+
+			dst = kzalloc(cmd->cmd_spdtl), GFP_KERNEL);
+			if (!(dst)) {
+				printk(KERN_ERR "Unable to allocate memory for"
+						" WRITE underflow\n");
+				transport_generic_request_failure(cmd, NULL,
+					PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
+				return;
+			}
+			memcpy(dst, buf, cmd->cmd_spdtl);
+
+			kfree(T_TASK(cmd)->t_task_buf);
+			T_TASK(cmd)->t_task_buf = dst;
+		} else {
+			struct scatterlist *sg =
+				(struct scatterlist *sg)T_TASK(cmd)->t_task_buf;
+			struct scatterlist *orig_sg;
+
+			orig_sg = kzalloc(sizeof(struct scatterlist) *
+					T_TASK(cmd)->t_task_se_num,
+					GFP_KERNEL))) {
+			if (!(orig_sg)) {
+				printk(KERN_ERR "Unable to allocate memory"
+						" for WRITE underflow\n");
+				transport_generic_request_failure(cmd, NULL,
+					PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
+				return;
+			}
+
+			memcpy(orig_sg, T_TASK(cmd)->t_task_buf,
+					sizeof(struct scatterlist) *
+					T_TASK(cmd)->t_task_se_num);
+
+			cmd->data_length = cmd->cmd_spdtl;
+			/*
+			 * FIXME, clear out original se_task_t and state
+			 * information.
+			 */
+			if (transport_generic_new_cmd(cmd) < 0) {
+				transport_generic_request_failure(cmd, NULL,
+					PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
+				kfree(orig_sg);
+				return;
+			}
+
+			transport_memcpy_write_sg(cmd, orig_sg);
+		}
+	}
+#endif
+	transport_execute_tasks(cmd);
+}
+
+/*	transport_generic_write_pending():
+ *
+ *
+ */
+static int transport_generic_write_pending(se_cmd_t *cmd)
+{
+	unsigned long flags;
+	int ret;
+
+	if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH) {
+		if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC)) {
+			complete(&T_TASK(cmd)->t_transport_passthrough_wcomp);
+			transport_cmd_check_stop(cmd, 1, 0);
+			return PYX_TRANSPORT_WRITE_PENDING;
+		}
+
+		transport_generic_process_write(cmd);
+		transport_cmd_check_stop(cmd, 1, 0);
+		return PYX_TRANSPORT_WRITE_PENDING;
+	}
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	cmd->t_state = TRANSPORT_WRITE_PENDING;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+	/*
+	 * Call the fabric write_pending function here to let the
+	 * frontend know that WRITE buffers are ready.
+	 */
+	ret = CMD_TFO(cmd)->write_pending(cmd);
+	if (ret < 0) {
+		transport_cmd_check_stop(cmd, 1, 0);
+		return ret;
+	}
+
+	transport_cmd_check_stop(cmd, 1, 0);
+	return PYX_TRANSPORT_WRITE_PENDING;
+}
+
+/*	transport_release_cmd_to_pool():
+ *
+ *
+ */
+void transport_release_cmd_to_pool(se_cmd_t *cmd)
+{
+	if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+		kfree(cmd->se_lun);
+	/*
+	 * Release se_cmd_t->se_fabric_cmd_ptr in fabric
+	 */
+	CMD_TFO(cmd)->release_cmd_to_pool(cmd);
+
+	transport_free_se_cmd(cmd);
+}
+EXPORT_SYMBOL(transport_release_cmd_to_pool);
+
+/*	transport_generic_free_cmd():
+ *
+ *	Called from processing frontend to release storage engine resources
+ */
+void transport_generic_free_cmd(
+	se_cmd_t *cmd,
+	int wait_for_tasks,
+	int release_to_pool,
+	int session_reinstatement)
+{
+	if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD) || !T_TASK(cmd))
+		transport_release_cmd_to_pool(cmd);
+	else {
+		core_dec_lacl_count(cmd->se_sess->se_node_acl, cmd);
+
+		if (SE_LUN(cmd)) {
+#if 0
+			printk(KERN_INFO "cmd: %p ITT: 0x%08x contains"
+				" SE_LUN(cmd)\n", cmd,
+				CMD_TFO(cmd)->get_task_tag(cmd));
+#endif
+			transport_lun_remove_cmd(cmd);
+		}
+
+		if (wait_for_tasks && cmd->transport_wait_for_tasks)
+			cmd->transport_wait_for_tasks(cmd, 0, 0);
+
+		transport_generic_remove(cmd, release_to_pool,
+				session_reinstatement);
+	}
+}
+EXPORT_SYMBOL(transport_generic_free_cmd);
+
+static void transport_nop_wait_for_tasks(
+	se_cmd_t *cmd,
+	int remove_cmd,
+	int session_reinstatement)
+{
+	return;
+}
+
+/*	transport_lun_wait_for_tasks():
+ *
+ *	Called from ConfigFS context to stop the passed se_cmd_t to allow
+ *	an se_lun_t to be successfully shutdown.
+ */
+int transport_lun_wait_for_tasks(se_cmd_t *cmd, se_lun_t *lun)
+{
+	unsigned long flags;
+	/*
+	 * If the frontend has already requested this se_cmd_t to
+	 * be stopped, we can safely ignore this se_cmd_t.
+	 */
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (atomic_read(&T_TASK(cmd)->t_transport_stop)) {
+		atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+		DEBUG_TRANSPORT_S("ConfigFS ITT[0x%08x] - t_transport_stop =="
+			" TRUE, skipping\n", CMD_TFO(cmd)->get_task_tag(cmd));
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return -1;
+	}
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	CMD_ORIG_OBJ_API(cmd)->notify_obj(cmd->se_orig_obj_ptr);
+
+	DEBUG_TRANSPORT_S("ConfigFS: ITT[0x%08x] - stopping cmd....\n",
+		CMD_TFO(cmd)->get_task_tag(cmd));
+	wait_for_completion(&T_TASK(cmd)->transport_lun_stop_comp);
+	DEBUG_TRANSPORT_S("ConfigFS: ITT[0x%08x] - stopped cmd....\n",
+		CMD_TFO(cmd)->get_task_tag(cmd));
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	return 0;
+}
+EXPORT_SYMBOL(transport_lun_wait_for_tasks);
+
+/* #define DEBUG_CLEAR_LUN */
+#ifdef DEBUG_CLEAR_LUN
+#define DEBUG_CLEAR_L(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CLEAR_L(x...)
+#endif
+
+void transport_clear_lun_from_sessions(se_lun_t *lun)
+{
+	se_cmd_t *cmd = NULL, *cmd_p = NULL;
+	unsigned long flags;
+	/*
+	 * Do exception processing and return CHECK_CONDITION status to the
+	 * Initiator Port.
+	 */
+	spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+	list_for_each_entry_safe(cmd, cmd_p, &lun->lun_cmd_list, se_lun_list) {
+		if (!(T_TASK(cmd))) {
+			printk(KERN_ERR "ITT: 0x%08x, T_TASK(cmd) = NULL"
+				"[i,t]_state: %u/%u\n",
+				CMD_TFO(cmd)->get_task_tag(cmd),
+				CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state);
+			BUG();
+		}
+		list_del(&cmd->se_lun_list);
+		atomic_set(&T_TASK(cmd)->transport_lun_active, 0);
+		/*
+		 * This will notify iscsi_target_transport.c:
+		 * transport_cmd_check_stop() that a LUN shutdown is in
+		 * progress for the iscsi_cmd_t.
+		 */
+		spin_lock(&T_TASK(cmd)->t_state_lock);
+		DEBUG_CLEAR_L("SE_LUN[%d] - Setting T_TASK(cmd)->transport"
+			"_lun_stop for  ITT: 0x%08x\n",
+			SE_LUN(cmd)->unpacked_lun,
+			CMD_TFO(cmd)->get_task_tag(cmd));
+		atomic_set(&T_TASK(cmd)->transport_lun_stop, 1);
+		spin_unlock(&T_TASK(cmd)->t_state_lock);
+
+		spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
+
+		if (!(SE_LUN(cmd))) {
+			printk(KERN_ERR "ITT: 0x%08x, [i,t]_state: %u/%u\n",
+				CMD_TFO(cmd)->get_task_tag(cmd),
+				CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state);
+			BUG();
+		}
+		/*
+		 * If the Storage engine still owns the iscsi_cmd_t, determine
+		 * and/or stop its context.
+		 */
+		DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x before transport"
+			"_lun_wait_for_tasks()\n", SE_LUN(cmd)->unpacked_lun,
+			CMD_TFO(cmd)->get_task_tag(cmd));
+
+		if (transport_lun_wait_for_tasks(cmd, SE_LUN(cmd)) < 0) {
+			spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+			continue;
+		}
+
+		DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x after transport_lun"
+			"_wait_for_tasks(): SUCCESS\n",
+			SE_LUN(cmd)->unpacked_lun,
+			CMD_TFO(cmd)->get_task_tag(cmd));
+		/*
+		 * The Storage engine stopped this se_cmd_t before it was
+		 * send to the fabric frontend for delivery back to the
+		 * Initiator Node.  Return this SCSI CDB back with an
+		 * CHECK_CONDITION status.
+		 */
+		transport_send_check_condition_and_sense(cmd,
+				NON_EXISTENT_LUN, 0);
+		/*
+		 *  If the fabric frontend is waiting for this iscsi_cmd_t to
+		 * be released, notify the waiting thread now that LU has
+		 * finished accessing it.
+		 */
+		spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+		if (atomic_read(&T_TASK(cmd)->transport_lun_fe_stop)) {
+			DEBUG_CLEAR_L("SE_LUN[%d] - Detected FE stop for"
+				" se_cmd_t: %p ITT: 0x%08x\n",
+				lun->unpacked_lun,
+				cmd, CMD_TFO(cmd)->get_task_tag(cmd));
+
+			spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+					flags);
+			complete(&T_TASK(cmd)->transport_lun_fe_stop_comp);
+			spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+			continue;
+		}
+		atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+
+		DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x finished processing\n",
+			lun->unpacked_lun, CMD_TFO(cmd)->get_task_tag(cmd));
+
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+	}
+	spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
+}
+EXPORT_SYMBOL(transport_clear_lun_from_sessions);
+
+/*	transport_generic_wait_for_tasks():
+ *
+ *	Called from frontend or passthrough context to wait for storage engine
+ *	to pause and/or release frontend generated se_cmd_t.
+ */
+static void transport_generic_wait_for_tasks(
+	se_cmd_t *cmd,
+	int remove_cmd,
+	int session_reinstatement)
+{
+	unsigned long flags;
+
+	if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD))
+		return;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	/*
+	 * If we are already stopped due to an external event (ie: LUN shutdown)
+	 * sleep until the connection can have the passed se_cmd_t back.
+	 * The T_TASK(cmd)->transport_lun_stopped_sem will be upped by
+	 * transport_clear_lun_from_sessions() once the ConfigFS context caller
+	 * has completed its operation on the se_cmd_t.
+	 */
+	if (atomic_read(&T_TASK(cmd)->transport_lun_stop)) {
+		atomic_set(&T_TASK(cmd)->transport_lun_fe_stop, 1);
+
+		DEBUG_TRANSPORT_S("wait_for_tasks: Stopping"
+			" wait_for_completion(&T_TASK(cmd)transport_lun_fe"
+			"_stop_comp); for ITT: 0x%08x\n",
+			CMD_TFO(cmd)->get_task_tag(cmd));
+		/*
+		 * There is a special case for WRITES where a FE exception +
+		 * LUN shutdown means ConfigFS context is still sleeping on
+		 * transport_lun_stop_comp in transport_lun_wait_for_tasks().
+		 * We go ahead and up transport_lun_stop_comp just to be sure
+		 * here.
+		 */
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		complete(&T_TASK(cmd)->transport_lun_stop_comp);
+		wait_for_completion(&T_TASK(cmd)->transport_lun_fe_stop_comp);
+		spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+		/*
+		 * At this point, the frontend who was the originator of this
+		 * se_cmd_t, now owns the structure and can be released through
+		 * normal means below.
+		 */
+		DEBUG_TRANSPORT_S("wait_for_tasks: Stopped"
+			" wait_for_completion(&T_TASK(cmd)transport_lun_fe_"
+			"stop_comp); for ITT: 0x%08x\n",
+			CMD_TFO(cmd)->get_task_tag(cmd));
+
+		atomic_set(&T_TASK(cmd)->transport_lun_fe_stop, 0);
+		atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+	}
+	if (!atomic_read(&T_TASK(cmd)->t_transport_active))
+		goto remove;
+
+	atomic_set(&T_TASK(cmd)->t_transport_stop, 1);
+
+	DEBUG_TRANSPORT_S("wait_for_tasks: Stopping %p ITT: 0x%08x"
+		" i_state: %d, t_state/def_t_state: %d/%d, t_transport_stop"
+		" = TRUE\n", cmd, CMD_TFO(cmd)->get_task_tag(cmd),
+		CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state,
+		cmd->deferred_t_state);
+
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	CMD_ORIG_OBJ_API(cmd)->notify_obj(cmd->se_orig_obj_ptr);
+
+	wait_for_completion(&T_TASK(cmd)->t_transport_stop_comp);
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+	atomic_set(&T_TASK(cmd)->t_transport_stop, 0);
+
+	DEBUG_TRANSPORT_S("wait_for_tasks: Stopped wait_for_compltion("
+		"&T_TASK(cmd)->t_transport_stop_comp) for ITT: 0x%08x\n",
+		CMD_TFO(cmd)->get_task_tag(cmd));
+remove:
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+	if (!remove_cmd)
+		return;
+
+	transport_generic_free_cmd(cmd, 0, 0, session_reinstatement);
+}
+
+int transport_get_sense_codes(
+	se_cmd_t *cmd,
+	u8 *asc,
+	u8 *ascq)
+{
+	*asc = cmd->scsi_asc;
+	*ascq = cmd->scsi_ascq;
+
+	return 0;
+}
+
+int transport_set_sense_codes(
+	se_cmd_t *cmd,
+	u8 asc,
+	u8 ascq)
+{
+	cmd->scsi_asc = asc;
+	cmd->scsi_ascq = ascq;
+
+	return 0;
+}
+
+int transport_send_check_condition_and_sense(
+	se_cmd_t *cmd,
+	u8 reason,
+	int from_transport)
+{
+	unsigned char *buffer = cmd->sense_buffer;
+	unsigned long flags;
+	int offset;
+	u8 asc = 0, ascq = 0;
+
+	spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+	if (cmd->se_cmd_flags & SCF_SENT_CHECK_CONDITION) {
+		spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+		return 0;
+	}
+	cmd->se_cmd_flags |= SCF_SENT_CHECK_CONDITION;
+	spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+	if (!reason && from_transport)
+		goto after_reason;
+
+	if (!from_transport)
+		cmd->se_cmd_flags |= SCF_EMULATED_TASK_SENSE;
+	/*
+	 * Data Segment and SenseLength of the fabric response PDU.
+	 *
+	 * TRANSPORT_SENSE_BUFFER is now set to SCSI_SENSE_BUFFERSIZE
+	 * from include/scsi/scsi_cmnd.h
+	 */
+	offset = CMD_TFO(cmd)->set_fabric_sense_len(cmd,
+				TRANSPORT_SENSE_BUFFER);
+	/*
+	 * Actual SENSE DATA, see SPC-3 7.23.2  SPC_SENSE_KEY_OFFSET uses
+	 * SENSE KEY values from include/scsi/scsi.h
+	 */
+	switch (reason) {
+	case NON_EXISTENT_LUN:
+	case UNSUPPORTED_SCSI_OPCODE:
+	case SECTOR_COUNT_TOO_MANY:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ILLEGAL REQUEST */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ILLEGAL_REQUEST;
+		/* INVALID COMMAND OPERATION CODE */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x20;
+		break;
+	case UNKNOWN_MODE_PAGE:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ILLEGAL REQUEST */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ILLEGAL_REQUEST;
+		/* INVALID FIELD IN CDB */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x24;
+		break;
+	case CHECK_CONDITION_ABORT_CMD:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* BUS DEVICE RESET FUNCTION OCCURRED */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x29;
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x03;
+		break;
+	case INCORRECT_AMOUNT_OF_DATA:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* WRITE ERROR */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x0c;
+		/* NOT ENOUGH UNSOLICITED DATA */
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x0d;
+		break;
+	case INVALID_CDB_FIELD:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* INVALID FIELD IN CDB */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x24;
+		break;
+	case INVALID_PARAMETER_LIST:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* INVALID FIELD IN PARAMETER LIST */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x26;
+		break;
+	case UNEXPECTED_UNSOLICITED_DATA:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* WRITE ERROR */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x0c;
+		/* UNEXPECTED_UNSOLICITED_DATA */
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x0c;
+		break;
+	case SERVICE_CRC_ERROR:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* PROTOCOL SERVICE CRC ERROR */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x47;
+		/* N/A */
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x05;
+		break;
+	case SNACK_REJECTED:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ABORTED COMMAND */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+		/* READ ERROR */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x11;
+		/* FAILED RETRANSMISSION REQUEST */
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x13;
+		break;
+	case WRITE_PROTECTED:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* DATA PROTECT */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = DATA_PROTECT;
+		/* WRITE PROTECTED */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x27;
+		break;
+	case CHECK_CONDITION_UNIT_ATTENTION:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* UNIT ATTENTION */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = UNIT_ATTENTION;
+		core_scsi3_ua_for_check_condition(cmd, &asc, &ascq);
+		buffer[offset+SPC_ASC_KEY_OFFSET] = asc;
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = ascq;
+		break;
+	case CHECK_CONDITION_NOT_READY:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* Not Ready */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = NOT_READY;
+		transport_get_sense_codes(cmd, &asc, &ascq);
+		buffer[offset+SPC_ASC_KEY_OFFSET] = asc;
+		buffer[offset+SPC_ASCQ_KEY_OFFSET] = ascq;
+		break;
+	case LOGICAL_UNIT_COMMUNICATION_FAILURE:
+	default:
+		/* CURRENT ERROR */
+		buffer[offset] = 0x70;
+		/* ILLEGAL REQUEST */
+		buffer[offset+SPC_SENSE_KEY_OFFSET] = ILLEGAL_REQUEST;
+		/* LOGICAL UNIT COMMUNICATION FAILURE */
+		buffer[offset+SPC_ASC_KEY_OFFSET] = 0x80;
+		break;
+	}
+	/*
+	 * This code uses linux/include/scsi/scsi.h SAM status codes!
+	 */
+	cmd->scsi_status = SAM_STAT_CHECK_CONDITION;
+	/*
+	 * Automatically padded, this value is encoded in the fabric's
+	 * data_length response PDU containing the SCSI defined sense data.
+	 */
+	cmd->scsi_sense_length  = TRANSPORT_SENSE_BUFFER + offset;
+
+after_reason:
+	if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+		CMD_TFO(cmd)->queue_status(cmd);
+
+	return 0;
+}
+EXPORT_SYMBOL(transport_send_check_condition_and_sense);
+
+int transport_check_aborted_status(se_cmd_t *cmd, int send_status)
+{
+	int ret = 0;
+
+	if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+		return 0;
+
+	if (atomic_read(&T_TASK(cmd)->t_transport_aborted) != 0) {
+		if (!(send_status) ||
+		     (cmd->se_cmd_flags & SCF_SENT_DELAYED_TAS))
+			return 1;
+#if 0
+		printk(KERN_INFO "Sending delayed SAM_STAT_TASK_ABORTED"
+			" status for CDB: 0x%02x ITT: 0x%08x\n",
+			T_TASK(cmd)->t_task_cdb[0],
+			CMD_TFO(cmd)->get_task_tag(cmd));
+#endif
+		cmd->se_cmd_flags |= SCF_SENT_DELAYED_TAS;
+		CMD_TFO(cmd)->queue_status(cmd);
+		ret = 1;
+	}
+	return ret;
+}
+EXPORT_SYMBOL(transport_check_aborted_status);
+
+void transport_send_task_abort(se_cmd_t *cmd)
+{
+	if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+		return;
+	/*
+	 * If there are still expected incoming fabric WRITEs, we wait
+	 * until until they have completed before sending a TASK_ABORTED
+	 * response.  This response with TASK_ABORTED status will be
+	 * queued back to fabric module by transport_check_aborted_status().
+	 */
+	if ((cmd->data_direction == SE_DIRECTION_WRITE) ||
+	    (cmd->data_direction == SE_DIRECTION_BIDI)) {
+		if (CMD_TFO(cmd)->write_pending_status(cmd) != 0) {
+			atomic_inc(&T_TASK(cmd)->t_transport_aborted);
+			smp_mb__after_atomic_inc();
+			cmd->scsi_status = SAM_STAT_TASK_ABORTED;
+			transport_new_cmd_failure(cmd);
+			return;
+		}
+	}
+	cmd->scsi_status = SAM_STAT_TASK_ABORTED;
+#if 0
+	printk(KERN_INFO "Setting SAM_STAT_TASK_ABORTED status for CDB: 0x%02x,"
+		" ITT: 0x%08x\n", T_TASK(cmd)->t_task_cdb[0],
+		CMD_TFO(cmd)->get_task_tag(cmd));
+#endif
+	CMD_TFO(cmd)->queue_status(cmd);
+}
+
+/*	transport_generic_do_tmr():
+ *
+ *
+ */
+int transport_generic_do_tmr(se_cmd_t *cmd)
+{
+	se_cmd_t *ref_cmd;
+	se_device_t *dev = SE_DEV(cmd);
+	se_tmr_req_t *tmr = cmd->se_tmr_req;
+	int ret;
+
+	switch (tmr->function) {
+	case ABORT_TASK:
+		ref_cmd = tmr->ref_cmd;
+		tmr->response = TMR_FUNCTION_REJECTED;
+		break;
+	case ABORT_TASK_SET:
+	case CLEAR_ACA:
+	case CLEAR_TASK_SET:
+		tmr->response = TMR_TASK_MGMT_FUNCTION_NOT_SUPPORTED;
+		break;
+	case LUN_RESET:
+		ret = core_tmr_lun_reset(dev, tmr, NULL, NULL);
+		tmr->response = (!ret) ? TMR_FUNCTION_COMPLETE :
+					 TMR_FUNCTION_REJECTED;
+		break;
+#if 0
+	case TARGET_WARM_RESET:
+		transport_generic_host_reset(dev->se_hba);
+		tmr->response = TMR_FUNCTION_REJECTED;
+		break;
+	case TARGET_COLD_RESET:
+		transport_generic_host_reset(dev->se_hba);
+		transport_generic_cold_reset(dev->se_hba);
+		tmr->response = TMR_FUNCTION_REJECTED;
+		break;
+#endif
+	default:
+		printk(KERN_ERR "Uknown TMR function: 0x%02x.\n",
+				tmr->function);
+		tmr->response = TMR_FUNCTION_REJECTED;
+		break;
+	}
+
+	cmd->t_state = TRANSPORT_ISTATE_PROCESSING;
+	CMD_TFO(cmd)->queue_tm_rsp(cmd);
+
+	return 0;
+}
+
+/*
+ *	Called with spin_lock_irq(&dev->execute_task_lock); held
+ *
+ */
+se_task_t *transport_get_task_from_state_list(se_device_t *dev)
+{
+	se_task_t *task;
+
+	if (list_empty(&dev->state_task_list))
+		return NULL;
+
+	list_for_each_entry(task, &dev->state_task_list, t_state_list)
+		break;
+
+	list_del(&task->t_state_list);
+	atomic_set(&task->task_state_active, 0);
+
+	return task;
+}
+
+static void transport_processing_shutdown(se_device_t *dev)
+{
+	se_cmd_t *cmd;
+	se_queue_req_t *qr;
+	se_task_t *task;
+	u8 state;
+	unsigned long flags;
+	/*
+	 * Empty the se_device_t's se_task_t state list.
+	 */
+	spin_lock_irqsave(&dev->execute_task_lock, flags);
+	while ((task = transport_get_task_from_state_list(dev))) {
+		if (!(TASK_CMD(task))) {
+			printk(KERN_ERR "TASK_CMD(task) is NULL!\n");
+			continue;
+		}
+		cmd = TASK_CMD(task);
+
+		if (!T_TASK(cmd)) {
+			printk(KERN_ERR "T_TASK(cmd) is NULL for task: %p cmd:"
+				" %p ITT: 0x%08x\n", task, cmd,
+				CMD_TFO(cmd)->get_task_tag(cmd));
+			continue;
+		}
+		spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+		spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+
+		DEBUG_DO("PT: cmd: %p task: %p ITT/CmdSN: 0x%08x/0x%08x,"
+			" i_state/def_i_state: %d/%d, t_state/def_t_state:"
+			" %d/%d cdb: 0x%02x\n", cmd, task,
+			CMD_TFO(cmd)->get_task_tag(cmd), cmd->cmd_sn,
+			CMD_TFO(cmd)->get_cmd_state(cmd), cmd->deferred_i_state,
+			cmd->t_state, cmd->deferred_t_state,
+			T_TASK(cmd)->t_task_cdb[0]);
+		DEBUG_DO("PT: ITT[0x%08x] - t_task_cdbs: %d t_task_cdbs_left:"
+			" %d t_task_cdbs_sent: %d -- t_transport_active: %d"
+			" t_transport_stop: %d t_transport_sent: %d\n",
+			CMD_TFO(cmd)->get_task_tag(cmd),
+			T_TASK(cmd)->t_task_cdbs,
+			atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
+			atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
+			atomic_read(&T_TASK(cmd)->t_transport_active),
+			atomic_read(&T_TASK(cmd)->t_transport_stop),
+			atomic_read(&T_TASK(cmd)->t_transport_sent));
+
+		if (atomic_read(&task->task_active)) {
+			atomic_set(&task->task_stop, 1);
+			spin_unlock_irqrestore(
+				&T_TASK(cmd)->t_state_lock, flags);
+
+			DEBUG_DO("Waiting for task: %p to shutdown for dev:"
+				" %p\n", task, dev);
+			wait_for_completion(&task->task_stop_comp);
+			DEBUG_DO("Completed task: %p shutdown for dev: %p\n",
+				task, dev);
+
+			spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+			atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+
+			atomic_set(&task->task_active, 0);
+			atomic_set(&task->task_stop, 0);
+		}
+		__transport_stop_task_timer(task, &flags);
+
+		if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_ex_left))) {
+			spin_unlock_irqrestore(
+					&T_TASK(cmd)->t_state_lock, flags);
+
+			DEBUG_DO("Skipping task: %p, dev: %p for"
+				" t_task_cdbs_ex_left: %d\n", task, dev,
+				atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left));
+
+			spin_lock_irqsave(&dev->execute_task_lock, flags);
+			continue;
+		}
+
+		if (atomic_read(&T_TASK(cmd)->t_transport_active)) {
+			DEBUG_DO("got t_transport_active = 1 for task: %p, dev:"
+					" %p\n", task, dev);
+
+			if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+				spin_unlock_irqrestore(
+					&T_TASK(cmd)->t_state_lock, flags);
+				transport_send_check_condition_and_sense(
+					cmd, LOGICAL_UNIT_COMMUNICATION_FAILURE,
+					0);
+				transport_remove_cmd_from_queue(cmd,
+					CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+						cmd->se_orig_obj_ptr));
+
+				transport_lun_remove_cmd(cmd);
+				if (!(transport_cmd_check_stop(cmd, 1, 0)))
+					transport_passthrough_check_stop(cmd);
+			} else {
+				spin_unlock_irqrestore(
+					&T_TASK(cmd)->t_state_lock, flags);
+
+				transport_remove_cmd_from_queue(cmd,
+					CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+						cmd->se_orig_obj_ptr));
+
+				transport_lun_remove_cmd(cmd);
+
+				if (!(transport_cmd_check_stop(cmd, 1, 0)))
+					transport_passthrough_check_stop(cmd);
+				else
+					transport_generic_remove(cmd, 0, 0);
+			}
+
+			spin_lock_irqsave(&dev->execute_task_lock, flags);
+			continue;
+		}
+		DEBUG_DO("Got t_transport_active = 0 for task: %p, dev: %p\n",
+				task, dev);
+
+		if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+			spin_unlock_irqrestore(
+				&T_TASK(cmd)->t_state_lock, flags);
+			transport_send_check_condition_and_sense(cmd,
+				LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
+			transport_remove_cmd_from_queue(cmd,
+				CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+					cmd->se_orig_obj_ptr));
+
+			transport_lun_remove_cmd(cmd);
+			if (!(transport_cmd_check_stop(cmd, 1, 0)))
+				transport_passthrough_check_stop(cmd);
+		} else {
+			spin_unlock_irqrestore(
+				&T_TASK(cmd)->t_state_lock, flags);
+
+			transport_remove_cmd_from_queue(cmd,
+				CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+					cmd->se_orig_obj_ptr));
+			transport_lun_remove_cmd(cmd);
+
+			if (!(transport_cmd_check_stop(cmd, 1, 0)))
+				transport_passthrough_check_stop(cmd);
+			else
+				transport_generic_remove(cmd, 0, 0);
+		}
+
+		spin_lock_irqsave(&dev->execute_task_lock, flags);
+	}
+	spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+	/*
+	 * Empty the se_device_t's se_cmd_t list.
+	 */
+	spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+	while ((qr = __transport_get_qr_from_queue(dev->dev_queue_obj))) {
+		spin_unlock_irqrestore(
+				&dev->dev_queue_obj->cmd_queue_lock, flags);
+		cmd = (se_cmd_t *)qr->cmd;
+		state = qr->state;
+		kfree(qr);
+
+		DEBUG_DO("From Device Queue: cmd: %p t_state: %d\n",
+				cmd, state);
+
+		if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+			transport_send_check_condition_and_sense(cmd,
+				LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
+
+			transport_lun_remove_cmd(cmd);
+			if (!(transport_cmd_check_stop(cmd, 1, 0)))
+				transport_passthrough_check_stop(cmd);
+		} else {
+			transport_lun_remove_cmd(cmd);
+
+			if (!(transport_cmd_check_stop(cmd, 1, 0)))
+				transport_passthrough_check_stop(cmd);
+			else
+				transport_generic_remove(cmd, 0, 0);
+		}
+		spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+	}
+	spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock, flags);
+}
+
+/*	transport_processing_thread():
+ *
+ *
+ */
+static int transport_processing_thread(void *param)
+{
+	int ret, t_state;
+	se_cmd_t *cmd;
+	se_device_t *dev = (se_device_t *) param;
+	se_queue_req_t *qr;
+
+	current->policy = SCHED_NORMAL;
+	set_user_nice(current, -20);
+	spin_lock_irq(&current->sighand->siglock);
+	siginitsetinv(&current->blocked, SHUTDOWN_SIGS);
+	recalc_sigpending();
+	spin_unlock_irq(&current->sighand->siglock);
+
+	complete(&dev->dev_queue_obj->thread_create_comp);
+
+	while (!(kthread_should_stop())) {
+		ret = wait_event_interruptible(dev->dev_queue_obj->thread_wq,
+				atomic_read(&dev->dev_queue_obj->queue_cnt) ||
+				kthread_should_stop());
+		if (ret < 0)
+			goto out;
+
+		spin_lock(&dev->dev_status_lock);
+		if (dev->dev_status & TRANSPORT_DEVICE_SHUTDOWN) {
+			spin_unlock(&dev->dev_status_lock);
+			transport_processing_shutdown(dev);
+			continue;
+		}
+		spin_unlock(&dev->dev_status_lock);
+
+get_cmd:
+		__transport_execute_tasks(dev);
+
+		qr = transport_get_qr_from_queue(dev->dev_queue_obj);
+		if (!(qr))
+			continue;
+
+		cmd = (se_cmd_t *)qr->cmd;
+		t_state = qr->state;
+		kfree(qr);
+
+		switch (t_state) {
+		case TRANSPORT_NEW_CMD:
+			ret = transport_generic_new_cmd(cmd);
+			if (ret < 0) {
+				cmd->transport_error_status = ret;
+				transport_generic_request_failure(cmd, NULL,
+					0, (cmd->data_direction !=
+					 SE_DIRECTION_WRITE));
+			}
+			break;
+		case TRANSPORT_PROCESS_WRITE:
+			transport_generic_process_write(cmd);
+			break;
+		case TRANSPORT_COMPLETE_OK:
+			transport_stop_all_task_timers(cmd);
+			transport_generic_complete_ok(cmd);
+			break;
+		case TRANSPORT_REMOVE:
+			transport_generic_remove(cmd, 1, 0);
+			break;
+		case TRANSPORT_PROCESS_TMR:
+			transport_generic_do_tmr(cmd);
+			break;
+		case TRANSPORT_COMPLETE_FAILURE:
+			transport_generic_request_failure(cmd, NULL, 1, 1);
+			break;
+		case TRANSPORT_COMPLETE_TIMEOUT:
+			transport_stop_all_task_timers(cmd);
+			transport_generic_request_timeout(cmd);
+			break;
+		default:
+			printk(KERN_ERR "Unknown t_state: %d deferred_t_state:"
+				" %d for ITT: 0x%08x i_state: %d on SE LUN:"
+				" %u\n", t_state, cmd->deferred_t_state,
+				CMD_TFO(cmd)->get_task_tag(cmd),
+				CMD_TFO(cmd)->get_cmd_state(cmd),
+				SE_LUN(cmd)->unpacked_lun);
+			BUG();
+		}
+
+		goto get_cmd;
+	}
+
+out:
+	transport_release_all_cmds(dev);
+	dev->process_thread = NULL;
+	complete(&dev->dev_queue_obj->thread_done_comp);
+	return 0;
+}
diff --git a/include/target/target_core_transport.h b/include/target/target_core_transport.h
new file mode 100644
index 0000000..fcce670
--- /dev/null
+++ b/include/target/target_core_transport.h
@@ -0,0 +1,529 @@
+/*******************************************************************************
+ * Filename:  target_core_transport.h
+ *
+ * This file contains the iSCSI Target Generic DAS Transport Layer definitions.
+ *
+ * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
+ * Copyright (c) 2005, 2006, 2007 SBE, Inc.
+ * Copyright (c) 2007-2009 Rising Tide Software, Inc.
+ * Copyright (c) 2008-2009 Linux-iSCSI.org
+ *
+ * Nicholas A. Bellinger <nab@...nel.org>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+ *
+ ******************************************************************************/
+
+
+#ifndef TARGET_CORE_TRANSPORT_H
+#define TARGET_CORE_TRANSPORT_H
+
+#define TARGET_CORE_VERSION			TARGET_CORE_MOD_VERSION
+
+/* Attempts before moving from SHORT to LONG */
+#define PYX_TRANSPORT_WINDOW_CLOSED_THRESHOLD	3
+#define PYX_TRANSPORT_WINDOW_CLOSED_WAIT_SHORT	3  /* In milliseconds */
+#define PYX_TRANSPORT_WINDOW_CLOSED_WAIT_LONG	10 /* In milliseconds */
+
+#define PYX_TRANSPORT_STATUS_INTERVAL		5 /* In seconds */
+
+#define PYX_TRANSPORT_SENT_TO_TRANSPORT		0
+#define PYX_TRANSPORT_WRITE_PENDING		1
+
+#define PYX_TRANSPORT_UNKNOWN_SAM_OPCODE	-1
+#define PYX_TRANSPORT_HBA_QUEUE_FULL		-2
+#define PYX_TRANSPORT_REQ_TOO_MANY_SECTORS	-3
+#define PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES	-4
+#define PYX_TRANSPORT_INVALID_CDB_FIELD		-5
+#define PYX_TRANSPORT_INVALID_PARAMETER_LIST	-6
+#define PYX_TRANSPORT_LU_COMM_FAILURE		-7
+#define PYX_TRANSPORT_UNKNOWN_MODE_PAGE		-8
+#define PYX_TRANSPORT_WRITE_PROTECTED		-9
+#define PYX_TRANSPORT_TASK_TIMEOUT		-10
+#define PYX_TRANSPORT_RESERVATION_CONFLICT	-11
+
+#ifndef SAM_STAT_RESERVATION_CONFLICT
+#define SAM_STAT_RESERVATION_CONFLICT		0x18
+#endif
+
+#define TRANSPORT_PLUGIN_FREE			0
+#define TRANSPORT_PLUGIN_REGISTERED		1
+
+#define TRANSPORT_PLUGIN_PHBA_PDEV		1
+#define TRANSPORT_PLUGIN_VHBA_PDEV		2
+#define TRANSPORT_PLUGIN_VHBA_VDEV		3
+
+/* For SE OBJ Plugins, in seconds */
+#define TRANSPORT_TIMEOUT_TUR			10
+#define TRANSPORT_TIMEOUT_TYPE_DISK		60
+#define TRANSPORT_TIMEOUT_TYPE_ROM		120
+#define TRANSPORT_TIMEOUT_TYPE_TAPE		600
+#define TRANSPORT_TIMEOUT_TYPE_OTHER		300
+
+/* For se_task->task_state_flags */
+#define TSF_EXCEPTION_CLEARED			0x01
+
+/*
+ * se_subsystem_dev_t->su_dev_flags
+*/
+#define SDF_FIRMWARE_VPD_UNIT_SERIAL		0x00000001
+#define SDF_EMULATED_VPD_UNIT_SERIAL		0x00000002
+#define SDF_USING_UDEV_PATH			0x00000004
+
+/*
+ * se_device_t->dev_flags
+ */
+#define DF_READAHEAD_ACTIVE                     0x00000001
+#define DF_TRANSPORT_DMA_ALLOC			0x00000002
+#define DF_TRANSPORT_BUF_ALLOC			0x00000004
+#define DF_DEV_DEBUG				0x00000008
+#define DF_CLAIMED_BLOCKDEV			0x00000010
+#define DF_PERSISTENT_CLAIMED_BLOCKDEV		0x00000020
+#define DF_READ_ONLY				0x00000040
+#define DF_SPC3_PERSISTENT_RESERVE		0x00000080
+#define DF_SPC2_RESERVATIONS			0x00000100
+
+/* se_dev_attrib_t sanity values */
+/* 10 Minutes, see transport_get_default_task_timeout()  */
+#define DA_TASK_TIMEOUT_MAX			600
+/* Emulation for UNIT ATTENTION Interlock Control */
+#define DA_EMULATE_UA_INTLLCK_CTRL		0
+/* Emulation for TASK_ABORTED status (TAS) by default */
+#define DA_EMULATE_TAS				1
+/* No Emulation for PSCSI by default */
+#define DA_EMULATE_RESERVATIONS			0
+/* No Emulation for PSCSI by default */
+#define DA_EMULATE_ALUA				0
+#define DA_STATUS_MAX_SECTORS_MIN		16
+#define DA_STATUS_MAX_SECTORS_MAX		8192
+
+#define SE_MODE_PAGE_BUF			512
+
+#define MOD_MAX_SECTORS(ms, bs)			(ms % (PAGE_SIZE / bs))
+
+struct se_mem_s;
+
+extern int init_se_global(void);
+extern void release_se_global(void);
+#ifdef DEBUG_DEV
+extern int __iscsi_debug_dev(se_device_t *);
+#endif
+extern unsigned char *transport_get_iqn_sn(void);
+extern void transport_init_queue_obj(struct se_queue_obj_s *);
+extern void transport_load_plugins(void);
+extern struct se_plugin_s *transport_core_get_plugin_by_name(const char *name);
+extern void transport_check_dev_params_delim(char *, char **);
+extern struct se_session_s *transport_init_session(void);
+extern void __transport_register_session(struct se_portal_group_s *,
+					struct se_node_acl_s *,
+					struct se_session_s *, void *);
+extern void transport_register_session(struct se_portal_group_s *,
+					struct se_node_acl_s *,
+					struct se_session_s *, void *);
+extern void transport_free_session(struct se_session_s *);
+extern void transport_deregister_session_configfs(struct se_session_s *);
+extern void transport_deregister_session(struct se_session_s *);
+extern void transport_task_dev_remove_state(struct se_task_s *,
+						struct se_device_s *);
+extern void transport_cmd_finish_abort(struct se_cmd_s *, int);
+extern void transport_cmd_finish_abort_tmr(struct se_cmd_s *);
+extern int transport_add_cmd_to_queue(struct se_cmd_s *,
+					struct se_queue_obj_s *, u8);
+extern struct se_queue_req_s *__transport_get_qr_from_queue(
+					struct se_queue_obj_s *);
+extern void transport_remove_cmd_from_queue(struct se_cmd_s *,
+					    struct se_queue_obj_s *);
+extern void transport_complete_cmd(se_cmd_t *, int);
+extern void transport_complete_task(struct se_task_s *, int);
+extern void transport_add_task_to_execute_queue(struct se_task_s *,
+						struct se_task_s *,
+						struct se_device_s *);
+extern void transport_add_tasks_from_cmd(struct se_cmd_s *);
+extern struct se_task_s *transport_get_task_from_execute_queue(
+						struct se_device_s *);
+extern se_queue_req_t *transport_get_qr_from_queue(struct se_queue_obj_s *);
+extern int transport_check_device_tcq(se_device_t *, u32, u32);
+unsigned char *transport_dump_cmd_direction (struct se_cmd_s *);
+extern void transport_dump_dev_state(struct se_device_s *, char *, int *);
+extern void transport_dump_dev_info(struct se_device_s *, struct se_lun_s *,
+					unsigned long long, char *, int *);
+extern void transport_dump_vpd_proto_id(struct t10_vpd_s *,
+					unsigned char *, int);
+extern int transport_dump_vpd_assoc(struct t10_vpd_s *,
+					unsigned char *, int);
+extern int transport_dump_vpd_ident_type(struct t10_vpd_s *,
+					unsigned char *, int);
+extern int transport_dump_vpd_ident(struct t10_vpd_s *,
+					unsigned char *, int);
+extern int transport_rescan_evpd_device_ident(struct se_device_s *);
+extern se_device_t *transport_add_device_to_core_hba(se_hba_t *,
+					struct se_subsystem_api_s *,
+					struct se_subsystem_dev_s *, u32,
+					void *);
+extern int transport_generic_activate_device(se_device_t *);
+extern void transport_generic_deactivate_device(se_device_t *);
+extern int transport_generic_claim_phydevice(se_device_t *);
+extern void transport_generic_release_phydevice(se_device_t *, int);
+extern void transport_generic_free_device(se_device_t *);
+extern int transport_allocate_iovecs_for_cmd(struct se_cmd_s *, u32);
+extern int transport_generic_obj_start(struct se_transform_info_s *,
+					struct se_obj_lun_type_s *, void *,
+					unsigned long long);
+extern void transport_device_setup_cmd(se_cmd_t *);
+extern int transport_check_alloc_task_attr(se_cmd_t *);
+extern se_cmd_t *transport_alloc_se_cmd(struct target_core_fabric_ops *,
+					struct se_session_s *, void *,
+					u32, int, int);
+extern void transport_free_se_cmd(struct se_cmd_s *);
+extern int transport_generic_allocate_tasks(se_cmd_t *, unsigned char *);
+extern int transport_generic_handle_cdb(se_cmd_t *);
+extern int transport_generic_handle_data(se_cmd_t *);
+extern int transport_generic_handle_tmr(se_cmd_t *);
+extern void transport_stop_tasks_for_cmd(struct se_cmd_s *);
+extern void transport_generic_request_failure(se_cmd_t *, se_device_t *,
+						int, int);
+extern void transport_direct_request_timeout(se_cmd_t *);
+extern void transport_generic_request_timeout(se_cmd_t *);
+extern int transport_generic_allocate_buf(se_cmd_t *, u32, u32);
+extern int __transport_execute_tasks(struct se_device_s *);
+extern void transport_new_cmd_failure(struct se_cmd_s *);
+extern u32 transport_get_default_task_timeout(struct se_device_s *);
+extern void transport_set_supported_SAM_opcode(struct se_cmd_s *);
+extern void transport_start_task_timer(struct se_task_s *);
+extern void __transport_stop_task_timer(struct se_task_s *, unsigned long *);
+extern void transport_stop_task_timer(struct se_task_s *);
+extern void transport_stop_all_task_timers(struct se_cmd_s *);
+extern int transport_execute_tasks(struct se_cmd_s *);
+extern int transport_generic_emulate_inquiry(struct se_cmd_s *, unsigned char,
+					unsigned char *, unsigned char *,
+					unsigned char *);
+extern int transport_generic_emulate_readcapacity(struct se_cmd_s *, u32);
+extern int transport_generic_emulate_readcapacity_16(struct se_cmd_s *,
+							unsigned long long);
+extern int transport_generic_emulate_modesense(struct se_cmd_s *,
+						unsigned char *,
+						unsigned char *, int, int);
+extern int transport_generic_emulate_request_sense(struct se_cmd_s *,
+						   unsigned char *);
+extern int transport_get_sense_data(struct se_cmd_s *);
+extern se_cmd_t *transport_allocate_passthrough(unsigned char *, int, u32,
+						void *, u32, u32,
+						struct se_obj_lun_type_s *,
+						void *);
+extern void transport_passthrough_release(se_cmd_t *);
+extern int transport_passthrough_complete(se_cmd_t *);
+extern int transport_generic_passthrough_async(se_cmd_t *cmd,
+				void(*callback)(se_cmd_t *cmd,
+				void *callback_arg, int complete_status),
+				void *callback_arg);
+extern int transport_generic_passthrough(se_cmd_t *);
+extern void transport_complete_task_attr(se_cmd_t *);
+extern void transport_generic_complete_ok(se_cmd_t *);
+extern void transport_free_dev_tasks(se_cmd_t *);
+extern void transport_release_tasks(se_cmd_t *);
+extern void transport_release_fe_cmd(se_cmd_t *);
+extern int transport_generic_remove(se_cmd_t *, int, int);
+extern int transport_lun_wait_for_tasks(se_cmd_t *, se_lun_t *);
+extern void transport_clear_lun_from_sessions(se_lun_t *);
+extern int transport_check_aborted_status(se_cmd_t *, int);
+extern int transport_get_sense_codes(se_cmd_t *, u8 *, u8 *);
+extern int transport_set_sense_codes(se_cmd_t *, u8, u8);
+extern int transport_send_check_condition_and_sense(se_cmd_t *, u8, int);
+extern void transport_send_task_abort(struct se_cmd_s *);
+extern void transport_release_cmd_to_pool(se_cmd_t *);
+extern void transport_generic_free_cmd(se_cmd_t *, int, int, int);
+extern void transport_generic_wait_for_cmds(se_cmd_t *, int);
+extern int transport_generic_do_transform(struct se_cmd_s *,
+					struct se_transform_info_s *);
+extern int transport_get_sectors(struct se_cmd_s *, struct se_obj_lun_type_s *,
+					void *);
+extern int transport_new_cmd_obj(struct se_cmd_s *,
+				struct se_transform_info_s *,
+				struct se_obj_lun_type_s *, void *, int);
+extern unsigned char *transport_get_vaddr(struct se_mem_s *);
+extern struct list_head *transport_init_se_mem_list(void);
+extern void transport_free_se_mem_list(struct list_head *);
+extern int transport_generic_get_mem(struct se_cmd_s *, u32, u32);
+extern u32 transport_calc_sg_num(struct se_task_s *, struct se_mem_s *, u32);
+extern int transport_map_sg_to_mem(struct se_task_s *, struct list_head *,
+					void *, struct se_mem_s *,
+					struct se_mem_s **, u32 *, u32 *);
+extern int transport_map_mem_to_mem(struct se_task_s *, struct list_head *,
+					void *, struct se_mem_s *,
+					struct se_mem_s **, u32 *, u32 *);
+extern int transport_map_mem_to_sg(struct se_task_s *, struct list_head *,
+					void *, struct se_mem_s *,
+					struct se_mem_s **, u32 *, u32 *);
+extern u32 transport_generic_get_cdb_count(struct se_cmd_s *,
+					struct se_transform_info_s *,
+					struct se_obj_lun_type_s *, void *,
+					unsigned long long, u32,
+					struct se_mem_s *, struct se_mem_s **,
+					u32 *);
+extern int transport_generic_new_cmd(se_cmd_t *);
+extern void transport_generic_process_write(se_cmd_t *);
+extern int transport_generic_do_tmr(se_cmd_t *);
+
+/*
+ * Each se_transport_task_t can have N number of possible se_task_t's
+ * for the storage transport(s) to possibly execute.
+ * Used primarily for splitting up CDBs that exceed the physical storage
+ * HBA's maximum sector count per task.
+ */
+typedef struct se_mem_s {
+	struct page	*se_page;
+	u32		se_len;
+	u32		se_off;
+	struct list_head se_list;
+} ____cacheline_aligned se_mem_t;
+
+/*
+ * Each type of DAS transport that uses the generic command sequencer needs
+ * each of the following function pointers set.
+ */
+typedef struct se_subsystem_spc_s {
+	int (*inquiry)(se_task_t *, u32);
+	int (*none)(se_task_t *, u32);
+	int (*read_non_SG)(se_task_t *, u32);
+	int (*read_SG)(se_task_t *, u32);
+	int (*write_non_SG)(se_task_t *, u32);
+	int (*write_SG)(se_task_t *, u32);
+} se_subsystem_spc_t;
+
+/*
+ * 	Each type of disk transport supported MUST have a template defined
+ *	within its .h file.
+ */
+typedef struct se_subsystem_api_s {
+	/*
+	 * The Name. :-)
+	 */
+	char name[16];
+	/*
+	 * Plugin Type.
+	 */
+	u8 type;
+	/*
+	 * Transport Type.
+	 */
+	u8 transport_type;
+	/*
+	 * attach_hba():
+	 */
+	int (*attach_hba)(se_hba_t *, u32);
+	/*
+	 * detach_hba():
+	 */
+	int (*detach_hba)(struct se_hba_s *);
+	/*
+	 * claim_phydevice(): Only for Physical HBAs
+	 */
+	int (*claim_phydevice)(struct se_hba_s *, struct se_device_s *);
+	/*
+	 * allocate_virtdevice():
+	 */
+	void *(*allocate_virtdevice)(struct se_hba_s *, const char *);
+	/*
+	 * create_virtdevice(): Only for Virtual HBAs
+	 */
+	se_device_t *(*create_virtdevice)(struct se_hba_s *,
+				struct se_subsystem_dev_s *, void *);
+	/*
+	 * activate_device():
+	 */
+	int (*activate_device)(struct se_device_s *);
+	/*
+	 * deactivate_device():
+	 */
+	void (*deactivate_device)(struct se_device_s *);
+	/*
+	 * release_phydevice():
+	 */
+	int (*release_phydevice)(struct se_device_s *);
+	/*
+	 * free_device():
+	 */
+	void (*free_device)(void *);
+	/*
+	 * cmd_sequencer():
+	 *
+	 * Use transport_generic_cmd_sequencer() for majority of DAS transport
+	 * drivers with a scsi_transport_spc_t struct as mentioned below.
+	 * Provided out of convenience.
+	 */
+	int (*cmd_sequencer)(se_cmd_t *cmd);
+	/*
+	 * do_tmr():
+	 *
+	 * Use transport_do_tmr() for majority of DAS transport drivers.
+	 * Provided out of convenience.
+	 */
+	int (*do_tmr)(se_cmd_t *cmd);
+	/*
+	 * transport_complete():
+	 *
+	 * Use transport_generic_complete() for majority of DAS transport
+	 * drivers.  Provided out of convenience.
+	 */
+	int (*transport_complete)(se_task_t *task);
+	/*
+	 * allocate_request():
+	 */
+	void *(*allocate_request)(se_task_t *, se_device_t *);
+	/*
+	 * allocate_buf():
+	 */
+	int (*allocate_buf)(se_cmd_t *, u32, u32);
+	/*
+	 * allocate_DMA();
+	 */
+	int (*allocate_DMA)(se_cmd_t *, u32, u32);
+	/*
+	 * free_buf():
+	 */
+	void (*free_buf)(se_cmd_t *);
+	/*
+	 * free_DMA():
+	 */
+	void (*free_DMA)(se_cmd_t *);
+	/*
+	 * do_task():
+	 */
+	int (*do_task)(se_task_t *);
+	/*
+	 * free_task():
+	 */
+	void (*free_task)(se_task_t *);
+	/*
+	 * check_configfs_dev_params():
+	 */
+	ssize_t (*check_configfs_dev_params)(se_hba_t *, se_subsystem_dev_t *);
+	/*
+	 * set_configfs_dev_params():
+	 */
+	ssize_t (*set_configfs_dev_params)(se_hba_t *, se_subsystem_dev_t *,
+						const char *, ssize_t);
+	/*
+	 * show_configfs_dev_params():
+	 */
+	ssize_t (*show_configfs_dev_params)(se_hba_t *, se_subsystem_dev_t *,
+						char *);
+	/*
+	 * create_virtdevice_from-fd():
+	 */
+	se_device_t *(*create_virtdevice_from_fd)(se_subsystem_dev_t *,
+						const char *);
+	/*
+	 * plugin_init():
+	 */
+	int (*plugin_init)(void);
+	/*
+	 * plugin_free():
+	 */
+	void (*plugin_free)(void);
+	/*
+	 * get_plugin_info():
+	 */
+	void (*get_plugin_info)(void *, char *, int *);
+	/*
+	 * get_hba_info():
+	 */
+	void (*get_hba_info)(se_hba_t *, char *, int *);
+	/*
+	 * get_dev_info():
+	 */
+	void (*get_dev_info)(se_device_t *, char *, int *);
+	/*
+	 * check_lba():
+	 */
+	int (*check_lba)(unsigned long long lba, se_device_t *);
+	/*
+	 * check_for_SG():
+	 */
+	int (*check_for_SG)(se_task_t *);
+	/*
+	 * get_cdb():
+	 */
+	unsigned char *(*get_cdb)(se_task_t *);
+	/*
+	 * get_blocksize():
+	 */
+	u32 (*get_blocksize)(se_device_t *);
+	/*
+	 * get_device_rev():
+	 */
+	u32 (*get_device_rev)(se_device_t *);
+	/*
+	 * get_device_type():
+	 */
+	u32 (*get_device_type)(se_device_t *);
+	/*
+	 * get_dma_length():
+	 */
+	u32 (*get_dma_length)(u32, se_device_t *);
+	/*
+	 * get_max_cdbs():
+	 */
+	u32 (*get_max_cdbs)(se_device_t *);
+	/*
+	 * get_max_sectors():
+	 */
+	 u32 (*get_max_sectors)(se_device_t *);
+	/*
+	 * get_queue_depth():
+	 *
+	 */
+	u32 (*get_queue_depth)(se_device_t *);
+	/*
+	 * get_max_queue_depth():
+	 */
+	u32 (*get_max_queue_depth)(se_device_t *);
+	/*
+	 * do_se_mem_map():
+	 */
+	int (*do_se_mem_map)(se_task_t *, struct list_head *, void *,
+				se_mem_t *, se_mem_t **, u32 *, u32 *);
+	/*
+	 * get_sense_buffer():
+	 */
+	unsigned char *(*get_sense_buffer)(se_task_t *);
+	/*
+	 * map_task_to_SG():
+	 */
+	void (*map_task_to_SG)(se_task_t *);
+	/*
+	 * set_iovec_ptrs():
+	 */
+	int (*set_iovec_ptrs)(se_map_sg_t *, se_unmap_sg_t *);
+	/*
+	 * write_pending():
+	 */
+	int (*write_pending)(se_task_t *);
+	/*
+	 * se_subsystem_spc_t structure:
+	 *
+	 * Contains function pointers of SPC opcodes to call from the generic
+	 * command sequencer into a transport driver if the generic command
+	 * sequencer is used. (ie: cmd_sequencer() is NULL)
+	 */
+	se_subsystem_spc_t *spc;
+} ____cacheline_aligned se_subsystem_api_t;
+
+#define TRANSPORT(dev)		((dev)->transport)
+#define TRANSPORT_SPC(dev)	((dev)->transport->spc)
+#define HBA_TRANSPORT(hba)	((hba)->transport)
+
+#endif /* TARGET_CORE_TRANSPORT_H */
+
diff --git a/include/target/target_core_transport_plugin.h b/include/target/target_core_transport_plugin.h
new file mode 100644
index 0000000..7e46360
--- /dev/null
+++ b/include/target/target_core_transport_plugin.h
@@ -0,0 +1,67 @@
+/*******************************************************************************
+ * Filename:  target_core_transport_plugin.h
+ *
+ * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
+ * Copyright (c) 2005, 2006, 2007 SBE, Inc.
+ * Copyright (c) 2007-2009 Rising Tide Software, Inc.
+ * Copyright (c) 2008-2009 Linux-iSCSI.org
+ *
+ * Nicholas A. Bellinger <nab@...nel.org>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+ *
+ ******************************************************************************/
+
+
+#ifndef _TARGET_CORE_TRANSPORT_PLUGIN_H_
+#define _TARGET_CORE_TRANSPORT_PLUGIN_H_
+
+#ifdef PARALLEL_SCSI
+#define PSCSI_INCLUDE_STRUCTS
+#include <target/target_core_pscsi.h>
+#undef PSCSI_INCLUDE_STRUCTS
+extern se_subsystem_api_t pscsi_template;
+#endif /* PARALLEL_SCSI */
+
+#ifdef STGT_PLUGIN
+#define STGT_INCLUDE_STRUCTS
+#include <target/target_core_stgt.h>
+#undef STGT_INCLUDE_STRUCTS
+extern se_subsystem_api_t stgt_template;
+#endif /* STGT_PLUGIN */
+
+#ifdef PYX_IBLOCK
+#define IBLOCK_INCLUDE_STRUCTS
+#include <target/target_core_iblock.h>
+#undef IBLOCK_INCLUDE_STRUCTS
+extern se_subsystem_api_t iblock_template;
+#endif /* PYX_IBLOCK */
+
+#ifdef PYX_RAMDISK
+#define RD_INCLUDE_STRUCTS
+#include <target/target_core_rd.h>
+#undef RD_INCLUDE_STRUCTS
+extern se_subsystem_api_t rd_dr_template;
+extern se_subsystem_api_t rd_mcp_template;
+#endif /* PYX_RAMDISK */
+
+#ifdef PYX_FILEIO
+#define FD_INCLUDE_STRUCTS
+#include <target/target_core_file.h>
+#undef FD_INCLUDE_STRUCTS
+extern se_subsystem_api_t fileio_template;
+#endif /* PYX_FILEIO */
+
+#endif    /*** _TARGET_CORE_TRANSPORT_PLUGIN_H_ ***/



--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ