[<prev] [next>] [day] [month] [year] [list]
Message-Id: <1238904141.4250.759.camel@haakon2.linux-iscsi.org>
Date: Sat, 04 Apr 2009 21:02:21 -0700
From: "Nicholas A. Bellinger" <nab@...ux-iscsi.org>
To: LKML <linux-kernel@...r.kernel.org>,
linux-scsi <linux-scsi@...r.kernel.org>
Cc: James Bottomley <James.Bottomley@...senPartnership.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Hannes Reinecke <hare@...e.de>,
FUJITA Tomonori <fujita.tomonori@....ntt.co.jp>,
Mike Christie <michaelc@...wisc.edu>,
"Martin K. Petersen" <martin.petersen@...cle.com>,
Ming Zhang <blackmagic02881@...il.com>,
Philipp Reisner <philipp.reisner@...bit.com>,
Linus Torvalds <torvalds@...ux-foundation.org>
Subject: [RFC PATCH 18/19] Target_Core_Mod Generic Transport Core
The patch adds the core generic target engine functionality including processing of
se_cmd_t, generation of se_task_ts, control patch CDB emulation, sense data, memory allocation,
and memory mapping from fabric module allocated memory using linked list struct scatterlist
style memory based on underlying storage object hardware.
Signed-off-by: Nicholas A. Bellinger <nab@...ux-iscsi.org>
diff --git a/drivers/target/target_core_transport.c b/drivers/target/target_core_transport.c
new file mode 100644
index 0000000..134d5a0
--- /dev/null
+++ b/drivers/target/target_core_transport.c
@@ -0,0 +1,7537 @@
+/*******************************************************************************
+ * Filename: target_core_transport.c
+ *
+ * This file contains the Generic Target Layer Core.
+ *
+ * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
+ * Copyright (c) 2005, 2006, 2007 SBE, Inc.
+ * Copyright (c) 2007-2009 Rising Tide Software, Inc.
+ * Copyright (c) 2008-2009 Linux-iSCSI.org
+ *
+ * Nicholas A. Bellinger <nab@...nel.org>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+ *
+ ******************************************************************************/
+
+
+#define TARGET_CORE_TRANSPORT_C
+
+#include <linux/version.h>
+#include <linux/net.h>
+#include <linux/delay.h>
+#include <linux/string.h>
+#include <linux/timer.h>
+#include <linux/slab.h>
+#include <linux/blkdev.h>
+#include <linux/spinlock.h>
+#include <linux/smp_lock.h>
+#include <linux/kthread.h>
+#include <linux/in.h>
+#include <linux/cdrom.h>
+#include <net/sock.h>
+#include <net/tcp.h>
+#include <scsi/scsi.h>
+#include <scsi/scsi_cmnd.h>
+
+#include <target/target_core_base.h>
+#include <target/target_core_device.h>
+#include <target/target_core_hba.h>
+#include <target/target_core_scdb.h>
+#include <target/target_core_pr.h>
+#include <target/target_core_alua.h>
+#include <target/target_core_tmr.h>
+#include <target/target_core_ua.h>
+#include <target/target_core_transport.h>
+#include <target/target_core_plugin.h>
+#include <target/target_core_seobj.h>
+#include <target/target_core_transport_plugin.h>
+#include <target/target_core_fabric_ops.h>
+#include <target/target_core_configfs.h>
+
+#undef TARGET_CORE_TRANSPORT_C
+
+/* #define DEBUG_CDB_HANDLER */
+#ifdef DEBUG_CDB_HANDLER
+#define DEBUG_CDB_H(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CDB_H(x...)
+#endif
+
+/* #define DEBUG_CMD_MAP */
+#ifdef DEBUG_CMD_MAP
+#define DEBUG_CMD_M(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CMD_M(x...)
+#endif
+
+/* #define DEBUG_MEM_ALLOC */
+#ifdef DEBUG_MEM_ALLOC
+#define DEBUG_MEM(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_MEM(x...)
+#endif
+
+/* #define DEBUG_MEM2_ALLOC */
+#ifdef DEBUG_MEM2_ALLOC
+#define DEBUG_MEM2(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_MEM2(x...)
+#endif
+
+/* #define DEBUG_SG_CALC */
+#ifdef DEBUG_SG_CALC
+#define DEBUG_SC(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_SC(x...)
+#endif
+
+/* #define DEBUG_SE_OBJ */
+#ifdef DEBUG_SE_OBJ
+#define DEBUG_SO(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_SO(x...)
+#endif
+
+/* #define DEBUG_CMD_VOL */
+#ifdef DEBUG_CMD_VOL
+#define DEBUG_VOL(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_VOL(x...)
+#endif
+
+/* #define DEBUG_CMD_STOP */
+#ifdef DEBUG_CMD_STOP
+#define DEBUG_CS(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_CS(x...)
+#endif
+
+/* #define DEBUG_PASSTHROUGH */
+#ifdef DEBUG_PASSTHROUGH
+#define DEBUG_PT(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_PT(x...)
+#endif
+
+/* #define DEBUG_TASK_STOP */
+#ifdef DEBUG_TASK_STOP
+#define DEBUG_TS(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TS(x...)
+#endif
+
+/* #define DEBUG_TRANSPORT_STOP */
+#ifdef DEBUG_TRANSPORT_STOP
+#define DEBUG_TRANSPORT_S(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TRANSPORT_S(x...)
+#endif
+
+/* #define DEBUG_TASK_FAILURE */
+#ifdef DEBUG_TASK_FAILURE
+#define DEBUG_TF(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TF(x...)
+#endif
+
+/* #define DEBUG_DEV_OFFLINE */
+#ifdef DEBUG_DEV_OFFLINE
+#define DEBUG_DO(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_DO(x...)
+#endif
+
+/* #define DEBUG_TASK_STATE */
+#ifdef DEBUG_TASK_STATE
+#define DEBUG_TSTATE(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TSTATE(x...)
+#endif
+
+/* #define DEBUG_STATUS_THR */
+#ifdef DEBUG_STATUS_THR
+#define DEBUG_ST(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_ST(x...)
+#endif
+
+/* #define DEBUG_TASK_TIMEOUT */
+#ifdef DEBUG_TASK_TIMEOUT
+#define DEBUG_TT(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_TT(x...)
+#endif
+
+/* #define DEBUG_GENERIC_REQUEST_FAILURE */
+#ifdef DEBUG_GENERIC_REQUEST_FAILURE
+#define DEBUG_GRF(x...) printk(KERN_INFO x)
+#else
+#define DEBUG_GRF(x...)
+#endif
+
+se_global_t *se_global;
+EXPORT_SYMBOL(se_global);
+
+struct kmem_cache *se_cmd_cache;
+struct kmem_cache *se_task_cache;
+struct kmem_cache *se_tmr_req_cache;
+struct kmem_cache *se_sess_cache;
+struct kmem_cache *se_ua_cache;
+struct kmem_cache *t10_pr_reg_cache;
+struct kmem_cache *t10_alua_lu_gp_cache;
+struct kmem_cache *t10_alua_lu_gp_mem_cache;
+struct kmem_cache *t10_alua_tg_pt_gp_cache;
+struct kmem_cache *t10_alua_tg_pt_gp_mem_cache;
+
+static int transport_generic_write_pending(se_cmd_t *);
+static int transport_processing_thread(void *);
+
+static char *transport_passthrough_get_fabric_name(void)
+{
+ return "PT";
+}
+
+static u32 transport_passthrough_get_task_tag(se_cmd_t *cmd)
+{
+ return 0;
+}
+
+static int transport_passthrough_get_cmd_state(se_cmd_t *cmd)
+{
+ return 0;
+}
+
+static void transport_passthrough_release_cmd_direct(se_cmd_t *cmd)
+{
+ return;
+}
+
+static u16 transport_passthrough_set_fabric_sense_len(
+ se_cmd_t *cmd,
+ u32 sense_len)
+{
+ return 0;
+}
+
+static u16 transport_passthrough_get_fabric_sense_len(void)
+{
+ return 0;
+}
+
+struct target_core_fabric_ops passthrough_fabric_ops = {
+ .release_cmd_direct = transport_passthrough_release_cmd_direct,
+ .get_fabric_name = transport_passthrough_get_fabric_name,
+ .get_task_tag = transport_passthrough_get_task_tag,
+ .get_cmd_state = transport_passthrough_get_cmd_state,
+ .set_fabric_sense_len = transport_passthrough_set_fabric_sense_len,
+ .get_fabric_sense_len = transport_passthrough_get_fabric_sense_len,
+};
+
+int init_se_global(void)
+{
+ se_global_t *global;
+ se_hba_t *hba;
+ int i;
+
+ global = kzalloc(sizeof(se_global_t), GFP_KERNEL);
+ if (!(global)) {
+ printk(KERN_ERR "Unable to allocate memory for se_global_t\n");
+ return -1;
+ }
+
+ INIT_LIST_HEAD(&global->g_lu_gps_list);
+ INIT_LIST_HEAD(&global->g_tg_pt_gps_list);
+ INIT_LIST_HEAD(&global->g_se_tpg_list);
+ spin_lock_init(&global->hba_lock);
+ spin_lock_init(&global->se_tpg_lock);
+ spin_lock_init(&global->lu_gps_lock);
+ spin_lock_init(&global->tg_pt_gps_lock);
+ spin_lock_init(&global->plugin_class_lock);
+
+ se_cmd_cache = kmem_cache_create("se_cmd_cache",
+ sizeof(se_cmd_t), __alignof__(se_cmd_t), 0, NULL);
+ if (!(se_cmd_cache)) {
+ printk(KERN_ERR "kmem_cache_create for se_cmd_t failed\n");
+ goto out;
+ }
+ se_task_cache = kmem_cache_create("se_task_cache",
+ sizeof(se_task_t), __alignof__(se_task_t), 0, NULL);
+ if (!(se_task_cache)) {
+ printk(KERN_ERR "kmem_cache_create for se_task_t failed\n");
+ goto out;
+ }
+ se_tmr_req_cache = kmem_cache_create("se_tmr_cache",
+ sizeof(se_tmr_req_t), __alignof__(se_tmr_req_t),
+ 0, NULL);
+ if (!(se_tmr_req_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for se_tmr_req_t"
+ " failed\n");
+ goto out;
+ }
+ se_sess_cache = kmem_cache_create("se_sess_cache",
+ sizeof(se_session_t), __alignof__(se_session_t),
+ 0, NULL);
+ if (!(se_sess_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for se_session_t"
+ " failed\n");
+ goto out;
+ }
+ se_ua_cache = kmem_cache_create("se_ua_cache",
+ sizeof(se_ua_t), __alignof__(se_ua_t), 0, NULL);
+ if (!(se_ua_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for se_ua_t failed\n");
+ goto out;
+ }
+ t10_pr_reg_cache = kmem_cache_create("t10_pr_reg_cache",
+ sizeof(t10_pr_registration_t),
+ __alignof__(t10_pr_registration_t), 0, NULL);
+ if (!(t10_pr_reg_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for t10_pr_registration_t"
+ " failed\n");
+ goto out;
+ }
+ t10_alua_lu_gp_cache = kmem_cache_create("t10_alua_lu_gp_cache",
+ sizeof(t10_alua_lu_gp_t), __alignof__(t10_alua_lu_gp_t),
+ 0, NULL);
+ if (!(t10_alua_lu_gp_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for t10_alua_lu_gp_cache"
+ " failed\n");
+ goto out;
+ }
+ t10_alua_lu_gp_mem_cache = kmem_cache_create("t10_alua_lu_gp_mem_cache",
+ sizeof(t10_alua_lu_gp_member_t),
+ __alignof__(t10_alua_lu_gp_member_t), 0, NULL);
+ if (!(t10_alua_lu_gp_mem_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for t10_alua_lu_gp_mem_"
+ "cache failed\n");
+ goto out;
+ }
+ t10_alua_tg_pt_gp_cache = kmem_cache_create("t10_alua_tg_pt_gp_cache",
+ sizeof(t10_alua_tg_pt_gp_t),
+ __alignof__(t10_alua_tg_pt_gp_t), 0, NULL);
+ if (!(t10_alua_tg_pt_gp_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for t10_alua_tg_pt_gp_"
+ "cache failed\n");
+ goto out;
+ }
+ t10_alua_tg_pt_gp_mem_cache = kmem_cache_create(
+ "t10_alua_tg_pt_gp_mem_cache",
+ sizeof(t10_alua_tg_pt_gp_member_t),
+ __alignof__(t10_alua_tg_pt_gp_member_t),
+ 0, NULL);
+ if (!(t10_alua_tg_pt_gp_mem_cache)) {
+ printk(KERN_ERR "kmem_cache_create() for t10_alua_tg_pt_gp_"
+ "mem_t failed\n");
+ goto out;
+ }
+
+ global->hba_list = kzalloc((sizeof(se_hba_t) *
+ TRANSPORT_MAX_GLOBAL_HBAS), GFP_KERNEL);
+ if (!(global->hba_list)) {
+ printk(KERN_ERR "Unable to allocate global->hba_list\n");
+ goto out;
+ }
+
+ for (i = 0; i < TRANSPORT_MAX_GLOBAL_HBAS; i++) {
+ hba = &global->hba_list[i];
+
+ hba->hba_status |= HBA_STATUS_FREE;
+ hba->hba_id = i;
+ INIT_LIST_HEAD(&hba->hba_dev_list);
+ spin_lock_init(&hba->device_lock);
+ spin_lock_init(&hba->hba_queue_lock);
+ init_MUTEX(&hba->hba_access_sem);
+#ifdef SNMP_SUPPORT
+ hba->hba_index = scsi_get_new_index(SCSI_INST_INDEX);
+#endif
+ }
+
+ global->plugin_class_list = kzalloc((sizeof(se_plugin_class_t) *
+ MAX_PLUGIN_CLASSES), GFP_KERNEL);
+ if (!(global->plugin_class_list)) {
+ printk(KERN_ERR "Unable to allocate global->"
+ "plugin_class_list\n");
+ goto out;
+ }
+
+ se_global = global;
+
+ return 0;
+out:
+ kfree(global->hba_list);
+ kfree(global->plugin_class_list);
+ if (se_cmd_cache)
+ kmem_cache_destroy(se_cmd_cache);
+ if (se_task_cache)
+ kmem_cache_destroy(se_task_cache);
+ if (se_tmr_req_cache)
+ kmem_cache_destroy(se_tmr_req_cache);
+ if (se_sess_cache)
+ kmem_cache_destroy(se_sess_cache);
+ if (se_ua_cache)
+ kmem_cache_destroy(se_ua_cache);
+ if (t10_pr_reg_cache)
+ kmem_cache_destroy(t10_pr_reg_cache);
+ if (t10_alua_lu_gp_cache)
+ kmem_cache_destroy(t10_alua_lu_gp_cache);
+ if (t10_alua_lu_gp_mem_cache)
+ kmem_cache_destroy(t10_alua_lu_gp_mem_cache);
+ if (t10_alua_tg_pt_gp_cache)
+ kmem_cache_destroy(t10_alua_tg_pt_gp_cache);
+ if (t10_alua_tg_pt_gp_mem_cache)
+ kmem_cache_destroy(t10_alua_tg_pt_gp_mem_cache);
+ kfree(global);
+ return -1;
+}
+
+void release_se_global(void)
+{
+ se_global_t *global;
+
+ global = se_global;
+ if (!(global))
+ return;
+
+ kfree(global->plugin_class_list);
+ kfree(global->hba_list);
+ kmem_cache_destroy(se_cmd_cache);
+ kmem_cache_destroy(se_task_cache);
+ kmem_cache_destroy(se_tmr_req_cache);
+ kmem_cache_destroy(se_sess_cache);
+ kmem_cache_destroy(se_ua_cache);
+ kmem_cache_destroy(t10_pr_reg_cache);
+ kmem_cache_destroy(t10_alua_lu_gp_cache);
+ kmem_cache_destroy(t10_alua_lu_gp_mem_cache);
+ kmem_cache_destroy(t10_alua_tg_pt_gp_cache);
+ kmem_cache_destroy(t10_alua_tg_pt_gp_mem_cache);
+ kfree(global);
+
+ se_global = NULL;
+}
+
+#ifdef DEBUG_DEV
+
+/* warning FIXME: PLUGIN API TODO */
+int __iscsi_debug_dev(se_device_t *dev)
+{
+ int fail_task = 0;
+ fd_dev_t *fd_dev;
+ iblock_dev_t *ib_dev;
+ rd_dev_t *rd_dev;
+ struct scsi_device *sd;
+
+ spin_lock(&se_global->debug_dev_lock);
+ switch (dev->se_hba->type) {
+ case PSCSI:
+ sd = (struct scsi_device *) dev->dev_ptr;
+ if (dev->dev_flags & DF_DEV_DEBUG) {
+ printk(KERN_INFO "HBA[%u] - Failing PSCSI Task for"
+ " %d/%d/%d\n", dev->se_hba->hba_id,
+ sd->channel, sd->id, sd->lun);
+ fail_task = 1;
+ }
+ break;
+ case IBLOCK:
+ ib_dev = (iblock_dev_t *) dev->dev_ptr;
+ if (dev->dev_flags & DF_DEV_DEBUG) {
+ printk(KERN_INFO "HBA[%u] - Failing IBLOCK Task for"
+ " %u/%u\n", dev->se_hba->hba_id,
+ ib_dev->ibd_major, ib_dev->ibd_minor);
+ fail_task = 1;
+ }
+ break;
+ case FILEIO:
+ fd_dev = (fd_dev_t *) dev->dev_ptr;
+ if (dev->dev_flags & DF_DEV_DEBUG) {
+ printk(KERN_INFO "HBA[%u] - Failing FILEIO Task for"
+ " %u\n", dev->se_hba->hba_id,
+ fd_dev->fd_dev_id);
+ fail_task = 1;
+ }
+ break;
+ case RAMDISK_DR:
+ case RAMDISK_MCP:
+ rd_dev = (rd_dev_t *) dev->dev_ptr;
+ if (dev->dev_flags & DF_DEV_DEBUG) {
+ printk(KERN_INFO "HBA[%u] - Failing RAMDISK Task for"
+ " %u\n", dev->se_hba->hba_id,
+ rd_dev->rd_dev_id);
+ fail_task = 1;
+ }
+ break;
+ default:
+ if (dev->dev_flags & DF_DEV_DEBUG) {
+ printk(KERN_INFO "HBA[%u] - Failing unknown Task\n",
+ dev->se_hba->hba_id);
+ fail_task = 1;
+ }
+ break;
+ }
+ spin_unlock(&se_global->debug_dev_lock);
+
+ return fail_task;
+}
+
+#endif /* DEBUG_DEV */
+
+/* #warning FIXME: transport_get_iqn_sn() for se_global_t */
+unsigned char *transport_get_iqn_sn(void)
+{
+ /*
+ * Assume that for production WWN information will come through
+ * ConfigFS at /sys/kernel/config/target/core/$HBA/$DEV/vpd_unit_serial
+ */
+ return "1234567890";
+}
+
+void transport_init_queue_obj(se_queue_obj_t *qobj)
+{
+ atomic_set(&qobj->queue_cnt, 0);
+ INIT_LIST_HEAD(&qobj->qobj_list);
+ init_waitqueue_head(&qobj->thread_wq);
+ init_MUTEX_LOCKED(&qobj->thread_create_sem);
+ init_MUTEX_LOCKED(&qobj->thread_done_sem);
+ spin_lock_init(&qobj->cmd_queue_lock);
+}
+
+void transport_load_plugins(void)
+{
+ int ret = 0;
+
+#ifdef PARALLEL_SCSI
+ plugin_register((void *)&pscsi_template, pscsi_template.type,
+ pscsi_template.name, PLUGIN_TYPE_TRANSPORT,
+ pscsi_template.get_plugin_info, &ret);
+#endif
+#ifdef PYX_IBLOCK
+ plugin_register((void *)&iblock_template, iblock_template.type,
+ iblock_template.name, PLUGIN_TYPE_TRANSPORT,
+ iblock_template.get_plugin_info, &ret);
+#endif
+#ifdef PYX_RAMDISK
+ plugin_register((void *)&rd_dr_template, rd_dr_template.type,
+ rd_dr_template.name, PLUGIN_TYPE_TRANSPORT,
+ rd_dr_template.get_plugin_info, &ret);
+ plugin_register((void *)&rd_mcp_template, rd_mcp_template.type,
+ rd_mcp_template.name, PLUGIN_TYPE_TRANSPORT,
+ rd_mcp_template.get_plugin_info, &ret);
+#endif
+#ifdef PYX_FILEIO
+ plugin_register((void *)&fileio_template, fileio_template.type,
+ fileio_template.name, PLUGIN_TYPE_TRANSPORT,
+ fileio_template.get_plugin_info, &ret);
+#endif
+}
+
+se_plugin_t *transport_core_get_plugin_by_name(const char *name)
+{
+ se_plugin_class_t *pc;
+ se_plugin_t *p;
+ int i;
+
+ pc = plugin_get_class(PLUGIN_TYPE_TRANSPORT);
+ if (!(pc))
+ return NULL;
+
+ for (i = 0; i < MAX_PLUGINS; i++) {
+ p = &pc->plugin_array[i];
+
+ if (!p->plugin_obj)
+ continue;
+
+ if (!(strncmp(name, p->plugin_name, strlen(p->plugin_name))))
+ return p;
+ }
+
+ return NULL;
+}
+
+void transport_check_dev_params_delim(char *ptr, char **cur)
+{
+ char *ptr2;
+
+ if (ptr) {
+ ptr2 = strstr(ptr, ",");
+ if ((ptr2)) {
+ *ptr2 = '\0';
+ *cur = (ptr2 + 1); /* Skip over comma */
+ } else
+ *cur = NULL;
+ }
+}
+
+se_session_t *transport_init_session(void)
+{
+ se_session_t *se_sess;
+
+ se_sess = kmem_cache_zalloc(se_sess_cache, GFP_KERNEL);
+ if (!(se_sess)) {
+ printk(KERN_ERR "Unable to allocate se_session_t from"
+ " se_sess_cache\n");
+ return ERR_PTR(-ENOMEM);
+ }
+ INIT_LIST_HEAD(&se_sess->sess_list);
+
+ return se_sess;
+}
+EXPORT_SYMBOL(transport_init_session);
+
+/*
+ * Called with spin_lock_bh(&se_portal_group_t->session_lock called.
+ */
+void __transport_register_session(
+ se_portal_group_t *se_tpg,
+ se_node_acl_t *se_nacl,
+ se_session_t *se_sess,
+ void *fabric_sess_ptr)
+{
+ se_sess->se_tpg = se_tpg;
+ se_sess->fabric_sess_ptr = fabric_sess_ptr;
+ /*
+ * Used by se_node_acl_t's under ConfigFS to locate active se_session-t
+ *
+ * Only set for se_session_t's that will actually be moving I/O.
+ * eg: *NOT* discovery sessions.
+ */
+ if (se_nacl) {
+ spin_lock_bh(&se_nacl->nacl_sess_lock);
+ se_nacl->nacl_sess = se_sess;
+ spin_unlock_bh(&se_nacl->nacl_sess_lock);
+ }
+ list_add_tail(&se_sess->sess_list, &se_tpg->tpg_sess_list);
+
+ printk(KERN_INFO "TARGET_CORE[%s]: Registered fabric_sess_ptr: %p\n",
+ TPG_TFO(se_tpg)->get_fabric_name(), se_sess->fabric_sess_ptr);
+}
+EXPORT_SYMBOL(__transport_register_session);
+
+void transport_register_session(
+ se_portal_group_t *se_tpg,
+ se_node_acl_t *se_nacl,
+ se_session_t *se_sess,
+ void *fabric_sess_ptr)
+{
+ spin_lock_bh(&se_tpg->session_lock);
+ __transport_register_session(se_tpg, se_nacl, se_sess, fabric_sess_ptr);
+ spin_unlock_bh(&se_tpg->session_lock);
+}
+EXPORT_SYMBOL(transport_register_session);
+
+void transport_deregister_session_configfs(se_session_t *se_sess)
+{
+ se_node_acl_t *se_nacl;
+
+ /*
+ * Used by se_node_acl_t's under ConfigFS to locate active se_session_t
+ */
+ se_nacl = se_sess->se_node_acl;
+ if ((se_nacl)) {
+ spin_lock_bh(&se_nacl->nacl_sess_lock);
+ se_nacl->nacl_sess = NULL;
+ spin_unlock_bh(&se_nacl->nacl_sess_lock);
+ }
+}
+EXPORT_SYMBOL(transport_deregister_session_configfs);
+
+void transport_free_session(se_session_t *se_sess)
+{
+ kmem_cache_free(se_sess_cache, se_sess);
+}
+EXPORT_SYMBOL(transport_free_session);
+
+void transport_deregister_session(se_session_t *se_sess)
+{
+ se_portal_group_t *se_tpg = se_sess->se_tpg;
+ se_node_acl_t *se_nacl;
+
+ if (!(se_tpg)) {
+ transport_free_session(se_sess);
+ return;
+ }
+
+ spin_lock_bh(&se_tpg->session_lock);
+ list_del(&se_sess->sess_list);
+ se_sess->se_tpg = NULL;
+ se_sess->fabric_sess_ptr = NULL;
+ spin_unlock_bh(&se_tpg->session_lock);
+
+ /*
+ * Determine if we need to do extra work for this initiator node's
+ * se_node_acl_t if it had been previously dynamically generated.
+ */
+ se_nacl = se_sess->se_node_acl;
+ if ((se_nacl)) {
+ spin_lock_bh(&se_tpg->acl_node_lock);
+ if (se_nacl->nodeacl_flags & NAF_DYNAMIC_NODE_ACL) {
+ if (!(TPG_TFO(se_tpg)->tpg_check_demo_mode_cache(
+ se_tpg))) {
+ list_del(&se_nacl->acl_list);
+ se_tpg->num_node_acls--;
+ spin_unlock_bh(&se_tpg->acl_node_lock);
+ core_free_device_list_for_node(se_nacl, se_tpg);
+ TPG_TFO(se_tpg)->tpg_release_fabric_acl(se_tpg,
+ se_nacl);
+ kfree(se_nacl);
+ spin_lock_bh(&se_tpg->acl_node_lock);
+ }
+ }
+ spin_unlock_bh(&se_tpg->acl_node_lock);
+ }
+
+ transport_free_session(se_sess);
+
+ printk(KERN_INFO "TARGET_CORE[%s]: Deregistered fabric_sess\n",
+ TPG_TFO(se_tpg)->get_fabric_name());
+}
+EXPORT_SYMBOL(transport_deregister_session);
+
+/*
+ * Called with T_TASK(cmd)->t_state_lock held.
+ */
+static void transport_all_task_dev_remove_state(se_cmd_t *cmd)
+{
+ se_device_t *dev;
+ se_task_t *task;
+ unsigned long flags;
+
+ if (!T_TASK(cmd))
+ return;
+
+ list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+ dev = task->se_dev;
+ if (!(dev))
+ continue;
+
+ if (atomic_read(&task->task_active))
+ continue;
+
+ if (!(atomic_read(&task->task_state_active)))
+ continue;
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ list_del(&task->t_state_list);
+ DEBUG_TSTATE("Removed ITT: 0x%08x dev: %p task[%p]\n",
+ CMD_TFO(cmd)->tfo_get_task_tag(cmd), dev, task);
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+ atomic_set(&task->task_state_active, 0);
+ atomic_dec(&T_TASK(cmd)->t_task_cdbs_ex_left);
+ }
+}
+
+/*
+ * Called with T_TASK(cmd)->t_state_lock held.
+ */
+void transport_task_dev_remove_state(se_task_t *task, se_device_t *dev)
+{
+ se_cmd_t *cmd = task->task_se_cmd;
+ unsigned long flags;
+
+ /*
+ * We cannot remove the task from the state list while said task is
+ * still active and probably timed out.
+ */
+ if (atomic_read(&task->task_active)) {
+#if 0
+ printk(KERN_ERR "Skipping Removal of state for ITT: 0x%08x"
+ " dev: %p task[%p]\n"
+ CMD_TFO(task->task_se_cmd)->tfo_get_task_tag(
+ task->task_se_cmd), dev, task);
+#endif
+ return;
+ }
+
+ if (atomic_read(&task->task_state_active)) {
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ list_del(&task->t_state_list);
+ DEBUG_TSTATE("Removed ITT: 0x%08x dev: %p task[%p]\n",
+ CMD_TFO(task->task_se_cmd)->tfo_get_task_tag(
+ task->task_se_cmd), dev, task);
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+ atomic_set(&task->task_state_active, 0);
+ atomic_dec(&T_TASK(cmd)->t_task_cdbs_ex_left);
+ }
+}
+
+static void transport_passthrough_check_stop(se_cmd_t *cmd)
+{
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+ return;
+
+ if (!cmd->transport_passthrough_done) {
+ if (cmd->callback) {
+ cmd->callback(cmd, cmd->callback_arg,
+ transport_passthrough_complete(cmd));
+ } else
+ up(&T_TASK(cmd)->t_transport_passthrough_sem);
+
+ return;
+ }
+
+ cmd->transport_passthrough_done(cmd);
+}
+
+/* transport_cmd_check_stop():
+ *
+ * 'transport_off = 1' determines if t_transport_active should be cleared.
+ * 'transport_off = 2' determines if task_dev_state should be removed.
+ *
+ * A non-zero u8 t_state sets cmd->t_state.
+ * Returns 1 when command is stopped, else 0.
+ */
+static int transport_cmd_check_stop(
+ se_cmd_t *cmd,
+ int transport_off,
+ u8 t_state)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ /*
+ * Determine if IOCTL context caller in requesting the stopping of this
+ * command for LUN shutdown purposes.
+ */
+ if (atomic_read(&T_TASK(cmd)->transport_lun_stop)) {
+ DEBUG_CS("%s:%d atomic_read(&T_TASK(cmd)->transport_lun_stop)"
+ " == TRUE for ITT: 0x%08x\n", __func__, __LINE__,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+
+ cmd->deferred_t_state = cmd->t_state;
+ cmd->t_state = TRANSPORT_DEFERRED_CMD;
+ atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+ if (transport_off == 2)
+ transport_all_task_dev_remove_state(cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ up(&T_TASK(cmd)->transport_lun_stop_sem);
+ return 1;
+ }
+ /*
+ * Determine if frontend context caller is requesting the stopping of
+ * this command for frontend excpections.
+ */
+ if (atomic_read(&T_TASK(cmd)->t_transport_stop)) {
+ DEBUG_CS("%s:%d atomic_read(&T_TASK(cmd)->t_transport_stop) =="
+ " TRUE for ITT: 0x%08x\n", __func__, __LINE__,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+
+ cmd->deferred_t_state = cmd->t_state;
+ cmd->t_state = TRANSPORT_DEFERRED_CMD;
+ if (transport_off == 2)
+ transport_all_task_dev_remove_state(cmd);
+
+ /*
+ * Clear se_cmd_t->se_lun before the transport_off == 2 handoff
+ * to FE.
+ */
+ if ((transport_off == 2) && !(cmd->se_cmd_flags &
+ SCF_CMD_PASSTHROUGH))
+ cmd->se_lun = NULL;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ up(&T_TASK(cmd)->t_transport_stop_sem);
+ return 1;
+ }
+ if (transport_off) {
+ atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+ if (transport_off == 2)
+ transport_all_task_dev_remove_state(cmd);
+
+ /*
+ * Clear se_cmd_t->se_lun before the transport_off == 2 handoff
+ * to FE.
+ */
+ if ((transport_off == 2) && !(cmd->se_cmd_flags &
+ SCF_CMD_PASSTHROUGH))
+ cmd->se_lun = NULL;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ return 0;
+ } else if (t_state)
+ cmd->t_state = t_state;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ return 0;
+}
+
+static void transport_lun_remove_cmd(se_cmd_t *cmd)
+{
+ se_lun_t *lun = SE_LUN(cmd);
+ unsigned long flags;
+
+ if (!lun)
+ return;
+ /*
+ * Do not track passthrough se_cmd_t for now..
+ */
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+ return;
+
+ spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+ if (atomic_read(&T_TASK(cmd)->transport_lun_active)) {
+ list_del(&cmd->se_lun_list);
+ atomic_set(&T_TASK(cmd)->transport_lun_active, 0);
+#if 0
+ printk(KERN_INFO "Removed ITT: 0x%08x from LUN LIST[%d]\n"
+ CMD_TFO(cmd)->get_task_tag(cmd), lun->unpacked_lun);
+#endif
+ }
+ spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
+}
+
+void transport_cmd_finish_abort(se_cmd_t *cmd, int remove)
+{
+ transport_remove_cmd_from_queue(cmd,
+ CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+ cmd->se_orig_obj_ptr));
+
+ transport_lun_remove_cmd(cmd);
+
+ if (!(transport_cmd_check_stop(cmd, 1, 0))) {
+ transport_passthrough_check_stop(cmd);
+ return;
+ }
+ if (remove)
+ transport_generic_remove(cmd, 0, 0);
+}
+
+void transport_cmd_finish_abort_tmr(se_cmd_t *cmd)
+{
+ transport_remove_cmd_from_queue(cmd,
+ CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+ cmd->se_orig_obj_ptr));
+
+ if (!(transport_cmd_check_stop(cmd, 1, 0))) {
+ transport_passthrough_check_stop(cmd);
+ return;
+ }
+ transport_generic_remove(cmd, 0, 0);
+}
+
+int transport_add_cmd_to_queue(
+ se_cmd_t *cmd,
+ se_queue_obj_t *qobj,
+ u8 t_state)
+{
+ se_queue_req_t *qr;
+ unsigned long flags;
+
+ qr = kzalloc(sizeof(se_queue_req_t), GFP_ATOMIC);
+ if (!(qr)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " se_queue_req_t\n");
+ return -1;
+ }
+ INIT_LIST_HEAD(&qr->qr_list);
+
+ qr->cmd = (void *)cmd;
+ qr->state = t_state;
+
+ if (t_state) {
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ cmd->t_state = t_state;
+ atomic_set(&T_TASK(cmd)->t_transport_active, 1);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ }
+
+ spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
+ list_add_tail(&qr->qr_list, &qobj->qobj_list);
+ atomic_inc(&T_TASK(cmd)->t_transport_queue_active);
+ spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+ atomic_inc(&qobj->queue_cnt);
+ wake_up_interruptible(&qobj->thread_wq);
+ return 0;
+}
+
+static int transport_add_cmd_to_dev_queue(se_cmd_t *cmd, u8 t_state)
+{
+ se_device_t *dev = cmd->se_dev;
+
+ return transport_add_cmd_to_queue(cmd, dev->dev_queue_obj, t_state);
+}
+
+/*
+ * Called with se_queue_obj_t->cmd_queue_lock held.
+ */
+se_queue_req_t *__transport_get_qr_from_queue(se_queue_obj_t *qobj)
+{
+ se_cmd_t *cmd;
+ se_queue_req_t *qr = NULL;
+
+ if (list_empty(&qobj->qobj_list))
+ return NULL;
+
+ list_for_each_entry(qr, &qobj->qobj_list, qr_list)
+ break;
+
+ if (qr->cmd) {
+ cmd = (se_cmd_t *)qr->cmd;
+ atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
+ }
+ list_del(&qr->qr_list);
+ atomic_dec(&qobj->queue_cnt);
+
+ return qr;
+}
+
+se_queue_req_t *transport_get_qr_from_queue(se_queue_obj_t *qobj)
+{
+ se_cmd_t *cmd;
+ se_queue_req_t *qr;
+ unsigned long flags;
+
+ spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
+ if (list_empty(&qobj->qobj_list)) {
+ spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+ return NULL;
+ }
+
+ list_for_each_entry(qr, &qobj->qobj_list, qr_list)
+ break;
+
+ if (qr->cmd) {
+ cmd = (se_cmd_t *)qr->cmd;
+ atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
+ }
+ list_del(&qr->qr_list);
+ atomic_dec(&qobj->queue_cnt);
+ spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+ return qr;
+}
+
+void transport_remove_cmd_from_queue(se_cmd_t *cmd, se_queue_obj_t *qobj)
+{
+ se_cmd_t *q_cmd;
+ se_queue_req_t *qr = NULL, *qr_p = NULL;
+ unsigned long flags;
+
+ spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
+ if (!(atomic_read(&T_TASK(cmd)->t_transport_queue_active))) {
+ spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+ return;
+ }
+
+ list_for_each_entry_safe(qr, qr_p, &qobj->qobj_list, qr_list) {
+ q_cmd = (se_cmd_t *)qr->cmd;
+ if (q_cmd != cmd)
+ continue;
+
+ atomic_dec(&T_TASK(q_cmd)->t_transport_queue_active);
+ atomic_dec(&qobj->queue_cnt);
+ list_del(&qr->qr_list);
+ kfree(qr);
+ }
+ spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+ if (atomic_read(&T_TASK(cmd)->t_transport_queue_active)) {
+ printk(KERN_ERR "ITT: 0x%08x t_transport_queue_active: %d\n",
+ CMD_TFO(cmd)->get_task_tag(cmd),
+ atomic_read(&T_TASK(cmd)->t_transport_queue_active));
+ }
+}
+
+void transport_complete_cmd(se_cmd_t *cmd, int success)
+{
+ int t_state;
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (!success) {
+ cmd->transport_error_status =
+ PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ t_state = TRANSPORT_COMPLETE_FAILURE;
+ } else {
+ t_state = TRANSPORT_COMPLETE_OK;
+ }
+ atomic_set(&T_TASK(cmd)->t_transport_complete, 1);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ cmd->transport_add_cmd_to_queue(cmd, t_state);
+}
+
+/* transport_complete_task():
+ *
+ * Called from interrupt and non interrupt context depending
+ * on the transport plugin.
+ */
+void transport_complete_task(se_task_t *task, int success)
+{
+ se_cmd_t *cmd = TASK_CMD(task);
+ se_device_t *dev = task->se_dev;
+ int t_state;
+ unsigned long flags;
+#if 0
+ printk(KERN_INFO "task: %p CDB: 0x%02x obj_ptr: %p\n", task,
+ T_TASK(cmd)->t_task_cdb[0], dev);
+#endif
+ if (dev) {
+ spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
+ atomic_inc(&dev->depth_left);
+ atomic_inc(&SE_HBA(dev)->left_queue_depth);
+ spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+ }
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ atomic_set(&task->task_active, 0);
+
+ /*
+ * See if any sense data exists, if so set the TASK_SENSE flag.
+ * Also check for any other post completion work that needs to be
+ * done by the plugins.
+ */
+ if (!dev)
+ goto check_task_stop;
+
+ if (TRANSPORT(dev)->transport_complete(task) != 0) {
+ cmd->se_cmd_flags |= SCF_TRANSPORT_TASK_SENSE;
+ task->task_sense = 1;
+ success = 1;
+ }
+
+ /*
+ * See if we are waiting for outstanding se_task_t
+ * to complete for an exception condition
+ */
+check_task_stop:
+ if (atomic_read(&task->task_stop)) {
+ /*
+ * Decrement T_TASK(cmd)->t_se_count if this task had
+ * previously thrown its timeout exception handler.
+ */
+ if (atomic_read(&task->task_timeout)) {
+ atomic_dec(&T_TASK(cmd)->t_se_count);
+ atomic_set(&task->task_timeout, 0);
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ up(&task->task_stop_sem);
+ return;
+ }
+ /*
+ * If the task's timeout handler has fired, use the t_task_cdbs_timeout
+ * left counter to determine when the se_cmd_t is ready to be queued to
+ * the processing thread.
+ */
+ if (atomic_read(&task->task_timeout)) {
+ if (!(atomic_dec_and_test(
+ &T_TASK(cmd)->t_task_cdbs_timeout_left))) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+ return;
+ }
+ t_state = TRANSPORT_COMPLETE_TIMEOUT;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ cmd->transport_add_cmd_to_queue(cmd, t_state);
+ return;
+ }
+ atomic_dec(&T_TASK(cmd)->t_task_cdbs_timeout_left);
+
+#ifdef DEBUG_DEV
+ if (dev) {
+ if (__iscsi_debug_dev(dev) != 0) {
+ success = 0;
+ task->task_scsi_status = 1;
+ cmd->transport_error_status =
+ PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ }
+ }
+#endif /* DEBUG_DEV */
+
+ /*
+ * Decrement the outstanding t_task_cdbs_left count. The last
+ * se_task_t from se_cmd_t will complete itself into the
+ * device queue depending upon int success.
+ */
+ if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_left))) {
+ if (!success)
+ T_TASK(cmd)->t_tasks_failed = 1;
+
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+
+ if (!success || T_TASK(cmd)->t_tasks_failed) {
+ t_state = TRANSPORT_COMPLETE_FAILURE;
+ if (!task->task_error_status) {
+ task->task_error_status =
+ PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+ cmd->transport_error_status =
+ PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+ }
+ } else {
+ atomic_set(&T_TASK(cmd)->t_transport_complete, 1);
+ t_state = TRANSPORT_COMPLETE_OK;
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ cmd->transport_add_cmd_to_queue(cmd, t_state);
+}
+
+/* __transport_add_task_to_execute_queue():
+ *
+ * Called with se_dev_t->execute_task_lock called.
+ */
+static void __transport_add_task_to_execute_queue(
+ se_task_t *task,
+ se_device_t *dev)
+{
+ list_add_tail(&task->t_execute_list, &dev->execute_task_list);
+ atomic_inc(&dev->execute_tasks);
+
+ if (atomic_read(&task->task_state_active))
+ return;
+
+ list_add_tail(&task->t_state_list, &dev->state_task_list);
+ atomic_set(&task->task_state_active, 1);
+
+ DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
+ CMD_TFO(task->task_se_cmd)->get_task_tag(task->task_se_cmd),
+ task, dev);
+}
+
+/* transport_add_task_to_execute_queue():
+ *
+ *
+ */
+void transport_add_task_to_execute_queue(se_task_t *task, se_device_t *dev)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ list_add_tail(&task->t_execute_list, &dev->execute_task_list);
+ atomic_inc(&dev->execute_tasks);
+
+ if (atomic_read(&task->task_state_active)) {
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+ return;
+ }
+
+ list_add_tail(&task->t_state_list, &dev->state_task_list);
+ atomic_set(&task->task_state_active, 1);
+
+ DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
+ CMD_TFO(task->task_se_cmd)->get_task_tag(task->task_se_cmd),
+ task, dev);
+
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+}
+
+static void transport_add_tasks_to_state_queue(se_cmd_t *cmd)
+{
+ se_device_t *dev;
+ se_task_t *task;
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+ dev = task->se_dev;
+
+ if (atomic_read(&task->task_state_active))
+ continue;
+
+ spin_lock(&dev->execute_task_lock);
+ list_add_tail(&task->t_state_list, &dev->state_task_list);
+ atomic_set(&task->task_state_active, 1);
+
+ DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
+ CMD_TFO(task->task_se_cmd)->get_task_tag(
+ task->task_se_cmd), task, dev);
+
+ spin_unlock(&dev->execute_task_lock);
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+/* transport_add_tasks_from_cmd():
+ *
+ *
+ */
+void transport_add_tasks_from_cmd(se_cmd_t *cmd)
+{
+ se_device_t *dev = SE_DEV(cmd);
+ se_task_t *task;
+ unsigned long flags;
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+ if (atomic_read(&task->task_execute_queue))
+ continue;
+
+ __transport_add_task_to_execute_queue(task, dev);
+ atomic_set(&task->task_execute_queue, 1);
+ }
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+ return;
+}
+
+/* transport_get_task_from_execute_queue():
+ *
+ * Called with dev->execute_task_lock held.
+ */
+se_task_t *transport_get_task_from_execute_queue(se_device_t *dev)
+{
+ se_task_t *task;
+
+ if (list_empty(&dev->execute_task_list))
+ return NULL;
+
+ list_for_each_entry(task, &dev->execute_task_list, t_execute_list)
+ break;
+
+ list_del(&task->t_execute_list);
+ atomic_dec(&dev->execute_tasks);
+
+ return task;
+}
+
+/* transport_remove_task_from_execute_queue():
+ *
+ *
+ */
+static void transport_remove_task_from_execute_queue(
+ se_task_t *task,
+ se_device_t *dev)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ list_del(&task->t_execute_list);
+ atomic_dec(&dev->execute_tasks);
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+}
+
+/* transport_check_device_tcq():
+ *
+ *
+ */
+int transport_check_device_tcq(
+ se_device_t *dev,
+ u32 unpacked_lun,
+ u32 device_tcq)
+{
+ if (device_tcq > dev->queue_depth) {
+ printk(KERN_ERR "Attempting to set storage device queue depth"
+ " to %d while transport maximum is %d on LUN: %u,"
+ " ignoring request\n", device_tcq, dev->queue_depth,
+ unpacked_lun);
+ return -1;
+ } else if (!device_tcq) {
+ printk(KERN_ERR "Attempting to set storage device queue depth"
+ " to 0 on LUN: %u, ignoring request\n", unpacked_lun);
+ return -1;
+ }
+
+ dev->queue_depth = device_tcq;
+ atomic_set(&dev->depth_left, dev->queue_depth);
+ printk(KERN_INFO "Reset Device Queue Depth to %u for Logical Unit"
+ " Number: %u\n", dev->queue_depth, unpacked_lun);
+
+ return 0;
+}
+EXPORT_SYMBOL(transport_check_device_tcq);
+
+unsigned char *transport_dump_cmd_direction (se_cmd_t *cmd)
+{
+ switch (cmd->data_direction) {
+ case SE_DIRECTION_NONE:
+ return "NONE";
+ case SE_DIRECTION_READ:
+ return "READ";
+ case SE_DIRECTION_WRITE:
+ return "WRITE";
+ case SE_DIRECTION_BIDI:
+ return "BIDI";
+ default:
+ break;
+ }
+
+ return "UNKNOWN";
+}
+
+void transport_dump_dev_state(
+ se_device_t *dev,
+ char *b,
+ int *bl)
+{
+ *bl += sprintf(b + *bl, "Status: ");
+ switch (dev->dev_status) {
+ case TRANSPORT_DEVICE_ACTIVATED:
+ *bl += sprintf(b + *bl, "ACTIVATED");
+ break;
+ case TRANSPORT_DEVICE_DEACTIVATED:
+ *bl += sprintf(b + *bl, "DEACTIVATED");
+ break;
+ case TRANSPORT_DEVICE_SHUTDOWN:
+ *bl += sprintf(b + *bl, "SHUTDOWN");
+ break;
+ case TRANSPORT_DEVICE_OFFLINE_ACTIVATED:
+ case TRANSPORT_DEVICE_OFFLINE_DEACTIVATED:
+ *bl += sprintf(b + *bl, "OFFLINE");
+ break;
+ default:
+ *bl += sprintf(b + *bl, "UNKNOWN=%d", dev->dev_status);
+ break;
+ }
+
+ *bl += sprintf(b + *bl, " Execute/Left/Max Queue Depth: %d/%d/%d",
+ atomic_read(&dev->execute_tasks), atomic_read(&dev->depth_left),
+ dev->queue_depth);
+ *bl += sprintf(b + *bl, " SectorSize: %u MaxSectors: %u\n",
+ DEV_ATTRIB(dev)->block_size, DEV_ATTRIB(dev)->max_sectors);
+ *bl += sprintf(b + *bl, " ");
+}
+
+void transport_dump_dev_info(
+ se_device_t *dev,
+ se_lun_t *lun,
+ unsigned long long total_bytes,
+ char *b, /* Pointer to info buffer */
+ int *bl)
+{
+ se_subsystem_api_t *t;
+ int ret = 0;
+
+ t = (se_subsystem_api_t *)plugin_get_obj(PLUGIN_TYPE_TRANSPORT,
+ dev->type, &ret);
+ if (!t || (ret != 0))
+ return;
+
+ t->get_dev_info(dev, b, bl);
+ *bl += sprintf(b + *bl, " ");
+ *bl += sprintf(b + *bl, "Type: %s ",
+ scsi_device_type(TRANSPORT(dev)->get_device_type(dev)));
+ *bl += sprintf(b + *bl, "ANSI SCSI revision: %02x ",
+ TRANSPORT(dev)->get_device_rev(dev));
+
+ if (DEV_OBJ_API(dev)->get_t10_wwn) {
+ t10_wwn_t *wwn = DEV_OBJ_API(dev)->get_t10_wwn((void *)dev);
+
+ *bl += sprintf(b + *bl, "Unit Serial: %s ",
+ ((strlen(wwn->unit_serial) != 0) ?
+ (char *)wwn->unit_serial : "None"));
+ }
+ *bl += sprintf(b + *bl, "%s", "DIRECT");
+
+ if ((DEV_OBJ_API(dev)->check_count(&dev->dev_access_obj)) ||
+ (DEV_OBJ_API(dev)->check_count(&dev->dev_feature_obj)))
+ *bl += sprintf(b + *bl, " ACCESSED\n");
+ else if (DEV_OBJ_API(dev)->check_count(&dev->dev_export_obj))
+ *bl += sprintf(b + *bl, " EXPORTED\n");
+ else
+ *bl += sprintf(b + *bl, " FREE\n");
+
+ if (lun) {
+ *bl += sprintf(b + *bl, " Core Host ID: %u LUN: %u",
+ dev->se_hba->hba_id, lun->unpacked_lun);
+ if (!(TRANSPORT(dev)->get_device_type(dev))) {
+ *bl += sprintf(b + *bl, " Active Cmds: %d Total Bytes"
+ ": %llu\n", atomic_read(&dev->active_cmds),
+ total_bytes);
+ } else {
+ *bl += sprintf(b + *bl, " Active Cmds: %d\n",
+ atomic_read(&dev->active_cmds));
+ }
+ } else {
+ if (!(TRANSPORT(dev)->get_device_type(dev))) {
+ *bl += sprintf(b + *bl, " Core Host ID: %u"
+ " Active Cmds: %d Total Bytes: %llu\n",
+ dev->se_hba->hba_id,
+ atomic_read(&dev->active_cmds), total_bytes);
+ } else {
+ *bl += sprintf(b + *bl, " CoreI Host ID: %u"
+ " Active Cmds: %d\n", dev->se_hba->hba_id,
+ atomic_read(&dev->active_cmds));
+ }
+ }
+}
+
+/* transport_release_all_cmds():
+ *
+ *
+ */
+static void transport_release_all_cmds(se_device_t *dev)
+{
+ se_cmd_t *cmd = NULL;
+ se_queue_req_t *qr = NULL, *qr_p = NULL;
+ int bug_out = 0, t_state;
+ unsigned long flags;
+
+ spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+ list_for_each_entry_safe(qr, qr_p, &dev->dev_queue_obj->qobj_list,
+ qr_list) {
+
+ cmd = (se_cmd_t *)qr->cmd;
+ t_state = qr->state;
+ list_del(&qr->qr_list);
+ kfree(qr);
+ spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock,
+ flags);
+
+ printk(KERN_ERR "Releasing %s ITT: 0x%08x, i_state: %u,"
+ " t_state: %u directly\n",
+ (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH) ?
+ "Passthrough" : "Normal",
+ CMD_TFO(cmd)->get_task_tag(cmd),
+ CMD_TFO(cmd)->get_cmd_state(cmd), t_state);
+
+ transport_release_fe_cmd(cmd);
+ bug_out = 1;
+
+ spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+ }
+ spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock, flags);
+#if 0
+ if (bug_out)
+ BUG();
+#endif
+}
+
+/* transport_dev_write_pending_nop():
+ *
+ *
+ */
+static int transport_dev_write_pending_nop(se_task_t *task)
+{
+ return 0;
+}
+
+static int transport_get_inquiry(
+ se_obj_lun_type_t *obj_api,
+ t10_wwn_t *wwn,
+ void *obj_ptr)
+{
+ se_cmd_t *cmd;
+ unsigned char *buf;
+ int i;
+ unsigned char cdb[SCSI_CDB_SIZE];
+
+ memset(cdb, 0, SCSI_CDB_SIZE);
+ cdb[0] = INQUIRY;
+ cdb[3] = (INQUIRY_LEN >> 8) & 0xff;
+ cdb[4] = (INQUIRY_LEN & 0xff);
+
+ cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+ 0, NULL, 0, INQUIRY_LEN, obj_api, obj_ptr);
+ if (!(cmd))
+ return -1;
+
+ if (transport_generic_passthrough(cmd) < 0) {
+ transport_passthrough_release(cmd);
+ return -1;
+ }
+
+ buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+ /*
+ * Save the basic Vendor, Model and Revision in passed t10_wwn_t.
+ * We will obtain the VPD in a seperate passthrough operation.
+ */
+ memcpy((void *)&wwn->vendor[0], (void *)&buf[8],
+ sizeof(wwn->vendor));
+ memcpy((void *)&wwn->model[0], (void *)&buf[16],
+ sizeof(wwn->model));
+ memcpy((void *)&wwn->revision[0], (void *)&buf[32],
+ sizeof(wwn->revision));
+
+ printk(" Vendor: ");
+ for (i = 8; i < 16; i++)
+ if (buf[i] >= 0x20 && i < buf[4] + 5)
+ printk("%c", buf[i]);
+ else
+ printk(" ");
+
+ printk(" Model: ");
+ for (i = 16; i < 32; i++)
+ if (buf[i] >= 0x20 && i < buf[4] + 5)
+ printk("%c", buf[i]);
+ else
+ printk(" ");
+
+ printk(" Revision: ");
+ for (i = 32; i < 36; i++)
+ if (buf[i] >= 0x20 && i < buf[4] + 5)
+ printk("%c", buf[i]);
+ else
+ printk(" ");
+
+ printk("\n");
+
+ i = buf[0] & 0x1f;
+
+ printk(" Type: %s ", scsi_device_type(i));
+ printk(" ANSI SCSI revision: %02x",
+ buf[2] & 0x07);
+ if ((buf[2] & 0x07) == 1 && (buf[3] & 0x0f) == 1)
+ printk(" CCS\n");
+ else
+ printk("\n");
+
+ transport_passthrough_release(cmd);
+ return 0;
+}
+
+static int transport_get_inquiry_vpd_serial(
+ se_obj_lun_type_t *obj_api,
+ t10_wwn_t *wwn,
+ void *obj_ptr)
+{
+ unsigned char *buf;
+ se_cmd_t *cmd;
+ unsigned char cdb[SCSI_CDB_SIZE];
+
+ memset(cdb, 0, SCSI_CDB_SIZE);
+ cdb[0] = INQUIRY;
+ cdb[1] = 0x01; /* Query VPD */
+ cdb[2] = 0x80; /* Unit Serial Number */
+ cdb[3] = (INQUIRY_VPD_SERIAL_LEN >> 8) & 0xff;
+ cdb[4] = (INQUIRY_VPD_SERIAL_LEN & 0xff);
+
+ cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+ 0, NULL, 0, INQUIRY_VPD_SERIAL_LEN, obj_api, obj_ptr);
+ if (!(cmd))
+ return -1;
+
+ if (transport_generic_passthrough(cmd) < 0) {
+ transport_passthrough_release(cmd);
+ return -1;
+ }
+
+ buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+
+ printk(KERN_INFO "T10 VPD Unit Serial Number: %s\n", &buf[4]);
+ snprintf(&wwn->unit_serial[0], INQUIRY_VPD_SERIAL_LEN, "%s", &buf[4]);
+
+ transport_passthrough_release(cmd);
+ return 0;
+}
+
+static const char hex_str[] = "0123456789abcdef";
+
+void transport_dump_vpd_proto_id(
+ t10_vpd_t *vpd,
+ unsigned char *p_buf,
+ int p_buf_len)
+{
+ unsigned char buf[VPD_TMP_BUF_SIZE];
+ int len;
+
+ memset(buf, 0, VPD_TMP_BUF_SIZE);
+ len = sprintf(buf, "T10 VPD Protocol Identifier: ");
+
+ switch (vpd->protocol_identifier) {
+ case 0x00:
+ sprintf(buf+len, "Fibre Channel\n");
+ break;
+ case 0x10:
+ sprintf(buf+len, "Parallel SCSI\n");
+ break;
+ case 0x20:
+ sprintf(buf+len, "SSA\n");
+ break;
+ case 0x30:
+ sprintf(buf+len, "IEEE 1394\n");
+ break;
+ case 0x40:
+ sprintf(buf+len, "SCSI Remote Direct Memory Access"
+ " Protocol\n");
+ break;
+ case 0x50:
+ sprintf(buf+len, "Internet SCSI (iSCSI)\n");
+ break;
+ case 0x60:
+ sprintf(buf+len, "SAS Serial SCSI Protocol\n");
+ break;
+ case 0x70:
+ sprintf(buf+len, "Automation/Drive Interface Transport"
+ " Protocol\n");
+ break;
+ case 0x80:
+ sprintf(buf+len, "AT Attachment Interface ATA/ATAPI\n");
+ break;
+ default:
+ sprintf(buf+len, "Unknown 0x%02x\n",
+ vpd->protocol_identifier);
+ break;
+ }
+
+ if (p_buf)
+ strncpy(p_buf, buf, p_buf_len);
+ else
+ printk(KERN_INFO "%s", buf);
+}
+
+void transport_set_vpd_proto_id(t10_vpd_t *vpd, unsigned char *page_83)
+{
+ /*
+ * Check if the Protocol Identifier Valid (PIV) bit is set..
+ *
+ * from spc3r23.pdf section 7.5.1
+ */
+ if (page_83[1] & 0x80) {
+ vpd->protocol_identifier = (page_83[0] & 0xf0);
+ vpd->protocol_identifier_set = 1;
+ transport_dump_vpd_proto_id(vpd, NULL, 0);
+ }
+}
+
+int transport_dump_vpd_assoc(
+ t10_vpd_t *vpd,
+ unsigned char *p_buf,
+ int p_buf_len)
+{
+ unsigned char buf[VPD_TMP_BUF_SIZE];
+ int ret = 0, len;
+
+ memset(buf, 0, VPD_TMP_BUF_SIZE);
+ len = sprintf(buf, "T10 VPD Identifier Association: ");
+
+ switch (vpd->association) {
+ case 0x00:
+ sprintf(buf+len, "addressed logical unit\n");
+ break;
+ case 0x10:
+ sprintf(buf+len, "target port\n");
+ break;
+ case 0x20:
+ sprintf(buf+len, "SCSI target device\n");
+ break;
+ default:
+ sprintf(buf+len, "Unknown 0x%02x\n", vpd->association);
+ ret = -1;
+ break;
+ }
+
+ if (p_buf)
+ strncpy(p_buf, buf, p_buf_len);
+ else
+ printk("%s", buf);
+
+ return ret;
+}
+
+static int transport_set_vpd_assoc(t10_vpd_t *vpd, unsigned char *page_83)
+{
+ /*
+ * The VPD identification association..
+ *
+ * from spc3r23.pdf Section 7.6.3.1 Table 297
+ */
+ vpd->association = (page_83[1] & 0x30);
+ return transport_dump_vpd_assoc(vpd, NULL, 0);
+}
+
+int transport_dump_vpd_ident_type(
+ t10_vpd_t *vpd,
+ unsigned char *p_buf,
+ int p_buf_len)
+{
+ unsigned char buf[VPD_TMP_BUF_SIZE];
+ int ret = 0, len;
+
+ memset(buf, 0, VPD_TMP_BUF_SIZE);
+ len = sprintf(buf, "T10 VPD Identifier Type: ");
+
+ switch (vpd->device_identifier_type) {
+ case 0x00:
+ sprintf(buf+len, "Vendor specific\n");
+ break;
+ case 0x01:
+ sprintf(buf+len, "T10 Vendor ID based\n");
+ break;
+ case 0x02:
+ sprintf(buf+len, "EUI-64 based\n");
+ break;
+ case 0x03:
+ sprintf(buf+len, "NAA\n");
+ break;
+ case 0x04:
+ sprintf(buf+len, "Relative target port identifier\n");
+ break;
+ case 0x08:
+ sprintf(buf+len, "SCSI name string\n");
+ break;
+ default:
+ sprintf(buf+len, "Unsupported: 0x%02x\n",
+ vpd->device_identifier_type);
+ ret = -1;
+ break;
+ }
+
+ if (p_buf)
+ strncpy(p_buf, buf, p_buf_len);
+ else
+ printk("%s", buf);
+
+ return ret;
+}
+
+int transport_set_vpd_ident_type(t10_vpd_t *vpd, unsigned char *page_83)
+{
+ /*
+ * The VPD identifier type..
+ *
+ * from spc3r23.pdf Section 7.6.3.1 Table 298
+ */
+ vpd->device_identifier_type = (page_83[1] & 0x0f);
+ return transport_dump_vpd_ident_type(vpd, NULL, 0);
+}
+
+int transport_dump_vpd_ident(
+ t10_vpd_t *vpd,
+ unsigned char *p_buf,
+ int p_buf_len)
+{
+ unsigned char buf[VPD_TMP_BUF_SIZE];
+ int ret = 0;
+
+ memset(buf, 0, VPD_TMP_BUF_SIZE);
+
+ switch (vpd->device_identifier_code_set) {
+ case 0x01: /* Binary */
+ sprintf(buf, "T10 VPD Binary Device Identifier: %s\n",
+ &vpd->device_identifier[0]);
+ break;
+ case 0x02: /* ASCII */
+ sprintf(buf, "T10 VPD ASCII Device Identifier: %s\n",
+ &vpd->device_identifier[0]);
+ break;
+ case 0x03: /* UTF-8 */
+ sprintf(buf, "T10 VPD UTF-8 Device Identifier: %s\n",
+ &vpd->device_identifier[0]);
+ break;
+ default:
+ sprintf(buf, "T10 VPD Device Identifier encoding unsupported:"
+ " 0x%02x", vpd->device_identifier_code_set);
+ ret = -1;
+ break;
+ }
+
+ if (p_buf)
+ strncpy(p_buf, buf, p_buf_len);
+ else
+ printk("%s", buf);
+
+ return ret;
+}
+
+int transport_set_vpd_ident(t10_vpd_t *vpd, unsigned char *page_83)
+{
+ int j = 0, i = 4; /* offset to start of the identifer */
+
+ /*
+ * The VPD Code Set (encoding)
+ *
+ * from spc3r23.pdf Section 7.6.3.1 Table 296
+ */
+ vpd->device_identifier_code_set = (page_83[0] & 0x0f);
+ switch (vpd->device_identifier_code_set) {
+ case 0x01: /* Binary */
+ vpd->device_identifier[j++] =
+ hex_str[vpd->device_identifier_type];
+ while (i < (4 + page_83[3])) {
+ vpd->device_identifier[j++] =
+ hex_str[(page_83[i] & 0xf0) >> 4];
+ vpd->device_identifier[j++] =
+ hex_str[page_83[i] & 0x0f];
+ i++;
+ }
+ break;
+ case 0x02: /* ASCII */
+ case 0x03: /* UTF-8 */
+ while (i < (4 + page_83[3]))
+ vpd->device_identifier[j++] = page_83[i++];
+
+ break;
+ default:
+ break;
+ }
+
+ return transport_dump_vpd_ident(vpd, NULL, 0);
+}
+
+static int transport_get_inquiry_vpd_device_ident(
+ se_obj_lun_type_t *obj_api,
+ t10_wwn_t *wwn,
+ void *obj_ptr)
+{
+ unsigned char *buf, *page_83;
+ se_cmd_t *cmd;
+ t10_vpd_t *vpd;
+ unsigned char cdb[SCSI_CDB_SIZE];
+ int ident_len, page_len, off = 4, ret = 0;
+
+ memset(cdb, 0, SCSI_CDB_SIZE);
+ cdb[0] = INQUIRY;
+ cdb[1] = 0x01; /* Query VPD */
+ cdb[2] = 0x83; /* Device Identifier */
+ cdb[3] = (INQUIRY_VPD_DEVICE_IDENTIFIER_LEN >> 8) & 0xff;
+ cdb[4] = (INQUIRY_VPD_DEVICE_IDENTIFIER_LEN & 0xff);
+
+ cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+ 0, NULL, 0, INQUIRY_VPD_DEVICE_IDENTIFIER_LEN,
+ obj_api, obj_ptr);
+ if (!(cmd))
+ return -1;
+
+ if (transport_generic_passthrough(cmd) < 0) {
+ transport_passthrough_release(cmd);
+ return -1;
+ }
+
+ buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+ page_len = (buf[2] << 8) | buf[3];
+ printk("T10 VPD Page Length: %d\n", page_len);
+
+ while (page_len > 0) {
+ /* Grab a pointer to the Identification descriptor */
+ page_83 = &buf[off];
+ ident_len = page_83[3];
+ if (!(ident_len)) {
+ printk(KERN_ERR "page_83[3]: identifier"
+ " length zero!\n");
+ break;
+ }
+ printk(KERN_INFO "T10 VPD Identifer Length: %d\n", ident_len);
+
+ vpd = kzalloc(sizeof(t10_vpd_t), GFP_KERNEL);
+ if (!(vpd)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " t10_vpd_t\n");
+ ret = -1;
+ goto out;
+ }
+ INIT_LIST_HEAD(&vpd->vpd_list);
+
+ transport_set_vpd_proto_id(vpd, page_83);
+ transport_set_vpd_assoc(vpd, page_83);
+
+ if (transport_set_vpd_ident_type(vpd, page_83) < 0) {
+ off += (ident_len + 4);
+ page_len -= (ident_len + 4);
+ kfree(vpd);
+ continue;
+ }
+ if (transport_set_vpd_ident(vpd, page_83) < 0) {
+ off += (ident_len + 4);
+ page_len -= (ident_len + 4);
+ kfree(vpd);
+ continue;
+ }
+
+ list_add_tail(&vpd->vpd_list, &wwn->t10_vpd_list);
+ off += (ident_len + 4);
+ page_len -= (ident_len + 4);
+ }
+out:
+ transport_passthrough_release(cmd);
+ return 0;
+}
+
+static int transport_get_read_capacity(se_device_t *dev)
+{
+ unsigned char cdb[SCSI_CDB_SIZE], *buf;
+ u32 blocks, v1, v2;
+ se_cmd_t *cmd;
+ unsigned long long blocks_long;
+
+ memset(cdb, 0, SCSI_CDB_SIZE);
+ cdb[0] = 0x25; /* READ_CAPACITY */
+
+ cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+ 0, NULL, 0, READ_CAP_LEN, DEV_OBJ_API(dev),
+ (void *)dev);
+ if (!(cmd))
+ return -1;
+
+ if (transport_generic_passthrough(cmd) < 0) {
+ transport_passthrough_release(cmd);
+ return -1;
+ }
+
+ buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+ blocks = (buf[0] << 24) | (buf[1] << 16) | (buf[2] << 8) | buf[3];
+
+ transport_passthrough_release(cmd);
+
+ if (blocks != 0xFFFFFFFF) {
+ dev->dev_sectors_total = blocks;
+ dev->dev_generate_cdb = &split_cdb_RW_10;
+ return 0;
+ }
+
+ printk(KERN_INFO "READ_CAPACITY returned 0xFFFFFFFF, issuing"
+ " SAI_READ_CAPACITY_16\n");
+
+ memset(cdb, 0, SCSI_CDB_SIZE);
+ cdb[0] = 0x9e; /* SERVICE_ACTION_IN */
+ cdb[1] = 0x10; /* SAI_READ_CAPACITY_16 */
+ cdb[13] = 12;
+
+ cmd = transport_allocate_passthrough(&cdb[0], SE_DIRECTION_READ,
+ 0, NULL, 0, 12, DEV_OBJ_API(dev), (void *)dev);
+ if (!(cmd))
+ return -1;
+
+ if (transport_generic_passthrough(cmd) < 0) {
+ transport_passthrough_release(cmd);
+ return -1;
+ }
+
+ buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+ v1 = (buf[0] << 24) | (buf[1] << 16) | (buf[2] << 8) | buf[3];
+ v2 = (buf[4] << 24) | (buf[5] << 16) | (buf[6] << 8) | buf[7];
+ blocks_long = ((unsigned long long)v2 | (unsigned long long)v1 << 32);
+
+ transport_passthrough_release(cmd);
+
+ dev->dev_sectors_total = blocks_long;
+ dev->dev_generate_cdb = &split_cdb_RW_16;
+
+ return 0;
+}
+
+/* transport_add_device_to_core_hba():
+ *
+ * Note that some plugins (IBLOCK) will pass device_flags ==
+ * DF_CLAIMED_BLOCKDEV signifying OS that a dependent block_device
+ * has been claimed. In exception cases we will release said
+ * block_device ourselves.
+ */
+se_device_t *transport_add_device_to_core_hba(
+ se_hba_t *hba,
+ se_subsystem_api_t *transport,
+ se_subsystem_dev_t *se_dev,
+ u32 device_flags,
+ void *transport_dev)
+{
+ int ret = 0;
+ se_device_t *dev;
+
+ dev = kzalloc(sizeof(se_device_t), GFP_KERNEL);
+ if (!(dev)) {
+ printk(KERN_ERR "Unable to allocate memory for se_dev_t\n");
+ return NULL;
+ }
+ dev->dev_queue_obj = kzalloc(sizeof(se_queue_obj_t), GFP_KERNEL);
+ if (!(dev->dev_queue_obj)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " dev->dev_queue_obj\n");
+ kfree(dev);
+ return NULL;
+ }
+ transport_init_queue_obj(dev->dev_queue_obj);
+
+ dev->dev_status_queue_obj = kzalloc(sizeof(se_queue_obj_t),
+ GFP_KERNEL);
+ if (!(dev->dev_status_queue_obj)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " dev->dev_status_queue_obj\n");
+ kfree(dev->dev_queue_obj);
+ kfree(dev);
+ return NULL;
+ }
+ transport_init_queue_obj(dev->dev_status_queue_obj);
+
+ dev->dev_flags = device_flags;
+ dev->dev_status |= TRANSPORT_DEVICE_DEACTIVATED;
+ dev->type = transport->type;
+ dev->dev_ptr = (void *) transport_dev;
+ dev->se_hba = hba;
+ dev->se_sub_dev = se_dev;
+ dev->transport = transport;
+ atomic_set(&dev->active_cmds, 0);
+ INIT_LIST_HEAD(&dev->dev_list);
+ INIT_LIST_HEAD(&dev->dev_sep_list);
+ INIT_LIST_HEAD(&dev->dev_tmr_list);
+ INIT_LIST_HEAD(&dev->execute_task_list);
+ INIT_LIST_HEAD(&dev->state_task_list);
+ spin_lock_init(&dev->execute_task_lock);
+ spin_lock_init(&dev->state_task_lock);
+ spin_lock_init(&dev->dev_reservation_lock);
+ spin_lock_init(&dev->dev_status_lock);
+ spin_lock_init(&dev->dev_status_thr_lock);
+ spin_lock_init(&dev->se_port_lock);
+ spin_lock_init(&dev->se_tmr_lock);
+
+ dev->queue_depth = TRANSPORT(dev)->get_queue_depth(dev);
+ atomic_set(&dev->depth_left, dev->queue_depth);
+
+ se_dev_set_default_attribs(dev);
+
+ dev->write_pending = (transport->write_pending) ?
+ transport->write_pending : &transport_dev_write_pending_nop;
+
+#ifdef SNMP_SUPPORT
+ dev->dev_index = scsi_get_new_index(SCSI_DEVICE_INDEX);
+ dev->creation_time = get_jiffies_64();
+ spin_lock_init(&dev->stats_lock);
+#endif /* SNMP_SUPPORT */
+
+ spin_lock(&hba->device_lock);
+ list_add_tail(&dev->dev_list, &hba->hba_dev_list);
+ hba->dev_count++;
+ spin_unlock(&hba->device_lock);
+
+ /*
+ * Get this se_device_t's API from the device object plugin.
+ */
+ dev->dev_obj_api = se_obj_get_api(TRANSPORT_LUN_TYPE_DEVICE);
+ if (!(dev->dev_obj_api))
+ goto out;
+ /*
+ * Setup the Reservations infrastructure for se_device_t
+ */
+ core_setup_reservations(dev);
+ /*
+ * Setup the Asymmetric Logical Unit Assignment for se_device_t
+ */
+ if (core_setup_alua(dev) < 0)
+ goto out;
+ /*
+ * Startup the se_device_t processing thread
+ */
+ if (transport_generic_activate_device(dev) < 0)
+ goto out;
+ /*
+ * This HBA is reserved for internal storage engine / feature plugin use
+ */
+ if (hba->hba_flags & HBA_FLAGS_INTERNAL_USE) {
+ ret = transport_get_read_capacity(dev);
+ if (ret < 0)
+ goto out;
+
+ return dev;
+ }
+
+ ret = transport_get_inquiry(DEV_OBJ_API(dev),
+ DEV_T10_WWN(dev), (void *)dev);
+ if (ret < 0)
+ goto out;
+ /*
+ * Locate VPD WWN Information used for various purposes within
+ * the Storage Engine.
+ */
+ if (!(transport_get_inquiry_vpd_serial(DEV_OBJ_API(dev),
+ DEV_T10_WWN(dev), (void *)dev))) {
+ /*
+ * If VPD Unit Serial returned GOOD status, try
+ * VPD Device Identification page (0x83).
+ */
+ transport_get_inquiry_vpd_device_ident(DEV_OBJ_API(dev),
+ DEV_T10_WWN(dev), (void *)dev);
+ }
+
+ /*
+ * Only perform the volume scan for peripheral type TYPE_DISK
+ */
+ if (TRANSPORT(dev)->get_device_type(dev) != 0)
+ return dev;
+
+ /*
+ * Get the sector count via READ_CAPACITY
+ */
+ ret = transport_get_read_capacity(dev);
+ if (ret < 0)
+ goto out;
+out:
+ if (!ret)
+ return dev;
+
+ /*
+ * Release claim to OS dependant block_device that may have been
+ * set by plugin with passed dev_flags.
+ */
+ transport_generic_release_phydevice(dev, 0);
+
+ /*
+ * Release newly allocated state for se_device_t
+ */
+ transport_generic_deactivate_device(dev);
+
+ spin_lock(&hba->device_lock);
+ list_del(&dev->dev_list);
+ hba->dev_count--;
+ spin_unlock(&hba->device_lock);
+
+ se_release_vpd_for_dev(dev);
+
+ kfree(dev->dev_status_queue_obj);
+ kfree(dev->dev_queue_obj);
+ kfree(dev);
+
+ return NULL;
+}
+
+/* transport_generic_activate_device():
+ *
+ *
+ */
+int transport_generic_activate_device(se_device_t *dev)
+{
+ char name[16];
+
+ if (TRANSPORT(dev)->activate_device)
+ TRANSPORT(dev)->activate_device(dev);
+
+ memset(name, 0, 16);
+ snprintf(name, 16, "LIO_%s", TRANSPORT(dev)->name);
+
+ dev->process_thread = kthread_run(transport_processing_thread,
+ (void *)dev, name);
+ if (IS_ERR(dev->process_thread)) {
+ printk(KERN_ERR "Unable to create kthread: %s\n", name);
+ return -1;
+ }
+
+ down(&dev->dev_queue_obj->thread_create_sem);
+
+ return 0;
+}
+
+/* transport_generic_deactivate_device():
+ *
+ *
+ */
+void transport_generic_deactivate_device(se_device_t *dev)
+{
+ if (TRANSPORT(dev)->deactivate_device)
+ TRANSPORT(dev)->deactivate_device(dev);
+
+ kthread_stop(dev->process_thread);
+
+ down(&dev->dev_queue_obj->thread_done_sem);
+}
+
+/* transport_generic_claim_phydevice()
+ *
+ * Obtain exclusive access to OS dependant block-device via
+ * Storage Transport Plugin API.
+ *
+ * In Linux v2.6 this means calling fs/block_dev.c:bd_claim()
+ * that is called in an plugin dependent method for claiming
+ * struct block_device.
+ *
+ * Returns 0 - Already claimed or not able to claim
+ * Returns 1 - Successfuly claimed
+ * Returns < 0 - Error
+ */
+int transport_generic_claim_phydevice(se_device_t *dev)
+{
+ int ret;
+ se_hba_t *hba;
+
+ /*
+ * This function pointer is present when handling access
+ * control to a OS dependant block subsystem.
+ */
+ if (!TRANSPORT(dev)->claim_phydevice)
+ return 0;
+
+ if (dev->dev_flags & DF_READ_ONLY)
+ return 0;
+
+ if (dev->dev_flags & DF_CLAIMED_BLOCKDEV)
+ return 0;
+
+ hba = dev->se_hba;
+ if (!(hba)) {
+ printk(KERN_ERR "se_device_t->se_hba is NULL!\n");
+ return -1;
+ }
+
+ ret = TRANSPORT(dev)->claim_phydevice(hba, dev);
+ if (ret < 0)
+ return ret;
+
+ dev->dev_flags |= DF_CLAIMED_BLOCKDEV;
+
+ return 1;
+}
+EXPORT_SYMBOL(transport_generic_claim_phydevice);
+
+/* transport_generic_release_phydevice():
+ *
+ * Release exclusive access from OS dependant block-device via
+ * Storage Transport Plugin API.
+ *
+ * In Linux v2.6 this means calling fs/block_dev.c:bd_release()
+ * see iscsi_target_pscsi.c and iscsi_target_iblock.c functions for
+ * se_subsystem_api_t->[claim,release]_phydevice()
+ */
+void transport_generic_release_phydevice(se_device_t *dev, int check_pscsi)
+{
+ if (!TRANSPORT(dev)->release_phydevice)
+ return;
+
+ if (dev->dev_flags & DF_READ_ONLY) {
+ if (check_pscsi &&
+ (TRANSPORT(dev)->transport_type !=
+ TRANSPORT_PLUGIN_PHBA_PDEV))
+ return;
+
+ TRANSPORT(dev)->release_phydevice(dev);
+ return;
+ }
+
+ if (!(dev->dev_flags & DF_CLAIMED_BLOCKDEV))
+ return;
+
+ if (!dev->dev_ptr) {
+ printk(KERN_ERR "se_device_t->dev_ptr is NULL!\n");
+ BUG();
+ }
+
+ if (check_pscsi) {
+ if (TRANSPORT(dev)->transport_type !=
+ TRANSPORT_PLUGIN_PHBA_PDEV)
+ return;
+
+ if (dev->dev_flags & DF_PERSISTENT_CLAIMED_BLOCKDEV)
+ return;
+ }
+
+ TRANSPORT(dev)->release_phydevice(dev);
+ dev->dev_flags &= ~DF_CLAIMED_BLOCKDEV;
+}
+
+/* transport_generic_free_device():
+ *
+ *
+ */
+void transport_generic_free_device(se_device_t *dev)
+{
+ if (!(dev->dev_ptr))
+ return;
+
+ transport_generic_deactivate_device(dev);
+
+ transport_generic_release_phydevice(dev, 0);
+
+ if (TRANSPORT(dev)->free_device)
+ TRANSPORT(dev)->free_device(dev->dev_ptr);
+}
+EXPORT_SYMBOL(transport_generic_free_device);
+
+int transport_allocate_iovecs_for_cmd(
+ se_cmd_t *cmd,
+ u32 iov_count)
+{
+ cmd->iov_data = kzalloc(iov_count * sizeof(struct iovec), GFP_ATOMIC);
+ if (!(cmd->iov_data)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " iscsi_cmd_t->iov_data.\n");
+ return -1;
+ }
+ cmd->orig_iov_data_count = iov_count;
+
+ return 0;
+}
+EXPORT_SYMBOL(transport_allocate_iovecs_for_cmd);
+
+/* transport_generic_allocate_iovecs():
+ *
+ * Called from transport_generic_new_cmd() in Transport Processing Thread.
+ */
+static int transport_generic_allocate_iovecs(
+ se_cmd_t *cmd)
+{
+ u32 iov_count;
+
+ iov_count = T_TASK(cmd)->t_task_se_num;
+ if (!(iov_count))
+ iov_count = 1;
+#if 0
+ printk(KERN_INFO "Allocated %d iovecs for ITT: 0x%08x t_task_se_num:"
+ " %u\n", iov_count, CMD_TFO(cmd)->get_task_tag(cmd),
+ T_TASK(cmd)->t_task_se_num);
+#endif
+ iov_count += TRANSPORT_IOV_DATA_BUFFER;
+
+ if (transport_allocate_iovecs_for_cmd(cmd, iov_count))
+ return -1;
+
+ return 0;
+}
+
+/* transport_generic_prepare_cdb():
+ *
+ * Since the Initiator sees iSCSI devices as LUNs, the SCSI CDB will
+ * contain the iSCSI LUN in bits 7-5 of byte 1 as per SAM-2.
+ * The point of this is since we are mapping iSCSI LUNs to
+ * SCSI Target IDs having a non-zero LUN in the CDB will throw the
+ * devices and HBAs for a loop.
+ */
+static inline void transport_generic_prepare_cdb(
+ unsigned char *cdb)
+{
+ switch (cdb[0]) {
+ case READ_10: /* SBC - RDProtect */
+ case READ_12: /* SBC - RDProtect */
+ case READ_16: /* SBC - RDProtect */
+ case SEND_DIAGNOSTIC: /* SPC - SELF-TEST Code */
+ case VERIFY: /* SBC - VRProtect */
+ case VERIFY_16: /* SBC - VRProtect */
+ case WRITE_VERIFY: /* SBC - VRProtect */
+ case WRITE_VERIFY_12: /* SBC - VRProtect */
+ break;
+ default:
+ cdb[1] &= 0x1f; /* clear logical unit number */
+ break;
+ }
+}
+
+/* transport_check_device_cdb_sector_count():
+ *
+ * returns:
+ * 0 on supported request sector count.
+ * 1 on unsupported request sector count.
+ */
+static inline int transport_check_device_cdb_sector_count(
+ se_obj_lun_type_t *se_obj_api,
+ void *se_obj_ptr,
+ u32 sectors)
+{
+ u32 max_sectors;
+
+ max_sectors = se_obj_api->max_sectors(se_obj_ptr);
+ if (!(max_sectors)) {
+ printk(KERN_ERR "TRANSPORT->get_max_sectors returned zero!\n");
+ return 1;
+ }
+
+ if (sectors > max_sectors)
+ return -1;
+
+ return 0;
+}
+
+/* transport_generic_get_task():
+ *
+ *
+ */
+static se_task_t *transport_generic_get_task(
+ se_transform_info_t *ti,
+ se_cmd_t *cmd,
+ void *se_obj_ptr,
+ se_obj_lun_type_t *se_obj_api)
+{
+ se_task_t *task;
+ unsigned long flags;
+
+ task = kmem_cache_zalloc(se_task_cache, GFP_KERNEL);
+ if (!(task)) {
+ printk(KERN_ERR "Unable to allocate se_task_t\n");
+ return NULL;
+ }
+
+ INIT_LIST_HEAD(&task->t_list);
+ INIT_LIST_HEAD(&task->t_execute_list);
+ INIT_LIST_HEAD(&task->t_state_list);
+ init_MUTEX_LOCKED(&task->task_stop_sem);
+ task->task_no = T_TASK(cmd)->t_task_no++;
+ task->task_se_cmd = cmd;
+
+ DEBUG_SO("se_obj_ptr: %p\n", se_obj_ptr);
+ DEBUG_SO("se_obj_api: %p\n", se_obj_api);
+ DEBUG_SO("Plugin: %s\n", se_obj_api->obj_plugin->plugin_name);
+
+ task->transport_req = se_obj_api->get_transport_req(se_obj_ptr, task);
+ if (!(task->transport_req))
+ return NULL;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ list_add_tail(&task->t_list, &T_TASK(cmd)->t_task_list);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ task->se_obj_api = se_obj_api;
+ task->se_obj_ptr = se_obj_ptr;
+
+ return task;
+}
+
+int transport_generic_obj_start(
+ se_transform_info_t *ti,
+ se_obj_lun_type_t *obj_api,
+ void *p,
+ unsigned long long starting_lba)
+{
+ ti->ti_lba = starting_lba;
+ ti->ti_obj_api = obj_api;
+ ti->ti_obj_ptr = p;
+
+ return 0;
+}
+
+static int transport_process_data_sg_transform(
+ se_cmd_t *cmd,
+ se_transform_info_t *ti)
+{
+ /*
+ * Already handled in transport_generic_get_cdb_count()
+ */
+ return 0;
+}
+
+/* transport_process_control_sg_transform():
+ *
+ *
+ */
+static int transport_process_control_sg_transform(
+ se_cmd_t *cmd,
+ se_transform_info_t *ti)
+{
+ unsigned char *cdb;
+ se_task_t *task;
+ se_mem_t *se_mem, *se_mem_lout = NULL;
+ int ret;
+ u32 se_mem_cnt = 0, task_offset = 0;
+
+ list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_list, se_list)
+ break;
+
+ if (!se_mem) {
+ printk(KERN_ERR "se_mem is NULL!\n");
+ return -1;
+ }
+
+ task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
+ ti->se_obj_api);
+ if (!(task))
+ return -1;
+
+ task->transport_map_task = ti->se_obj_api->get_map_SG(
+ ti->se_obj_ptr, cmd->data_direction);
+
+ cdb = ti->se_obj_api->get_cdb(ti->se_obj_ptr, task);
+ if (cdb)
+ memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+
+ task->task_size = cmd->data_length;
+ task->task_sg_num = 1;
+
+ atomic_inc(&T_TASK(cmd)->t_fe_count);
+ atomic_inc(&T_TASK(cmd)->t_se_count);
+
+ ret = ti->se_obj_api->do_se_mem_map(ti->se_obj_ptr, task,
+ T_TASK(cmd)->t_mem_list, NULL, se_mem, &se_mem_lout,
+ &se_mem_cnt, &task_offset);
+ if (ret < 0)
+ return ret;
+
+ DEBUG_CDB_H("task_no[%u]: SCF_SCSI_CONTROL_SG_IO_CDB task_size: %d\n",
+ task->task_no, task->task_size);
+ return 0;
+}
+
+/* transport_process_control_nonsg_transform():
+ *
+ *
+ */
+static int transport_process_control_nonsg_transform(
+ se_cmd_t *cmd,
+ se_transform_info_t *ti)
+{
+ unsigned char *cdb;
+ se_task_t *task;
+
+ task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
+ ti->se_obj_api);
+ if (!(task))
+ return -1;
+
+ task->transport_map_task = ti->se_obj_api->get_map_non_SG(
+ ti->se_obj_ptr, cmd->data_direction);
+
+ cdb = ti->se_obj_api->get_cdb(ti->se_obj_ptr, task);
+ if (cdb)
+ memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+
+ task->task_size = cmd->data_length;
+ task->task_sg_num = 0;
+
+ atomic_inc(&T_TASK(cmd)->t_fe_count);
+ atomic_inc(&T_TASK(cmd)->t_se_count);
+
+ DEBUG_CDB_H("task_no[%u]: SCF_SCSI_CONTROL_NONSG_IO_CDB task_size:"
+ " %d\n", task->task_no, task->task_size);
+ return 0;
+}
+
+/* transport_process_non_data_transform():
+ *
+ *
+ */
+static int transport_process_non_data_transform(
+ se_cmd_t *cmd,
+ se_transform_info_t *ti)
+{
+ unsigned char *cdb;
+ se_task_t *task;
+
+ task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
+ ti->se_obj_api);
+ if (!(task))
+ return -1;
+
+ task->transport_map_task = ti->se_obj_api->get_map_none(ti->se_obj_ptr);
+
+ cdb = ti->se_obj_api->get_cdb(ti->se_obj_ptr, task);
+ if (cdb)
+ memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+
+ task->task_size = cmd->data_length;
+ task->task_sg_num = 0;
+
+ atomic_inc(&T_TASK(cmd)->t_fe_count);
+ atomic_inc(&T_TASK(cmd)->t_se_count);
+
+ DEBUG_CDB_H("task_no[%u]: SCF_SCSI_NON_DATA_CDB task_size: %d\n",
+ task->task_no, task->task_size);
+ return 0;
+}
+
+static int transport_generic_cmd_sequencer(se_cmd_t *, unsigned char *);
+
+void transport_device_setup_cmd(se_cmd_t *cmd)
+{
+ cmd->transport_add_cmd_to_queue = &transport_add_cmd_to_dev_queue;
+ cmd->se_dev = SE_LUN(cmd)->se_dev;
+}
+
+se_cmd_t *__transport_alloc_se_cmd(
+ struct target_core_fabric_ops *tfo,
+ se_session_t *se_sess,
+ void *fabric_cmd_ptr,
+ u32 data_length,
+ int data_direction)
+{
+ se_cmd_t *cmd;
+
+ if (data_direction == SE_DIRECTION_BIDI) {
+ printk(KERN_ERR "SCSI BiDirectional mode not supported yet\n");
+ return ERR_PTR(-ENOSYS);
+ }
+
+ cmd = kmem_cache_zalloc(se_cmd_cache, GFP_KERNEL);
+ if (!(cmd)) {
+ printk(KERN_ERR "kmem_cache_alloc() failed for se_cmd_cache\n");
+ return ERR_PTR(-ENOMEM);
+ }
+ INIT_LIST_HEAD(&cmd->se_lun_list);
+
+ cmd->t_task = kzalloc(sizeof(se_transport_task_t), GFP_KERNEL);
+ if (!(cmd->t_task)) {
+ printk(KERN_ERR "Unable to allocate cmd->t_task\n");
+ kmem_cache_free(se_cmd_cache, cmd);
+ return NULL;
+ }
+
+ cmd->sense_buffer = kzalloc(
+ TRANSPORT_SENSE_BUFFER + tfo->get_fabric_sense_len(),
+ GFP_KERNEL);
+ if (!(cmd->sense_buffer)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " cmd->sense_buffer\n");
+ kfree(cmd->t_task);
+ kmem_cache_free(se_cmd_cache, cmd);
+ return NULL;
+ }
+ INIT_LIST_HEAD(&T_TASK(cmd)->t_task_list);
+ init_MUTEX_LOCKED(&T_TASK(cmd)->transport_lun_fe_stop_sem);
+ init_MUTEX_LOCKED(&T_TASK(cmd)->transport_lun_stop_sem);
+ init_MUTEX_LOCKED(&T_TASK(cmd)->t_transport_stop_sem);
+ init_MUTEX_LOCKED(&T_TASK(cmd)->t_transport_passthrough_sem);
+ init_MUTEX_LOCKED(&T_TASK(cmd)->t_transport_passthrough_wsem);
+ spin_lock_init(&T_TASK(cmd)->t_state_lock);
+
+ cmd->se_tfo = tfo;
+ cmd->se_sess = se_sess;
+ cmd->se_fabric_cmd_ptr = fabric_cmd_ptr;
+ cmd->data_length = data_length;
+ cmd->data_direction = data_direction;
+
+ return cmd;
+}
+
+se_cmd_t *transport_alloc_se_cmd(
+ struct target_core_fabric_ops *tfo_api,
+ se_session_t *se_sess,
+ void *fabric_cmd_ptr,
+ u32 data_length,
+ int data_direction)
+{
+ return __transport_alloc_se_cmd(tfo_api, se_sess, fabric_cmd_ptr,
+ data_length, data_direction);
+}
+EXPORT_SYMBOL(transport_alloc_se_cmd);
+
+void transport_free_se_cmd(
+ se_cmd_t *se_cmd)
+{
+ if (se_cmd->se_tmr_req)
+ core_tmr_release_req(se_cmd->se_tmr_req);
+
+ kfree(se_cmd->iov_data);
+ kfree(se_cmd->sense_buffer);
+ kfree(se_cmd->t_task);
+ kmem_cache_free(se_cmd_cache, se_cmd);
+}
+EXPORT_SYMBOL(transport_free_se_cmd);
+
+static void transport_generic_wait_for_tasks(se_cmd_t *, int, int);
+
+/* transport_generic_allocate_tasks():
+ *
+ * Called from the iSCSI RX Thread.
+ */
+int transport_generic_allocate_tasks(
+ se_cmd_t *cmd,
+ unsigned char *cdb)
+{
+ int non_data_cdb;
+
+ transport_generic_prepare_cdb(cdb);
+
+ /*
+ * This is needed for early exceptions.
+ */
+ cmd->transport_wait_for_tasks = &transport_generic_wait_for_tasks;
+
+ CMD_ORIG_OBJ_API(cmd)->transport_setup_cmd(cmd->se_orig_obj_ptr, cmd);
+ /*
+ * See if this is a CDB which follows SAM, also grab a function
+ * pointer to see if we need to do extra work.
+ */
+ non_data_cdb = transport_generic_cmd_sequencer(cmd, cdb);
+ if (non_data_cdb < 0)
+ return -1;
+ /*
+ * Copy the original CDB into T_TASK(cmd).
+ */
+ memcpy(T_TASK(cmd)->t_task_cdb, cdb, SCSI_CDB_SIZE);
+
+#ifdef SNMP_SUPPORT
+ spin_lock(&cmd->se_lun->lun_sep_lock);
+ if (cmd->se_lun->lun_sep)
+ cmd->se_lun->lun_sep->sep_stats.cmd_pdus++;
+ spin_unlock(&cmd->se_lun->lun_sep_lock);
+#endif /* SNMP_SUPPORT */
+
+ switch (non_data_cdb) {
+ case 0:
+ DEBUG_CDB_H("Set cdb[0]: 0x%02x to "
+ "SCF_SCSI_DATA_SG_IO_CDB\n", cdb[0]);
+ cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
+
+ /*
+ * Get the initial Logical Block Address from the Original
+ * Command Descriptor Block that arrived on the iSCSI wire.
+ */
+ T_TASK(cmd)->t_task_lba = (cmd->transport_get_long_lba) ?
+ cmd->transport_get_long_lba(cdb) :
+ cmd->transport_get_lba(cdb);
+
+ break;
+ case 1:
+ DEBUG_CDB_H("Set cdb[0]: 0x%02x to"
+ " SCF_SCSI_CONTROL_SG_IO_CDB\n", cdb[0]);
+ cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
+ cmd->transport_cdb_transform =
+ &transport_process_control_sg_transform;
+ break;
+ case 2:
+ DEBUG_CDB_H("Set cdb[0]: 0x%02x to "
+ "SCF_SCSI_CONTROL_NONSG_IO_CDB\n", cdb[0]);
+ cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+ cmd->transport_cdb_transform =
+ &transport_process_control_nonsg_transform;
+ break;
+ case 3:
+ DEBUG_CDB_H("Set cdb[0]: 0x%02x to "
+ "SCF_SCSI_NON_DATA_CDB\n", cdb[0]);
+ cmd->se_cmd_flags |= SCF_SCSI_NON_DATA_CDB;
+ cmd->transport_cdb_transform =
+ &transport_process_non_data_transform;
+ break;
+ case 4:
+ DEBUG_CDB_H("Set cdb[0]: 0x%02x to"
+ " SCF_SCSI_UNSUPPORTED_CDB\n", cdb[0]);
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+ return -2;
+ case 5:
+ DEBUG_CDB_H("Set cdb[0]: 0x%02x to"
+ " SCF_SCSI_RESERVATION_CONFLICT\n", cdb[0]);
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->se_cmd_flags |= SCF_SCSI_RESERVATION_CONFLICT;
+ cmd->scsi_status = SAM_STAT_RESERVATION_CONFLICT;
+ /*
+ * For UA Interlock Code 11b, a RESERVATION CONFLICT will
+ * establish a UNIT ATTENTION with PREVIOUS RESERVATION
+ * CONFLICT STATUS.
+ *
+ * See spc4r17, section 7.4.6 Control Mode Page, Table 349
+ */
+ if (SE_SESS(cmd) &&
+ DEV_ATTRIB(cmd->se_dev)->emulate_ua_intlck_ctrl == 2)
+ core_scsi3_ua_allocate(SE_SESS(cmd)->se_node_acl,
+ cmd->orig_fe_lun, 0x2C,
+ ASCQ_2CH_PREVIOUS_RESERVATION_CONFLICT_STATUS);
+ return -2;
+ case 6:
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->scsi_sense_reason = INVALID_CDB_FIELD;
+ return -2;
+ case 7:
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->scsi_sense_reason = ILLEGAL_REQUEST;
+ return -2;
+ case 8:
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->scsi_sense_reason = CHECK_CONDITION_UNIT_ATTENTION;
+ return -2;
+ default:
+ break;
+ }
+
+ return 0;
+}
+EXPORT_SYMBOL(transport_generic_allocate_tasks);
+
+/* transport_generic_handle_cdb():
+ *
+ *
+ */
+int transport_generic_handle_cdb(
+ se_cmd_t *cmd)
+{
+ if (!SE_LUN(cmd)) {
+ printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
+ return -1;
+ }
+
+ cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_NEW_CMD);
+ return 0;
+}
+EXPORT_SYMBOL(transport_generic_handle_cdb);
+
+/* transport_generic_handle_data():
+ *
+ *
+ */
+int transport_generic_handle_data(
+ se_cmd_t *cmd)
+{
+ /*
+ * Make sure that the transport has been disabled by
+ * transport_write_pending() before readding this se_cmd_t to the
+ * processing queue. If it has not yet been reset to zero by the
+ * processing thread in cmd->transport_add_cmd_to_queue(), let other
+ * processes run. If a signal was received, then we assume the
+ * connection is being failed/shutdown, so we return a failure.
+ */
+ while (atomic_read(&T_TASK(cmd)->t_transport_active)) {
+ msleep_interruptible(10);
+ if (signal_pending(current))
+ return -1;
+ }
+
+ cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_PROCESS_WRITE);
+ return 0;
+}
+EXPORT_SYMBOL(transport_generic_handle_data);
+
+/* transport_generic_handle_tmr():
+ *
+ *
+ */
+int transport_generic_handle_tmr(
+ se_cmd_t *cmd)
+{
+ /*
+ * This is needed for early exceptions.
+ */
+ cmd->transport_wait_for_tasks = &transport_generic_wait_for_tasks;
+ CMD_ORIG_OBJ_API(cmd)->transport_setup_cmd(cmd->se_orig_obj_ptr, cmd);
+
+ cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_PROCESS_TMR);
+ return 0;
+}
+EXPORT_SYMBOL(transport_generic_handle_tmr);
+
+/* transport_stop_tasks_for_cmd():
+ *
+ *
+ */
+void transport_stop_tasks_for_cmd(se_cmd_t *cmd)
+{
+ se_task_t *task, *task_tmp;
+ unsigned long flags;
+
+ DEBUG_TS("ITT[0x%08x] - Stopping tasks\n",
+ CMD_TFO(cmd)->get_task_tag(cmd));
+
+ /*
+ * No tasks remain in the execution queue
+ */
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ list_for_each_entry_safe(task, task_tmp,
+ &T_TASK(cmd)->t_task_list, t_list) {
+ DEBUG_TS("task_no[%d] - Processing task %p\n",
+ task->task_no, task);
+ /*
+ * If the se_task_t has not been sent and is not active,
+ * remove the se_task_t from the execution queue.
+ */
+ if (!atomic_read(&task->task_sent) &&
+ !atomic_read(&task->task_active)) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+ transport_remove_task_from_execute_queue(task,
+ task->se_dev);
+
+ DEBUG_TS("task_no[%d] - Removed from execute queue\n",
+ task->task_no);
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ continue;
+ }
+
+ /*
+ * If the se_task_t is active, sleep until it is returned
+ * from the plugin.
+ */
+ if (atomic_read(&task->task_active)) {
+ atomic_set(&task->task_stop, 1);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+
+ DEBUG_TS("task_no[%d] - Waiting to complete\n",
+ task->task_no);
+ down(&task->task_stop_sem);
+ DEBUG_TS("task_no[%d] - Stopped successfully\n",
+ task->task_no);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+
+ atomic_set(&task->task_active, 0);
+ atomic_set(&task->task_stop, 0);
+ } else {
+ DEBUG_TS("task_no[%d] - Did nothing\n", task->task_no);
+ }
+
+ __transport_stop_task_timer(task, &flags);
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+static void transport_failure_reset_queue_depth(se_device_t *dev)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);;
+ atomic_inc(&dev->depth_left);
+ atomic_inc(&SE_HBA(dev)->left_queue_depth);
+ spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+}
+
+/* transport_generic_request_failure():
+ *
+ * Handle SAM-esque emulation for generic transport request failures.
+ */
+void transport_generic_request_failure(
+ se_cmd_t *cmd,
+ se_device_t *dev,
+ int complete,
+ int sc)
+{
+ DEBUG_GRF("-----[ Storage Engine Exception for cmd: %p ITT: 0x%08x"
+ " CDB: 0x%02x\n", cmd, CMD_TFO(cmd)->get_task_tag(cmd),
+ T_TASK(cmd)->t_task_cdb[0]);
+ DEBUG_GRF("-----[ se_obj_api: %p se_obj_ptr: %p\n", cmd->se_obj_api,
+ cmd->se_obj_ptr);
+ DEBUG_GRF("-----[ se_orig_obj_api: %p se_orig_obj_ptr: %p\n",
+ cmd->se_orig_obj_api, cmd->se_orig_obj_ptr);
+ DEBUG_GRF("-----[ i_state/def_i_state: %d/%d t_state/def_t_state:"
+ " %d/%d transport_error_status: %d\n",
+ CMD_TFO(cmd)->get_cmd_state(cmd), cmd->deferred_i_state,
+ cmd->t_state, cmd->deferred_t_state,
+ cmd->transport_error_status);
+ DEBUG_GRF("-----[ t_task_cdbs: %d t_task_cdbs_left: %d"
+ " t_task_cdbs_sent: %d t_task_cdbs_ex_left: %d --"
+ " t_transport_active: %d t_transport_stop: %d"
+ " t_transport_sent: %d\n", T_TASK(cmd)->t_task_cdbs,
+ atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
+ atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
+ atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left),
+ atomic_read(&T_TASK(cmd)->t_transport_active),
+ atomic_read(&T_TASK(cmd)->t_transport_stop),
+ atomic_read(&T_TASK(cmd)->t_transport_sent));
+
+ transport_stop_all_task_timers(cmd);
+
+ if (dev)
+ transport_failure_reset_queue_depth(dev);
+
+ if (complete) {
+ transport_direct_request_timeout(cmd);
+ cmd->transport_error_status =
+ PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ }
+
+ switch (cmd->transport_error_status) {
+ case PYX_TRANSPORT_UNKNOWN_SAM_OPCODE:
+ cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+ break;
+ case PYX_TRANSPORT_REQ_TOO_MANY_SECTORS:
+ cmd->scsi_sense_reason = SECTOR_COUNT_TOO_MANY;
+ break;
+ case PYX_TRANSPORT_INVALID_CDB_FIELD:
+ cmd->scsi_sense_reason = INVALID_CDB_FIELD;
+ break;
+ case PYX_TRANSPORT_INVALID_PARAMETER_LIST:
+ cmd->scsi_sense_reason = INVALID_PARAMETER_LIST;
+ break;
+ case PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES:
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)) {
+ if (!sc)
+ transport_new_cmd_failure(cmd);
+ /*
+ * Currently for PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES,
+ * we force this session to fall back to session
+ * recovery.
+ */
+ CMD_TFO(cmd)->fall_back_to_erl0(cmd->se_sess);
+ CMD_TFO(cmd)->stop_session(cmd->se_sess, 0, 0);
+
+ goto check_stop;
+ } else {
+ cmd->scsi_sense_reason =
+ LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ }
+ break;
+ case PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE:
+ cmd->scsi_sense_reason = LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ break;
+ case PYX_TRANSPORT_UNKNOWN_MODE_PAGE:
+ cmd->scsi_sense_reason = UNKNOWN_MODE_PAGE;
+ break;
+ case PYX_TRANSPORT_WRITE_PROTECTED:
+ cmd->scsi_sense_reason = WRITE_PROTECTED;
+ break;
+ case PYX_TRANSPORT_RESERVATION_CONFLICT:
+ /*
+ * No SENSE Data payload for this case, set SCSI Status
+ * and queue the response to $FABRIC_MOD.
+ *
+ * Uses linux/include/scsi/scsi.h SAM status codes defs
+ */
+ cmd->scsi_status = SAM_STAT_RESERVATION_CONFLICT;
+ /*
+ * For UA Interlock Code 11b, a RESERVATION CONFLICT will
+ * establish a UNIT ATTENTION with PREVIOUS RESERVATION
+ * CONFLICT STATUS.
+ *
+ * See spc4r17, section 7.4.6 Control Mode Page, Table 349
+ */
+ if (SE_SESS(cmd) &&
+ DEV_ATTRIB(cmd->se_dev)->emulate_ua_intlck_ctrl == 2)
+ core_scsi3_ua_allocate(SE_SESS(cmd)->se_node_acl,
+ cmd->orig_fe_lun, 0x2C,
+ ASCQ_2CH_PREVIOUS_RESERVATION_CONFLICT_STATUS);
+
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+ CMD_TFO(cmd)->queue_status(cmd);
+
+ goto check_stop;
+ default:
+ printk(KERN_ERR "Unknown transport error for CDB 0x%02x: %d\n",
+ T_TASK(cmd)->t_task_cdb[0],
+ cmd->transport_error_status);
+ cmd->scsi_sense_reason = UNSUPPORTED_SCSI_OPCODE;
+ break;
+ }
+
+ if (!sc)
+ transport_new_cmd_failure(cmd);
+ else
+ transport_send_check_condition_and_sense(cmd,
+ cmd->scsi_sense_reason, 0);
+check_stop:
+ transport_lun_remove_cmd(cmd);
+ if (!(transport_cmd_check_stop(cmd, 2, 0)))
+ transport_passthrough_check_stop(cmd);
+}
+
+void transport_direct_request_timeout(se_cmd_t *cmd)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (!(atomic_read(&T_TASK(cmd)->t_transport_timeout))) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+ if (atomic_read(&T_TASK(cmd)->t_task_cdbs_timeout_left)) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+
+ atomic_sub(atomic_read(&T_TASK(cmd)->t_transport_timeout),
+ &T_TASK(cmd)->t_se_count);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+void transport_generic_request_timeout(se_cmd_t *cmd)
+{
+ unsigned long flags;
+
+ /*
+ * Reset T_TASK(cmd)->t_se_count to allow transport_generic_remove()
+ * to allow last call to free memory resources.
+ */
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (atomic_read(&T_TASK(cmd)->t_transport_timeout) > 1) {
+ int tmp = (atomic_read(&T_TASK(cmd)->t_transport_timeout) - 1);
+
+ atomic_sub(tmp, &T_TASK(cmd)->t_se_count);
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ transport_generic_remove(cmd, 0, 0);
+}
+
+/* #define iscsi_linux_calculate_map_segment_DEBUG */
+#ifdef iscsi_linux_calculate_map_segment_DEBUG
+#define DEBUG_MAP_SEGMENTS(buf...) PYXPRINT(buf)
+#else
+#define DEBUG_MAP_SEGMENTS(buf...)
+#endif
+
+/* transport_calculate_map_segment():
+ *
+ *
+ */
+static inline void transport_calculate_map_segment(
+ u32 *data_length,
+ se_offset_map_t *lm)
+{
+ u32 sg_offset = 0;
+ se_mem_t *se_mem = lm->map_se_mem;
+
+ DEBUG_MAP_SEGMENTS(" START Mapping se_mem: %p, Length: %d"
+ " Remaining iSCSI Data: %u\n", se_mem, se_mem->se_len,
+ *data_length);
+ /*
+ * Still working on pages in the current se_mem_t.
+ */
+ if (!lm->map_reset) {
+ lm->iovec_length = (lm->sg_length > PAGE_SIZE) ?
+ PAGE_SIZE : lm->sg_length;
+ if (*data_length < lm->iovec_length) {
+ DEBUG_MAP_SEGMENTS("LINUX_MAP: Reset lm->iovec_length"
+ " to %d\n", *data_length);
+
+ lm->iovec_length = *data_length;
+ }
+ lm->iovec_base = page_address(lm->sg_page) + sg_offset;
+
+ DEBUG_MAP_SEGMENTS("LINUX_MAP: Set lm->iovec_base to %p from"
+ " lm->sg_page: %p\n", lm->iovec_base, lm->sg_page);
+ return;
+ }
+
+ /*
+ * First run of an iscsi_linux_map_t.
+ *
+ * OR:
+ *
+ * Mapped all of the pages in the current scatterlist, move
+ * on to the next one.
+ */
+ lm->map_reset = 0;
+ sg_offset = se_mem->se_off;
+ lm->sg_page = se_mem->se_page;
+ lm->sg_length = se_mem->se_len;
+
+ DEBUG_MAP_SEGMENTS("LINUX_MAP1[%p]: Starting to se_mem->se_len: %u,"
+ " se_mem->se_off: %u, se_mem->se_page: %p\n", se_mem,
+ se_mem->se_len, se_mem->se_off, se_mem->se_page);;
+ /*
+ * Get the base and length of the current page for use with the iovec.
+ */
+recalc:
+ lm->iovec_length = (lm->sg_length > (PAGE_SIZE - sg_offset)) ?
+ (PAGE_SIZE - sg_offset) : lm->sg_length;
+
+ DEBUG_MAP_SEGMENTS("LINUX_MAP: lm->iovec_length: %u, lm->sg_length: %u,"
+ " sg_offset: %u\n", lm->iovec_length, lm->sg_length, sg_offset);
+ /*
+ * See if there is any iSCSI offset we need to deal with.
+ */
+ if (!lm->current_offset) {
+ lm->iovec_base = page_address(lm->sg_page) + sg_offset;
+
+ if (*data_length < lm->iovec_length) {
+ DEBUG_MAP_SEGMENTS("LINUX_MAP1[%p]: Reset"
+ " lm->iovec_length to %d\n", se_mem,
+ *data_length);
+ lm->iovec_length = *data_length;
+ }
+
+ DEBUG_MAP_SEGMENTS("LINUX_MAP2[%p]: No current_offset,"
+ " set iovec_base to %p and set Current Page to %p\n",
+ se_mem, lm->iovec_base, lm->sg_page);
+
+ return;
+ }
+
+ /*
+ * We know the iSCSI offset is in the next page of the current
+ * scatterlist. Increase the lm->sg_page pointer and try again.
+ */
+ if (lm->current_offset >= lm->iovec_length) {
+ DEBUG_MAP_SEGMENTS("LINUX_MAP3[%p]: Next Page:"
+ " lm->current_offset: %u, iovec_length: %u"
+ " sg_offset: %u\n", se_mem, lm->current_offset,
+ lm->iovec_length, sg_offset);
+
+ lm->current_offset -= lm->iovec_length;
+ lm->sg_length -= lm->iovec_length;
+ lm->sg_page++;
+ sg_offset = 0;
+
+ DEBUG_MAP_SEGMENTS("LINUX_MAP3[%p]: ** Skipping to Next Page,"
+ " updated values: lm->current_offset: %u\n", se_mem,
+ lm->current_offset);
+
+ goto recalc;
+ }
+
+ /*
+ * The iSCSI offset is in the current page, increment the iovec
+ * base and reduce iovec length.
+ */
+ lm->iovec_base = page_address(lm->sg_page);
+
+ DEBUG_MAP_SEGMENTS("LINUX_MAP4[%p]: Set lm->iovec_base to %p\n", se_mem,
+ lm->iovec_base);
+
+ lm->iovec_base += sg_offset;
+ lm->iovec_base += lm->current_offset;
+ DEBUG_MAP_SEGMENTS("****** the OLD lm->iovec_length: %u lm->sg_length:"
+ " %u\n", lm->iovec_length, lm->sg_length);
+
+ if ((lm->iovec_length - lm->current_offset) < *data_length)
+ lm->iovec_length -= lm->current_offset;
+ else
+ lm->iovec_length = *data_length;
+
+ if ((lm->sg_length - lm->current_offset) < *data_length)
+ lm->sg_length -= lm->current_offset;
+ else
+ lm->sg_length = *data_length;
+
+ lm->current_offset = 0;
+
+ DEBUG_MAP_SEGMENTS("****** the NEW lm->iovec_length %u lm->sg_length:"
+ " %u\n", lm->iovec_length, lm->sg_length);
+}
+
+/* #define iscsi_linux_get_iscsi_offset_DEBUG */
+#ifdef iscsi_linux_get_iscsi_offset_DEBUG
+#define DEBUG_GET_ISCSI_OFFSET(buf...) PYXPRINT(buf)
+#else
+#define DEBUG_GET_ISCSI_OFFSET(buf...)
+#endif
+
+/* transport_get_iscsi_offset():
+ *
+ *
+ */
+static int transport_get_iscsi_offset(
+ se_offset_map_t *lmap,
+ se_unmap_sg_t *usg)
+{
+ u32 current_length = 0, current_iscsi_offset = lmap->iscsi_offset;
+ u32 total_offset = 0;
+ se_cmd_t *cmd = usg->se_cmd;
+ se_mem_t *se_mem;
+
+ list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_list, se_list)
+ break;
+
+ if (!se_mem) {
+ printk(KERN_ERR "Unable to locate se_mem from"
+ " T_TASK(cmd)->t_mem_list\n");
+ return -1;
+ }
+
+ /*
+ * Locate the current offset from the passed iSCSI Offset.
+ */
+ while (lmap->iscsi_offset != current_length) {
+ /*
+ * The iSCSI Offset is within the current se_mem_t.
+ *
+ * Or:
+ *
+ * The iSCSI Offset is outside of the current se_mem_t.
+ * Recalculate the values and obtain the next se_mem_t pointer.
+ */
+ total_offset += se_mem->se_len;
+
+ DEBUG_GET_ISCSI_OFFSET("ISCSI_OFFSET: current_length: %u,"
+ " total_offset: %u, sg->length: %u\n",
+ current_length, total_offset, se_mem->se_len);
+
+ if (total_offset > lmap->iscsi_offset) {
+ current_length += current_iscsi_offset;
+ lmap->orig_offset = lmap->current_offset =
+ usg->t_offset = current_iscsi_offset;
+ DEBUG_GET_ISCSI_OFFSET("ISCSI_OFFSET: Within Current"
+ " se_mem_t: %p, current_length incremented to"
+ " %u\n", se_mem, current_length);
+ } else {
+ current_length += se_mem->se_len;
+ current_iscsi_offset -= se_mem->se_len;
+
+ DEBUG_GET_ISCSI_OFFSET("ISCSI_OFFSET: Outside of"
+ " Current se_mem: %p, current_length"
+ " incremented to %u and current_iscsi_offset"
+ " decremented to %u\n", se_mem, current_length,
+ current_iscsi_offset);
+
+ list_for_each_entry_continue(se_mem,
+ T_TASK(cmd)->t_mem_list, se_list)
+ break;
+
+ if (!se_mem) {
+ printk(KERN_ERR "Unable to locate se_mem_t\n");
+ return -1;
+ }
+ }
+ }
+ lmap->map_orig_se_mem = se_mem;
+ usg->cur_se_mem = se_mem;
+
+ return 0;
+}
+
+/* #define iscsi_OS_set_SG_iovec_ptrs_DEBUG */
+#ifdef iscsi_OS_set_SG_iovec_ptrs_DEBUG
+#define DEBUG_IOVEC_SCATTERLISTS(buf...) PYXPRINT(buf)
+
+static void iscsi_check_iovec_map(
+ u32 iovec_count,
+ u32 map_length,
+ se_map_sg_t *map_sg,
+ se_unmap_sg_t *unmap_sg)
+{
+ u32 i, iovec_map_length = 0;
+ se_cmd_t *cmd = map_sg->se_cmd;
+ struct iovec *iov = map_sg->iov;
+ se_mem_t *se_mem;
+
+ for (i = 0; i < iovec_count; i++)
+ iovec_map_length += iov[i].iov_len;
+
+ if (iovec_map_length == map_length)
+ return;
+
+ printk(KERN_INFO "Calculated iovec_map_length: %u does not match passed"
+ " map_length: %u\n", iovec_map_length, map_length);
+ printk(KERN_INFO "ITT: 0x%08x data_length: %u data_direction %d\n",
+ CMD_TFO(cmd)->get_task_tag(cmd), cmd->data_length,
+ cmd->data_direction);
+
+ iovec_map_length = 0;
+
+ for (i = 0; i < iovec_count; i++) {
+ printk(KERN_INFO "iov[%d].iov_[base,len]: %p / %u bytes------"
+ "-->\n", i, iov[i].iov_base, iov[i].iov_len);
+
+ printk(KERN_INFO "iovec_map_length from %u to %u\n",
+ iovec_map_length, iovec_map_length + iov[i].iov_len);
+ iovec_map_length += iov[i].iov_len;
+
+ printk(KERN_INFO "XXXX_map_length from %u to %u\n", map_length,
+ (map_length - iov[i].iov_len));
+ map_length -= iov[i].iov_len;
+ }
+
+ list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_list, se_list) {
+ printk(KERN_INFO "se_mem[%p]: offset: %u length: %u\n",
+ se_mem, se_mem->se_off, se_mem->se_len);
+ }
+
+ BUG();
+}
+
+#else
+#define DEBUG_IOVEC_SCATTERLISTS(buf...)
+#define iscsi_check_iovec_map(a, b, c, d)
+#endif
+
+/* transport_generic_set_iovec_ptrs():
+ *
+ *
+ */
+static int transport_generic_set_iovec_ptrs(
+ se_map_sg_t *map_sg,
+ se_unmap_sg_t *unmap_sg)
+{
+ u32 i = 0 /* For iovecs */, j = 0 /* For scatterlists */;
+#ifdef iscsi_OS_set_SG_iovec_ptrs_DEBUG
+ u32 orig_map_length = map_sg->data_length;
+#endif
+ se_cmd_t *cmd = map_sg->se_cmd;
+ se_offset_map_t *lmap = &unmap_sg->lmap;
+ struct iovec *iov = map_sg->iov;
+
+ /*
+ * Used for non scatterlist operations, assume a single iovec.
+ */
+ if (!T_TASK(cmd)->t_task_se_num) {
+ DEBUG_IOVEC_SCATTERLISTS("ITT: 0x%08x No se_mem_t elements"
+ " present\n", CMD_TFO(cmd)->get_task_tag(cmd));
+ iov[0].iov_base = (unsigned char *) T_TASK(cmd)->t_task_buf +
+ map_sg->data_offset;
+ iov[0].iov_len = map_sg->data_length;
+ return 1;
+ }
+
+ /*
+ * Set lmap->map_reset = 1 so the first call to
+ * transport_calculate_map_segment() sets up the initial
+ * values for se_offset_map_t.
+ */
+ lmap->map_reset = 1;
+
+ DEBUG_IOVEC_SCATTERLISTS("[-------------------] ITT: 0x%08x OS"
+ " Independent Network POSIX defined iovectors to SE Memory"
+ " [-------------------]\n\n", CMD_TFO(cmd)->get_task_tag(cmd));
+
+ /*
+ * Get a pointer to the first used scatterlist based on the passed
+ * offset. Also set the rest of the needed values in iscsi_linux_map_t.
+ */
+ lmap->iscsi_offset = map_sg->data_offset;
+ if (map_sg->map_flags & MAP_SG_KMAP) {
+ unmap_sg->se_cmd = map_sg->se_cmd;
+ transport_get_iscsi_offset(lmap, unmap_sg);
+ unmap_sg->data_length = map_sg->data_length;
+ } else {
+ lmap->current_offset = lmap->orig_offset;
+ }
+ lmap->map_se_mem = lmap->map_orig_se_mem;
+
+ DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: Total map_sg->data_length: %d,"
+ " lmap->iscsi_offset: %d, cmd->orig_iov_data_count: %d\n",
+ map_sg->data_length, lmap->iscsi_offset,
+ cmd->orig_iov_data_count);
+
+ while (map_sg->data_length) {
+ /*
+ * Time to get the virtual address for use with iovec pointers.
+ * This function will return the expected iovec_base address
+ * and iovec_length.
+ */
+ transport_calculate_map_segment(&map_sg->data_length, lmap);
+
+ /*
+ * Set the iov.iov_base and iov.iov_len from the current values
+ * in iscsi_linux_map_t.
+ */
+ iov[i].iov_base = lmap->iovec_base;
+ iov[i].iov_len = lmap->iovec_length;
+
+ /*
+ * Subtract the final iovec length from the total length to be
+ * mapped, and the length of the current scatterlist. Also
+ * perform the paranoid check to make sure we are not going to
+ * overflow the iovecs allocated for this command in the next
+ * pass.
+ */
+ map_sg->data_length -= iov[i].iov_len;
+ lmap->sg_length -= iov[i].iov_len;
+
+ DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: iov[%u].iov_len: %u\n",
+ i, iov[i].iov_len);
+ DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: lmap->sg_length: from %u"
+ " to %u\n", lmap->sg_length + iov[i].iov_len,
+ lmap->sg_length);
+ DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: Changed total"
+ " map_sg->data_length from %u to %u\n",
+ map_sg->data_length + iov[i].iov_len,
+ map_sg->data_length);
+
+ if ((++i + 1) > cmd->orig_iov_data_count) {
+ printk(KERN_ERR "Current iovec count %u is greater than"
+ " se_cmd_t->orig_data_iov_count %u, cannot"
+ " continue.\n", i+1, cmd->orig_iov_data_count);
+ return -1;
+ }
+
+ /*
+ * All done mapping this scatterlist's pages, move on to
+ * the next scatterlist by setting lmap.map_reset = 1;
+ */
+ if (!lmap->sg_length || !map_sg->data_length) {
+ list_for_each_entry(lmap->map_se_mem,
+ &lmap->map_se_mem->se_list, se_list)
+ break;
+
+ if (!lmap->map_se_mem) {
+ printk(KERN_ERR "Unable to locate next"
+ " lmap->map_se_mem_t entry\n");
+ return -1;
+ }
+ j++;
+
+ lmap->sg_page = NULL;
+ lmap->map_reset = 1;
+
+ DEBUG_IOVEC_SCATTERLISTS("OS_IOVEC: Done with current"
+ " scatterlist, incremented Generic scatterlist"
+ " Counter to %d and reset = 1\n", j);
+ } else
+ lmap->sg_page++;
+ }
+
+ unmap_sg->sg_count = j;
+
+ iscsi_check_iovec_map(i, orig_map_length, map_sg, unmap_sg);
+
+ return i;
+}
+
+/* transport_generic_allocate_buf():
+ *
+ * Called from transport_generic_new_cmd() in Transport Processing Thread.
+ */
+int transport_generic_allocate_buf(
+ se_cmd_t *cmd,
+ u32 data_length,
+ u32 dma_size)
+{
+ unsigned char *buf;
+
+ buf = kzalloc(data_length, GFP_KERNEL);
+ if (!(buf)) {
+ printk(KERN_ERR "Unable to allocate memory for buffer\n");
+ return -1;
+ }
+
+ T_TASK(cmd)->t_task_se_num = 0;
+ T_TASK(cmd)->t_task_buf = buf;
+
+ return 0;
+}
+
+/* transport_generic_allocate_none():
+ *
+ *
+ */
+static int transport_generic_allocate_none(
+ se_cmd_t *cmd,
+ u32 data_length,
+ u32 dma_size)
+{
+ return 0;
+}
+
+/* transport_generic_map_SG_segments():
+ *
+ *
+ */
+static void transport_generic_map_SG_segments(se_unmap_sg_t *unmap_sg)
+{
+ u32 i = 0;
+ se_cmd_t *cmd = unmap_sg->se_cmd;
+ se_mem_t *se_mem = unmap_sg->cur_se_mem;
+
+ if (!(T_TASK(cmd)->t_task_se_num))
+ return;
+
+ list_for_each_entry_continue(se_mem, T_TASK(cmd)->t_mem_list, se_list) {
+ kmap(se_mem->se_page);
+
+ if (++i == unmap_sg->sg_count)
+ break;
+ }
+}
+
+/* transport_generic_unmap_SG_segments():
+ *
+ *
+ */
+static void transport_generic_unmap_SG_segments(se_unmap_sg_t *unmap_sg)
+{
+ u32 i = 0;
+ se_cmd_t *cmd = unmap_sg->se_cmd;
+ se_mem_t *se_mem = unmap_sg->cur_se_mem;
+
+ if (!(T_TASK(cmd)->t_task_se_num))
+ return;
+
+ list_for_each_entry_continue(se_mem, T_TASK(cmd)->t_mem_list, se_list) {
+ kunmap(se_mem->se_page);
+
+ if (++i == unmap_sg->sg_count)
+ break;
+ }
+
+ return;
+}
+
+static inline u32 transport_lba_21(unsigned char *cdb)
+{
+ return ((cdb[1] & 0x1f) << 16) | (cdb[2] << 8) | cdb[3];
+}
+
+static inline u32 transport_lba_32(unsigned char *cdb)
+{
+ return (cdb[2] << 24) | (cdb[3] << 16) | (cdb[4] << 8) | cdb[5];
+}
+
+static inline unsigned long long transport_lba_64(unsigned char *cdb)
+{
+ unsigned int __v1, __v2;
+
+ __v1 = (cdb[2] << 24) | (cdb[3] << 16) | (cdb[4] << 8) | cdb[5];
+ __v2 = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
+
+ return ((unsigned long long)__v2) | (unsigned long long)__v1 << 32;
+}
+
+/* transport_set_supported_SAM_opcode():
+ *
+ *
+ */
+void transport_set_supported_SAM_opcode(se_cmd_t *se_cmd)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(se_cmd)->t_state_lock, flags);
+ se_cmd->se_cmd_flags |= SCF_SUPPORTED_SAM_OPCODE;
+ spin_unlock_irqrestore(&T_TASK(se_cmd)->t_state_lock, flags);
+}
+
+/*
+ * Called from interrupt context.
+ */
+void transport_task_timeout_handler(unsigned long data)
+{
+ se_task_t *task = (se_task_t *)data;
+ se_cmd_t *cmd = TASK_CMD(task);
+ unsigned long flags;
+
+ DEBUG_TT("transport task timeout fired! task: %p cmd: %p\n", task, cmd);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (task->task_flags & TF_STOP) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+ task->task_flags &= ~TF_RUNNING;
+
+ /*
+ * Determine if transport_complete_task() has already been called.
+ */
+ if (!(atomic_read(&task->task_active))) {
+ DEBUG_TT("transport task: %p cmd: %p timeout task_active"
+ " == 0\n", task, cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+
+ atomic_inc(&T_TASK(cmd)->t_se_count);
+ atomic_inc(&T_TASK(cmd)->t_transport_timeout);
+ T_TASK(cmd)->t_tasks_failed = 1;
+
+ atomic_set(&task->task_timeout, 1);
+ task->task_error_status = PYX_TRANSPORT_TASK_TIMEOUT;
+ task->task_scsi_status = 1;
+
+ if (atomic_read(&task->task_stop)) {
+ DEBUG_TT("transport task: %p cmd: %p timeout task_stop"
+ " == 1\n", task, cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ up(&task->task_stop_sem);
+ return;
+ }
+
+ if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_left))) {
+ DEBUG_TT("transport task: %p cmd: %p timeout non zero"
+ " t_task_cdbs_left\n", task, cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+ DEBUG_TT("transport task: %p cmd: %p timeout ZERO t_task_cdbs_left\n",
+ task, cmd);
+
+ cmd->t_state = TRANSPORT_COMPLETE_FAILURE;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ cmd->transport_add_cmd_to_queue(cmd, TRANSPORT_COMPLETE_FAILURE);
+}
+
+u32 transport_get_default_task_timeout(se_device_t *dev)
+{
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_DISK)
+ return TRANSPORT_TIMEOUT_TYPE_DISK;
+
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_ROM)
+ return TRANSPORT_TIMEOUT_TYPE_ROM;
+
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+ return TRANSPORT_TIMEOUT_TYPE_TAPE;
+
+ return TRANSPORT_TIMEOUT_TYPE_OTHER;
+}
+EXPORT_SYMBOL(transport_get_default_task_timeout);
+
+/*
+ * Called with T_TASK(cmd)->t_state_lock held.
+ */
+void transport_start_task_timer(se_task_t *task)
+{
+ int timeout;
+
+ if (task->task_flags & TF_RUNNING)
+ return;
+ /*
+ * If the task_timeout is disabled, exit now.
+ */
+ timeout = task->se_obj_api->get_task_timeout(task->se_obj_ptr);
+ if (!(timeout))
+ return;
+
+ init_timer(&task->task_timer);
+ task->task_timer.expires = (get_jiffies_64() + timeout * HZ);
+ task->task_timer.data = (unsigned long) task;
+ task->task_timer.function = transport_task_timeout_handler;
+
+ task->task_flags |= TF_RUNNING;
+ add_timer(&task->task_timer);
+#if 0
+ printk(KERN_INFO "Starting task timer for cmd: %p task: %p seconds:"
+ " %d\n", task->task_se_cmd, task, timeout);
+#endif
+}
+
+/*
+ * Called with spin_lock_irq(&T_TASK(cmd)->t_state_lock) held.
+ */
+void __transport_stop_task_timer(se_task_t *task, unsigned long *flags)
+{
+ se_cmd_t *cmd = TASK_CMD(task);
+
+ if (!(task->task_flags & TF_RUNNING))
+ return;
+
+ task->task_flags |= TF_STOP;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, *flags);
+
+ del_timer_sync(&task->task_timer);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, *flags);
+ task->task_flags &= ~TF_RUNNING;
+ task->task_flags &= ~TF_STOP;
+}
+
+void transport_stop_task_timer(se_task_t *task)
+{
+ se_cmd_t *cmd = TASK_CMD(task);
+ unsigned long flags;
+#if 0
+ printk(KERN_INFO "Stopping task timer for cmd: %p task: %p\n",
+ cmd, task);
+#endif
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (!(task->task_flags & TF_RUNNING)) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return;
+ }
+ task->task_flags |= TF_STOP;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ del_timer_sync(&task->task_timer);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ task->task_flags &= ~TF_RUNNING;
+ task->task_flags &= ~TF_STOP;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+void transport_stop_all_task_timers(se_cmd_t *cmd)
+{
+ se_task_t *task = NULL, *task_tmp;
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ list_for_each_entry_safe(task, task_tmp,
+ &T_TASK(cmd)->t_task_list, t_list)
+ __transport_stop_task_timer(task, &flags);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+static inline int transport_tcq_window_closed(se_device_t *dev)
+{
+ if (dev->dev_tcq_window_closed++ <
+ PYX_TRANSPORT_WINDOW_CLOSED_THRESHOLD) {
+ msleep(PYX_TRANSPORT_WINDOW_CLOSED_WAIT_SHORT);
+ } else
+ msleep(PYX_TRANSPORT_WINDOW_CLOSED_WAIT_LONG);
+
+ wake_up_interruptible(&dev->dev_queue_obj->thread_wq);
+ return 0;
+}
+
+int transport_execute_tasks(se_cmd_t *cmd)
+{
+ if (!(cmd->se_cmd_flags & SCF_SE_DISABLE_ONLINE_CHECK)) {
+ if (CMD_ORIG_OBJ_API(cmd)->check_online(
+ cmd->se_orig_obj_ptr) != 0) {
+ cmd->transport_error_status =
+ PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ transport_generic_request_failure(cmd, NULL, 0, 1);
+ return 0;
+ }
+ }
+ /*
+ * Add the task(s) built for the passed se_cmd_t to the
+ * execution queue for this se_device_t.
+ */
+ if (!transport_cmd_check_stop(cmd, 0, TRANSPORT_PROCESSING))
+ CMD_ORIG_OBJ_API(cmd)->add_tasks(cmd->se_orig_obj_ptr, cmd);
+
+ CMD_ORIG_OBJ_API(cmd)->execute_tasks(cmd->se_orig_obj_ptr);
+
+ return 0;
+}
+
+int __transport_execute_tasks(se_device_t *dev)
+{
+ int error;
+ se_cmd_t *cmd = NULL;
+ se_task_t *task;
+ unsigned long flags;
+
+ /*
+ * Check if there is enough room in the device and HBA queue to send
+ * se_transport_task_t's to the selected transport.
+ */
+check_depth:
+ spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
+ if (!(atomic_read(&dev->depth_left)) ||
+ !(atomic_read(&SE_HBA(dev)->left_queue_depth))) {
+ spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+ return transport_tcq_window_closed(dev);
+ }
+ dev->dev_tcq_window_closed = 0;
+
+ spin_lock(&dev->execute_task_lock);
+ task = transport_get_task_from_execute_queue(dev);
+ spin_unlock(&dev->execute_task_lock);
+
+ if (!task) {
+ spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+ return 0;
+ }
+
+ atomic_dec(&dev->depth_left);
+ atomic_dec(&SE_HBA(dev)->left_queue_depth);
+ spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+
+ cmd = TASK_CMD(task);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ atomic_set(&task->task_active, 1);
+ atomic_set(&task->task_sent, 1);
+ atomic_inc(&T_TASK(cmd)->t_task_cdbs_sent);
+
+ if (atomic_read(&T_TASK(cmd)->t_task_cdbs_sent) ==
+ T_TASK(cmd)->t_task_cdbs)
+ atomic_set(&cmd->transport_sent, 1);
+
+ transport_start_task_timer(task);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ /*
+ * The se_cmd_t->transport_emulate_cdb() function pointer is used
+ * to grab REPORT_LUNS CDBs before they hit the
+ * se_subsystem_api_t->do_task() caller below.
+ */
+ if (cmd->transport_emulate_cdb) {
+ error = cmd->transport_emulate_cdb(cmd);
+ if (error != 0) {
+ cmd->transport_error_status = error;
+ atomic_set(&task->task_active, 0);
+ atomic_set(&cmd->transport_sent, 0);
+ transport_stop_tasks_for_cmd(cmd);
+ transport_generic_request_failure(cmd, dev, 0, 1);
+ goto check_depth;
+ }
+ /*
+ * Handle the successful completion for transport_emulate_cdb()
+ * usage.
+ */
+ cmd->scsi_status = SAM_STAT_GOOD;
+ task->task_scsi_status = GOOD;
+ transport_complete_task(task, 1);
+ } else {
+ error = TRANSPORT(dev)->do_task(task);
+ if (error != 0) {
+ cmd->transport_error_status = error;
+ atomic_set(&task->task_active, 0);
+ atomic_set(&cmd->transport_sent, 0);
+ transport_stop_tasks_for_cmd(cmd);
+ transport_generic_request_failure(cmd, dev, 0, 1);
+ }
+ }
+
+ goto check_depth;
+
+ return 0;
+}
+
+/* transport_new_cmd_failure():
+ *
+ *
+ */
+void transport_new_cmd_failure(se_cmd_t *se_cmd)
+{
+ unsigned long flags;
+ /*
+ * Any unsolicited data will get dumped for failed command inside of
+ * the fabric plugin
+ */
+ spin_lock_irqsave(&T_TASK(se_cmd)->t_state_lock, flags);
+ se_cmd->se_cmd_flags |= SCF_SE_CMD_FAILED;
+ se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ spin_unlock_irqrestore(&T_TASK(se_cmd)->t_state_lock, flags);
+
+ CMD_TFO(se_cmd)->new_cmd_failure(se_cmd);
+}
+
+static int transport_generic_map_buffers_to_tasks(se_cmd_t *);
+static void transport_nop_wait_for_tasks(se_cmd_t *, int, int);
+
+static inline u32 transport_get_sectors_6(
+ unsigned char *cdb,
+ se_cmd_t *cmd,
+ int *ret)
+{
+ se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+ /*
+ * Assume TYPE_DISK for non se_device_t objects.
+ * Use 8-bit sector value.
+ */
+ if (!dev)
+ goto type_disk;
+
+ /*
+ * Use 24-bit allocation length for TYPE_TAPE.
+ */
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+ return (u32)(cdb[2] << 16) + (cdb[3] << 8) + cdb[4];
+
+ /*
+ * Everything else assume TYPE_DISK Sector CDB location.
+ * Use 8-bit sector value.
+ */
+type_disk:
+ return (u32)cdb[4];
+}
+
+static inline u32 transport_get_sectors_10(
+ unsigned char *cdb,
+ se_cmd_t *cmd,
+ int *ret)
+{
+ se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+ /*
+ * Assume TYPE_DISK for non se_device_t objects.
+ * Use 16-bit sector value.
+ */
+ if (!dev)
+ goto type_disk;
+
+ /*
+ * XXX_10 is not defined in SSC, throw an exception
+ */
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
+ *ret = -1;
+ return 0;
+ }
+
+ /*
+ * Everything else assume TYPE_DISK Sector CDB location.
+ * Use 16-bit sector value.
+ */
+type_disk:
+ return (u32)(cdb[7] << 8) + cdb[8];
+}
+
+static inline u32 transport_get_sectors_12(
+ unsigned char *cdb,
+ se_cmd_t *cmd,
+ int *ret)
+{
+ se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+ /*
+ * Assume TYPE_DISK for non se_device_t objects.
+ * Use 32-bit sector value.
+ */
+ if (!dev)
+ goto type_disk;
+
+ /*
+ * XXX_12 is not defined in SSC, throw an exception
+ */
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
+ *ret = -1;
+ return 0;
+ }
+
+ /*
+ * Everything else assume TYPE_DISK Sector CDB location.
+ * Use 32-bit sector value.
+ */
+type_disk:
+ return (u32)(cdb[6] << 24) + (cdb[7] << 16) + (cdb[8] << 8) + cdb[9];
+}
+
+static inline u32 transport_get_sectors_16(
+ unsigned char *cdb,
+ se_cmd_t *cmd,
+ int *ret)
+{
+ se_device_t *dev = SE_LUN(cmd)->se_dev;
+
+ /*
+ * Assume TYPE_DISK for non se_device_t objects.
+ * Use 32-bit sector value.
+ */
+ if (!dev)
+ goto type_disk;
+
+ /*
+ * Use 24-bit allocation length for TYPE_TAPE.
+ */
+ if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+ return (u32)(cdb[12] << 16) + (cdb[13] << 8) + cdb[14];
+
+type_disk:
+ return (u32)(cdb[10] << 24) + (cdb[11] << 16) +
+ (cdb[12] << 8) + cdb[13];
+}
+
+static inline u32 transport_get_size(
+ u32 sectors,
+ unsigned char *cdb,
+ se_cmd_t *cmd)
+{
+ return CMD_ORIG_OBJ_API(cmd)->get_cdb_size(cmd->se_orig_obj_ptr,
+ sectors, cdb);
+}
+
+static inline void transport_get_maps(se_cmd_t *cmd)
+{
+ cmd->transport_map_SG_segments = &transport_generic_map_SG_segments;
+ cmd->transport_unmap_SG_segments = &transport_generic_unmap_SG_segments;
+}
+
+static unsigned char asciihex_to_binaryhex(unsigned char val[2])
+{
+ unsigned char result = 0;
+ /*
+ * MSB
+ */
+ if ((val[0] >= 'a') && (val[0] <= 'f'))
+ result = ((val[0] - 'a' + 10) & 0xf) << 4;
+ else
+ if ((val[0] >= 'A') && (val[0] <= 'F'))
+ result = ((val[0] - 'A' + 10) & 0xf) << 4;
+ else /* digit */
+ result = ((val[0] - '0') & 0xf) << 4;
+ /*
+ * LSB
+ */
+ if ((val[1] >= 'a') && (val[1] <= 'f'))
+ result |= ((val[1] - 'a' + 10) & 0xf);
+ else
+ if ((val[1] >= 'A') && (val[1] <= 'F'))
+ result |= ((val[1] - 'A' + 10) & 0xf);
+ else /* digit */
+ result |= ((val[1] - '0') & 0xf);
+
+ return result;
+}
+
+extern int transport_generic_emulate_inquiry(
+ se_cmd_t *cmd,
+ unsigned char type,
+ unsigned char *prod,
+ unsigned char *version,
+ unsigned char *se_location)
+{
+ se_device_t *dev = SE_DEV(cmd);
+ se_lun_t *lun = SE_LUN(cmd);
+ se_port_t *port = NULL;
+ se_portal_group_t *tpg = NULL;
+ t10_alua_lu_gp_member_t *lu_gp_mem;
+ t10_alua_tg_pt_gp_member_t *tg_pt_gp_mem;
+ unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+ unsigned char *cdb = T_TASK(cmd)->t_task_cdb;
+ unsigned char *iqn_sn, binary, binary_new;
+ u32 prod_len, iqn_sn_len, se_location_len;
+ u32 unit_serial_len, off = 0;
+ int i;
+ u16 len = 0, id_len;
+
+ /*
+ * Make sure we at least have 8 bytes of INQUIRY response payload
+ * going back.
+ */
+ if (cmd->data_length < 8) {
+ printk(KERN_ERR "SCSI Inquiry payload length: %u too small\n",
+ cmd->data_length);
+ return -1;
+ }
+ buf[0] = type;
+
+ if (!(cdb[1] & 0x1)) {
+ if (type == TYPE_TAPE)
+ buf[1] = 0x80;
+ buf[2] = TRANSPORT(dev)->get_device_rev(dev);
+ /*
+ * Enable SCCS and TPGS fields for Emulated ALUA
+ */
+ if (T10_ALUA(dev->se_sub_dev)->alua_type ==
+ SPC3_ALUA_EMULATED) {
+ /*
+ * Set SCCS for MAINTENANCE_IN +
+ * REPORT_TARGET_PORT_GROUPS
+ */
+ buf[5] = 0x80;
+ /*
+ * Set TPGS field, see spc4r17 section 6.4.2 Table 135
+ */
+ buf[5] |= TPGS_IMPLICT_ALUA;
+ }
+ buf[7] = 0x32; /* Sync=1 and CmdQue=1 */
+ /*
+ * Do not include vendor, product, reversion info in INQUIRY
+ * response payload for cdbs with a small allocation length.
+ */
+ if (cmd->data_length < 36) {
+ buf[4] = 3; /* Set additional length to 3 */
+ return 0;
+ }
+
+ snprintf((unsigned char *)&buf[8], 8, "LIO-ORG");
+ snprintf((unsigned char *)&buf[16], 16, "%s", prod);
+ snprintf((unsigned char *)&buf[32], 4, "%s", version);
+ buf[4] = 31; /* Set additional length to 31 */
+ return 0;
+ }
+
+ switch (cdb[2]) {
+ case 0x00: /* supported vital product data pages */
+ buf[1] = 0x00;
+ buf[3] = 3;
+ buf[4] = 0x0;
+ buf[5] = 0x80;
+ buf[6] = 0x83;
+ len = 3;
+ break;
+ case 0x80: /* unit serial number */
+ buf[1] = 0x80;
+ if (dev->se_sub_dev->su_dev_flags &
+ SDF_EMULATED_VPD_UNIT_SERIAL) {
+ unit_serial_len =
+ strlen(&DEV_T10_WWN(dev)->unit_serial[0]);
+ unit_serial_len++; /* For NULL Terminator */
+
+ if (((len + 4) + unit_serial_len) > cmd->data_length) {
+ len += unit_serial_len;
+ goto set_len;
+ }
+ len += sprintf((unsigned char *)&buf[4], "%s",
+ &DEV_T10_WWN(dev)->unit_serial[0]);
+ } else {
+ iqn_sn = transport_get_iqn_sn();
+ iqn_sn_len = strlen(iqn_sn);
+ iqn_sn_len++; /* For ":" */
+ se_location_len = strlen(se_location);
+ se_location_len++; /* For NULL Terminator */
+
+ if (((len + 4) + (iqn_sn_len + se_location_len)) >
+ cmd->data_length) {
+ len += (iqn_sn_len + se_location_len);
+ goto set_len;
+ }
+ len += sprintf((unsigned char *)&buf[4], "%s:%s",
+ iqn_sn, se_location);
+ }
+ len++; /* Extra Byte for NULL Terminator */
+ buf[3] = len;
+ break;
+ case 0x83:
+ /*
+ * Device identification VPD, for a complete list of
+ * DESIGNATOR TYPEs see spc4r17 Table 459.
+ */
+ buf[1] = 0x83;
+ off = 4;
+ /*
+ * NAA IEEE Registered Extended Assigned designator format,
+ * see spc4r17 section 7.7.3.6.5
+ *
+ * We depend upon a target_core_mod/ConfigFS provided
+ * /sys/kernel/config/target/core/$HBA/$DEV/wwn/vpd_unit_serial
+ * value in order to return the NAA id.
+ */
+ if (!(dev->se_sub_dev->su_dev_flags &
+ SDF_EMULATED_VPD_UNIT_SERIAL))
+ goto check_t10_vend_desc;
+ if ((off + 20) > cmd->data_length)
+ goto check_t10_vend_desc;
+ /* CODE SET == Binary */
+ buf[off++] = 0x1;
+ /* Set ASSOICATION == addressed logical unit: 0)b */
+ buf[off] = 0x00;
+ /* Identifier/Designator type == NAA identifier */
+ buf[off++] = 0x3;
+ off++;
+ /* Identifier/Designator length */
+ buf[off++] = 0x10;
+ /*
+ * Start NAA IEEE Registered Extended Identifier/Designator
+ */
+ buf[off++] = (0x6 << 4);
+ /*
+ * Use OpenFabrics IEEE Company ID: 00 14 05
+ */
+ buf[off++] = 0x01;
+ buf[off++] = 0x40;
+ buf[off] = (0x5 << 4);
+ /*
+ * Return ConfigFS Unit Serial Number information for
+ * VENDOR_SPECIFIC_IDENTIFIER and
+ * VENDOR_SPECIFIC_IDENTIFIER_EXTENTION
+ */
+ binary = asciihex_to_binaryhex(
+ &DEV_T10_WWN(dev)->unit_serial[0]);
+ buf[off++] |= (binary & 0xf0) >> 4;
+ for (i = 0; i < 24; i += 2) {
+ binary_new = asciihex_to_binaryhex(
+ &DEV_T10_WWN(dev)->unit_serial[i+2]);
+ buf[off] = (binary & 0x0f) << 4;
+ buf[off++] |= (binary_new & 0xf0) >> 4;
+ binary = binary_new;
+ }
+ len = 20;
+ off = (len + 4);
+check_t10_vend_desc:
+ /*
+ * T10 Vendor Identifier Page, see spc4r17 section 7.7.3.4
+ */
+ id_len = 8; /* For Vendor field */
+ prod_len = 4; /* For VPD Header */
+ prod_len += 8; /* For Vendor field */
+ prod_len += strlen(prod);
+ prod_len++; /* For : */
+
+ if (dev->se_sub_dev->su_dev_flags &
+ SDF_EMULATED_VPD_UNIT_SERIAL) {
+ unit_serial_len =
+ strlen(&DEV_T10_WWN(dev)->unit_serial[0]);
+ unit_serial_len++; /* For NULL Terminator */
+
+ if ((len + (id_len + 4) +
+ (prod_len + unit_serial_len)) >
+ cmd->data_length) {
+ len += (prod_len + unit_serial_len);
+ goto check_port;
+ }
+ id_len += sprintf((unsigned char *)&buf[off+12],
+ "%s:%s", prod,
+ &DEV_T10_WWN(dev)->unit_serial[0]);
+ } else {
+ iqn_sn = transport_get_iqn_sn();
+ iqn_sn_len = strlen(iqn_sn);
+ iqn_sn_len++; /* For ":" */
+ se_location_len = strlen(se_location);
+ se_location_len++; /* For NULL Terminator */
+
+ if ((len + (id_len + 4) + (prod_len + iqn_sn_len +
+ se_location_len)) > cmd->data_length) {
+ len += (prod_len + iqn_sn_len +
+ se_location_len);
+ goto check_port;
+ }
+ id_len += sprintf((unsigned char *)&buf[off+12],
+ "%s:%s:%s", prod, iqn_sn, se_location);
+ }
+ buf[off] = 0x2; /* ASCII */
+ buf[off+1] = 0x1; /* T10 Vendor ID */
+ buf[off+2] = 0x0;
+ memcpy((unsigned char *)&buf[off+4], "LIO-ORG", 8);
+ /* Extra Byte for NULL Terminator */
+ id_len++;
+ /* Identifier Length */
+ buf[off+3] = id_len;
+ /* Header size for Designation descriptor */
+ len += (id_len + 4);
+ off += (id_len + 4);
+ /*
+ * se_port_t is only set for INQUIRY VPD=1 through $FABRIC_MOD
+ */
+check_port:
+ port = lun->lun_sep;
+ if (port) {
+ t10_alua_lu_gp_t *lu_gp;
+ t10_alua_tg_pt_gp_t *tg_pt_gp;
+ u32 padding, scsi_name_len;
+ u16 lu_gp_id = 0;
+ u16 tg_pt_gp_id = 0;
+ u16 tpgt;
+
+ tpg = port->sep_tpg;
+ /*
+ * Relative target port identifer, see spc4r17
+ * section 7.7.3.7
+ *
+ * Get the PROTOCOL IDENTIFIER as defined by spc4r17
+ * section 7.5.1 Table 362
+ */
+ if (((len + 4) + 8) > cmd->data_length) {
+ len += 8;
+ goto check_tpgi;
+ }
+ buf[off] =
+ (TPG_TFO(tpg)->get_fabric_proto_ident() << 4);
+ buf[off++] |= 0x1; /* CODE SET == Binary */
+ buf[off] = 0x80; /* Set PIV=1 */
+ /* Set ASSOICATION == target port: 01b */
+ buf[off] |= 0x10;
+ /* DESIGNATOR TYPE == Relative target port identifer */
+ buf[off++] |= 0x4;
+ off++; /* Skip over Reserved */
+ buf[off++] = 4; /* DESIGNATOR LENGTH */
+ /* Skip over Obsolete field in RTPI payload
+ * in Table 472 */
+ off += 2;
+ buf[off++] = ((port->sep_rtpi >> 8) & 0xff);
+ buf[off++] = (port->sep_rtpi & 0xff);
+ len += 8; /* Header size + Designation descriptor */
+ /*
+ * Target port group identifier, see spc4r17
+ * section 7.7.3.8
+ *
+ * Get the PROTOCOL IDENTIFIER as defined by spc4r17
+ * section 7.5.1 Table 362
+ */
+check_tpgi:
+ if (T10_ALUA(dev->se_sub_dev)->alua_type !=
+ SPC3_ALUA_EMULATED)
+ goto check_scsi_name;
+
+ if (((len + 4) + 8) > cmd->data_length) {
+ len += 8;
+ goto check_lu_gp;
+ }
+ tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
+ if (!(tg_pt_gp_mem))
+ goto check_lu_gp;
+
+ spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+ tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
+ if (!(tg_pt_gp)) {
+ spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+ goto check_lu_gp;
+ }
+ tg_pt_gp_id = tg_pt_gp->tg_pt_gp_id;
+ spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
+
+ buf[off] =
+ (TPG_TFO(tpg)->get_fabric_proto_ident() << 4);
+ buf[off++] |= 0x1; /* CODE SET == Binary */
+ buf[off] = 0x80; /* Set PIV=1 */
+ /* Set ASSOICATION == target port: 01b */
+ buf[off] |= 0x10;
+ /* DESIGNATOR TYPE == Target port group identifier */
+ buf[off++] |= 0x5;
+ off++; /* Skip over Reserved */
+ buf[off++] = 4; /* DESIGNATOR LENGTH */
+ off += 2; /* Skip over Reserved Field */
+ buf[off++] = ((tg_pt_gp_id >> 8) & 0xff);
+ buf[off++] = (tg_pt_gp_id & 0xff);
+ len += 8; /* Header size + Designation descriptor */
+ /*
+ * Logical Unit Group identifier, see spc4r17
+ * section 7.7.3.8
+ */
+check_lu_gp:
+ if (((len + 4) + 8) > cmd->data_length) {
+ len += 8;
+ goto check_scsi_name;
+ }
+ lu_gp_mem = dev->dev_alua_lu_gp_mem;
+ if (!(lu_gp_mem))
+ goto check_scsi_name;
+
+ spin_lock(&lu_gp_mem->lu_gp_mem_lock);
+ lu_gp = lu_gp_mem->lu_gp;
+ if (!(lu_gp)) {
+ spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
+ goto check_scsi_name;
+ }
+ lu_gp_id = lu_gp->lu_gp_id;
+ spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
+
+ buf[off++] |= 0x1; /* CODE SET == Binary */
+ /* DESIGNATOR TYPE == Logical Unit Group identifier */
+ buf[off++] |= 0x6;
+ off++; /* Skip over Reserved */
+ buf[off++] = 4; /* DESIGNATOR LENGTH */
+ off += 2; /* Skip over Reserved Field */
+ buf[off++] = ((lu_gp_id >> 8) & 0xff);
+ buf[off++] = (lu_gp_id & 0xff);
+ len += 8; /* Header size + Designation descriptor */
+ /*
+ * SCSI name string designator, see spc4r17
+ * section 7.7.3.11
+ *
+ * Get the PROTOCOL IDENTIFIER as defined by spc4r17
+ * section 7.5.1 Table 362
+ */
+check_scsi_name:
+ scsi_name_len = strlen(TPG_TFO(tpg)->tpg_get_wwn(tpg));
+ /* UTF-8 ",t,0x<16-bit TPGT>" + NULL Terminator */
+ scsi_name_len += 10;
+ /* Header size + Designation descriptor */
+ scsi_name_len += 4;
+ padding = ((-scsi_name_len) & 3);
+ if (padding != 0)
+ scsi_name_len += padding;
+
+ if ((len + scsi_name_len) > cmd->data_length) {
+ len += scsi_name_len;
+ goto set_len;
+ }
+ buf[off] =
+ (TPG_TFO(tpg)->get_fabric_proto_ident() << 4);
+ buf[off++] |= 0x3; /* CODE SET == UTF-8 */
+ buf[off] = 0x80; /* Set PIV=1 */
+ /* Set ASSOICATION == target port: 01b */
+ buf[off] |= 0x10;
+ /* DESIGNATOR TYPE == SCSI name string */
+ buf[off++] |= 0x8;
+ off += 2; /* Skip over Reserved and length */
+ /*
+ * SCSI name string identifer containing, $FABRIC_MOD
+ * dependent information. For LIO-Target and iSCSI
+ * Target Port, this means "<iSCSI name>,t,0x<TPGT> in
+ * UTF-8 encoding.
+ */
+ tpgt = TPG_TFO(tpg)->tpg_get_tag(tpg);
+ scsi_name_len = sprintf(&buf[off], "%s,t,0x%04x",
+ TPG_TFO(tpg)->tpg_get_wwn(tpg), tpgt);
+ scsi_name_len += 1 /* Include NULL terminator */;
+ /*
+ * The null-terminated, null-padded (see 4.4.2) SCSI
+ * NAME STRING field contains a UTF-8 format string.
+ * The number of bytes in the SCSI NAME STRING field
+ * (i.e., the value in the DESIGNATOR LENGTH field)
+ * shall be no larger than 256 and shall be a multiple
+ * of four.
+ */
+ if (padding)
+ scsi_name_len += padding;
+
+ buf[off-1] = scsi_name_len;
+ off += scsi_name_len;
+ /* Header size + Designation descriptor */
+ len += (scsi_name_len + 4);
+ }
+set_len:
+ buf[2] = ((len >> 8) & 0xff);
+ buf[3] = (len & 0xff); /* Page Length for VPD 0x83 */
+ break;
+ default:
+ printk(KERN_ERR "Unknown VPD Code: 0x%02x\n", cdb[2]);
+ return -1;
+ }
+
+ return 0;
+}
+
+int transport_generic_emulate_readcapacity(
+ se_cmd_t *cmd,
+ u32 blocks)
+{
+ se_device_t *dev = SE_DEV(cmd);
+ unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+
+ buf[0] = (blocks >> 24) & 0xff;
+ buf[1] = (blocks >> 16) & 0xff;
+ buf[2] = (blocks >> 8) & 0xff;
+ buf[3] = blocks & 0xff;
+ buf[4] = (DEV_ATTRIB(dev)->block_size >> 24) & 0xff;
+ buf[5] = (DEV_ATTRIB(dev)->block_size >> 16) & 0xff;
+ buf[6] = (DEV_ATTRIB(dev)->block_size >> 8) & 0xff;
+ buf[7] = DEV_ATTRIB(dev)->block_size & 0xff;
+
+ return 0;
+}
+
+int transport_generic_emulate_readcapacity_16(
+ se_cmd_t *cmd,
+ unsigned long long blocks)
+{
+ se_device_t *dev = SE_DEV(cmd);
+ unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+
+ buf[0] = (blocks >> 56) & 0xff;
+ buf[1] = (blocks >> 48) & 0xff;
+ buf[2] = (blocks >> 40) & 0xff;
+ buf[3] = (blocks >> 32) & 0xff;
+ buf[4] = (blocks >> 24) & 0xff;
+ buf[5] = (blocks >> 16) & 0xff;
+ buf[6] = (blocks >> 8) & 0xff;
+ buf[7] = blocks & 0xff;
+ buf[8] = (DEV_ATTRIB(dev)->block_size >> 24) & 0xff;
+ buf[9] = (DEV_ATTRIB(dev)->block_size >> 16) & 0xff;
+ buf[10] = (DEV_ATTRIB(dev)->block_size >> 8) & 0xff;
+ buf[11] = DEV_ATTRIB(dev)->block_size & 0xff;
+
+ return 0;
+}
+
+static int transport_modesense_rwrecovery(unsigned char *p)
+{
+ p[0] = 0x01;
+ p[1] = 0x0a;
+
+ return 12;
+}
+
+static int transport_modesense_control(se_device_t *dev, unsigned char *p)
+{
+ p[0] = 0x0a;
+ p[1] = 0x0a;
+ p[2] = 2;
+ /*
+ * From spc4r17, section 7.4.6 Control mode Page
+ *
+ * Unit Attention interlocks control (UN_INTLCK_CTRL) to code 00b
+ *
+ * 00b: The logical unit shall clear any unit attention condition
+ * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
+ * status and shall not establish a unit attention condition when a com-
+ * mand is completed with BUSY, TASK SET FULL, or RESERVATION CONFLICT
+ * status.
+ *
+ * 10b: The logical unit shall not clear any unit attention condition
+ * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
+ * status and shall not establish a unit attention condition when
+ * a command is completed with BUSY, TASK SET FULL, or RESERVATION
+ * CONFLICT status.
+ *
+ * 11b a The logical unit shall not clear any unit attention condition
+ * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
+ * status and shall establish a unit attention condition for the
+ * initiator port associated with the I_T nexus on which the BUSY,
+ * TASK SET FULL, or RESERVATION CONFLICT status is being returned.
+ * Depending on the status, the additional sense code shall be set to
+ * PREVIOUS BUSY STATUS, PREVIOUS TASK SET FULL STATUS, or PREVIOUS
+ * RESERVATION CONFLICT STATUS. Until it is cleared by a REQUEST SENSE
+ * command, a unit attention condition shall be established only once
+ * for a BUSY, TASK SET FULL, or RESERVATION CONFLICT status regardless
+ * to the number of commands completed with one of those status codes.
+ */
+ p[4] = (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl == 2) ? 0x30 :
+ (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl == 1) ? 0x20 : 0x00;
+ /*
+ * From spc4r17, section 7.4.6 Control mode Page
+ *
+ * Task Aborted Status (TAS) bit set to zero.
+ *
+ * A task aborted status (TAS) bit set to zero specifies that aborted
+ * tasks shall be terminated by the device server without any response
+ * to the application client. A TAS bit set to one specifies that tasks
+ * aborted by the actions of an I_T nexus other than the I_T nexus on
+ * which the command was received shall be completed with TASK ABORTED
+ * status (see SAM-4).
+ */
+ p[5] = (DEV_ATTRIB(dev)->emulate_tas) ? 0x40 : 0x00;
+ p[8] = 0xff;
+ p[9] = 0xff;
+ p[11] = 30;
+
+ return 12;
+}
+
+static int transport_modesense_caching(unsigned char *p)
+{
+ p[0] = 0x08;
+ p[1] = 0x12;
+#if 0
+ p[2] = 0x04; /* Write Cache Enable */
+#endif
+ p[12] = 0x20; /* Disabled Read Ahead */
+
+ return 20;
+}
+
+#if 0
+static int transport_modesense_devicecaps(unsigned char *p)
+{
+ p[0] = 0x2a;
+ p[1] = 0x0a;
+
+ return 12;
+}
+#endif
+
+static void transport_modesense_write_protect(
+ unsigned char *buf,
+ int type)
+{
+ /*
+ * I believe that the WP bit (bit 7) in the mode header is the same for
+ * all device types..
+ */
+ switch (type) {
+ case TYPE_DISK:
+ case TYPE_TAPE:
+ default:
+ buf[0] |= 0x80; /* WP bit */
+ break;
+ }
+}
+
+int transport_generic_emulate_modesense(
+ se_cmd_t *cmd,
+ unsigned char *cdb,
+ unsigned char *rbuf,
+ int ten,
+ int type)
+{
+ se_device_t *dev = SE_DEV(cmd);
+ int offset = (ten) ? 8 : 4;
+ int length = 0;
+ unsigned char buf[SE_MODE_PAGE_BUF];
+
+ memset(buf, 0, SE_MODE_PAGE_BUF);
+
+ switch (cdb[2] & 0x3f) {
+ case 0x01:
+ length = transport_modesense_rwrecovery(&buf[offset]);
+ break;
+ case 0x08:
+ length = transport_modesense_caching(&buf[offset]);
+ break;
+ case 0x0a:
+ length = transport_modesense_control(dev, &buf[offset]);
+ break;
+#if 0
+ case 0x2a:
+ length = transport_modesense_devicecaps(&buf[offset]);
+ break;
+#endif
+ case 0x3f:
+ length = transport_modesense_rwrecovery(&buf[offset]);
+ length += transport_modesense_caching(&buf[offset+length]);
+ length += transport_modesense_control(dev, &buf[offset+length]);
+#if 0
+ length += transport_modesense_devicecaps(&buf[offset+length]);
+#endif
+ break;
+ default:
+ printk(KERN_ERR "Got Unknown Mode Page: 0x%02x\n",
+ cdb[2] & 0x3f);
+ return PYX_TRANSPORT_UNKNOWN_MODE_PAGE;
+ }
+ offset += length;
+
+ if (ten) {
+ offset -= 2;
+ buf[0] = (offset >> 8) & 0xff;
+ buf[1] = offset & 0xff;
+
+ if ((SE_LUN(cmd)->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
+ (cmd->se_deve &&
+ (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
+ transport_modesense_write_protect(&buf[3], type);
+
+ if ((offset + 2) > cmd->data_length)
+ offset = cmd->data_length;
+
+ } else {
+ offset -= 1;
+ buf[0] = offset & 0xff;
+
+ if ((SE_LUN(cmd)->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
+ (cmd->se_deve &&
+ (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
+ transport_modesense_write_protect(&buf[2], type);
+
+ if ((offset + 1) > cmd->data_length)
+ offset = cmd->data_length;
+ }
+ memcpy(rbuf, buf, offset);
+
+ return 0;
+}
+
+int transport_generic_emulate_request_sense(
+ se_cmd_t *cmd,
+ unsigned char *cdb)
+{
+ unsigned char *buf = (unsigned char *) T_TASK(cmd)->t_task_buf;
+ u8 ua_asc = 0, ua_ascq = 0;
+
+ if (cdb[1] & 0x01) {
+ printk(KERN_ERR "REQUEST_SENSE description emulation not"
+ " supported\n");
+ return PYX_TRANSPORT_INVALID_CDB_FIELD;
+ }
+ if (!(core_scsi3_ua_clear_for_request_sense(cmd, &ua_asc, &ua_ascq))) {
+ /*
+ * CURRENT ERROR, UNIT ATTENTION
+ */
+ buf[0] = 0x70;
+ buf[SPC_SENSE_KEY_OFFSET] = UNIT_ATTENTION;
+ /*
+ * Make sure request data length is enough for additional
+ * sense data.
+ */
+ if (cmd->data_length <= 18) {
+ buf[7] = 0x00;
+ return 0;
+ }
+ /*
+ * The Additional Sense Code (ASC) from the UNIT ATTENTION
+ */
+ buf[SPC_ASC_KEY_OFFSET] = ua_asc;
+ buf[SPC_ASCQ_KEY_OFFSET] = ua_ascq;
+ buf[7] = 0x0A;
+ } else {
+ /*
+ * CURRENT ERROR, NO SENSE
+ */
+ buf[0] = 0x70;
+ buf[SPC_SENSE_KEY_OFFSET] = NO_SENSE;
+ /*
+ * Make sure request data length is enough for additional
+ * sense data.
+ */
+ if (cmd->data_length <= 18) {
+ buf[7] = 0x00;
+ return 0;
+ }
+ /*
+ * NO ADDITIONAL SENSE INFORMATION
+ */
+ buf[SPC_ASC_KEY_OFFSET] = 0x00;
+ buf[7] = 0x0A;
+ }
+
+ return 0;
+}
+
+/*
+ * Used to obtain Sense Data from underlying Linux/SCSI struct scsi_cmnd
+ */
+int transport_get_sense_data(se_cmd_t *cmd)
+{
+ unsigned char *buffer = cmd->sense_buffer, *sense_buffer = NULL;
+ se_device_t *dev;
+ se_task_t *task = NULL, *task_tmp;
+ unsigned long flags;
+ u32 offset = 0;
+
+ if (!SE_LUN(cmd)) {
+ printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
+ return -1;
+ }
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (cmd->se_cmd_flags & SCF_SENT_CHECK_CONDITION) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return 0;
+ }
+
+ list_for_each_entry_safe(task, task_tmp,
+ &T_TASK(cmd)->t_task_list, t_list) {
+
+ if (!task->task_sense)
+ continue;
+
+ dev = task->se_dev;
+ if (!(dev))
+ continue;
+
+ if (!TRANSPORT(dev)->get_sense_buffer) {
+ printk(KERN_ERR "TRANSPORT(dev)->get_sense_buffer"
+ " is NULL\n");
+ continue;
+ }
+
+ sense_buffer = TRANSPORT(dev)->get_sense_buffer(task);
+ if (!(sense_buffer)) {
+ printk(KERN_ERR "ITT[0x%08x]_TASK[%d]: Unable to locate"
+ " sense buffer for task with sense\n",
+ CMD_TFO(cmd)->get_task_tag(cmd), task->task_no);
+ continue;
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ offset = CMD_TFO(cmd)->set_fabric_sense_len(cmd,
+ TRANSPORT_SENSE_BUFFER);
+
+ memcpy((void *)&buffer[offset], (void *)sense_buffer,
+ TRANSPORT_SENSE_BUFFER);
+ cmd->scsi_status = task->task_scsi_status;
+ /* Automatically padded */
+ cmd->scsi_sense_length =
+ (TRANSPORT_SENSE_BUFFER + offset);
+
+ printk(KERN_INFO "HBA_[%u]_PLUG[%s]: Set SAM STATUS: 0x%02x"
+ " and sense\n",
+ dev->se_hba->hba_id, TRANSPORT(dev)->name,
+ cmd->scsi_status);
+ return 0;
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ return -1;
+}
+
+/*
+ * Generic function pointers for target_core_mod/ConfigFS
+ */
+#define SET_GENERIC_TRANSPORT_FUNCTIONS(cmd) \
+do { \
+ cmd->transport_allocate_iovecs = \
+ &transport_generic_allocate_iovecs; \
+ cmd->transport_get_task = &transport_generic_get_task; \
+ cmd->transport_map_buffers_to_tasks = \
+ &transport_generic_map_buffers_to_tasks; \
+ cmd->transport_set_iovec_ptrs = \
+ &transport_generic_set_iovec_ptrs; \
+} while (0)
+
+/* transport_generic_cmd_sequencer():
+ *
+ * Generic Command Sequencer that should work for most DAS transport
+ * drivers.
+ *
+ * Called from transport_generic_allocate_tasks() in the $FABRIC_MOD
+ * RX Thread.
+ *
+ * FIXME: Need to support other SCSI OPCODES where as well.
+ */
+static int transport_generic_cmd_sequencer(
+ se_cmd_t *cmd,
+ unsigned char *cdb)
+{
+ se_device_t *dev = SE_DEV(cmd);
+ se_subsystem_dev_t *su_dev = dev->se_sub_dev;
+ int ret = 0, sector_ret = 0;
+ u32 sectors = 0, size = 0, pr_reg_type = 0;
+ /*
+ * Check for an existing UNIT ATTENTION condition
+ */
+ if (core_scsi3_ua_check(cmd, cdb) < 0) {
+ cmd->transport_wait_for_tasks =
+ &transport_nop_wait_for_tasks;
+ transport_get_maps(cmd);
+ return 8; /* UNIT ATTENTION */
+ }
+ /*
+ * Check status for SPC-3 Persistent Reservations
+ */
+ if (T10_RES(su_dev)->t10_reservation_check(cmd, &pr_reg_type) != 0) {
+ if (T10_RES(su_dev)->t10_seq_non_holder(
+ cmd, cdb, pr_reg_type) != 0) {
+ cmd->transport_wait_for_tasks =
+ &transport_nop_wait_for_tasks;
+ transport_get_maps(cmd);
+ return 5; /* RESERVATION CONFLIT */
+ }
+ /*
+ * This means the CDB is allowed for the SCSI Initiator port
+ * when said port is *NOT* holding the legacy SPC-2 or
+ * SPC-3 Persistent Reservation.
+ */
+ }
+
+ switch (cdb[0]) {
+ case READ_6:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_6(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_6;
+ cmd->transport_get_lba = &transport_lba_21;
+ break;
+ case READ_10:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_10(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_10;
+ cmd->transport_get_lba = &transport_lba_32;
+ break;
+ case READ_12:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_12(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_12;
+ cmd->transport_get_lba = &transport_lba_32;
+ break;
+ case READ_16:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_16(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_16;
+ cmd->transport_get_long_lba = &transport_lba_64;
+ break;
+ case WRITE_6:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_6(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_6;
+ cmd->transport_get_lba = &transport_lba_21;
+ break;
+ case WRITE_10:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_10(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_10;
+ cmd->transport_get_lba = &transport_lba_32;
+ break;
+ case WRITE_12:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_12(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_12;
+ cmd->transport_get_lba = &transport_lba_32;
+ break;
+ case WRITE_16:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = transport_get_sectors_16(cdb, cmd, §or_ret);
+ if (sector_ret)
+ return 4;
+ size = transport_get_size(sectors, cdb, cmd);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ cmd->transport_split_cdb = &split_cdb_XX_16;
+ cmd->transport_get_long_lba = &transport_lba_64;
+ break;
+ case 0xa3:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
+ /* MAINTENANCE_IN from SCC-2 */
+ /*
+ * Check for emulated MI_REPORT_TARGET_PGS.
+ */
+ if (cdb[1] == MI_REPORT_TARGET_PGS) {
+ cmd->transport_emulate_cdb =
+ (T10_ALUA(su_dev)->alua_type ==
+ SPC3_ALUA_EMULATED) ?
+ &core_scsi3_emulate_report_target_port_groups :
+ NULL;
+ }
+ size = (cdb[6] << 24) | (cdb[7] << 16) |
+ (cdb[8] << 8) | cdb[9];
+ } else {
+ /* GPCMD_SEND_KEY from multi media commands */
+ size = (cdb[8] << 8) + cdb[9];
+ }
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case MODE_SELECT:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = cdb[4];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 1;
+ break;
+ case MODE_SELECT_10:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[7] << 8) + cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 1;
+ break;
+ case MODE_SENSE:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = cdb[4];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case MODE_SENSE_10:
+ case GPCMD_READ_BUFFER_CAPACITY:
+ case GPCMD_SEND_OPC:
+ case LOG_SELECT:
+ case LOG_SENSE:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[7] << 8) + cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case READ_BLOCK_LIMITS:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = READ_BLOCK_LEN;
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case GPCMD_GET_CONFIGURATION:
+ case GPCMD_READ_FORMAT_CAPACITIES:
+ case GPCMD_READ_DISC_INFO:
+ case GPCMD_READ_TRACK_RZONE_INFO:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[7] << 8) + cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 1;
+ break;
+ case PERSISTENT_RESERVE_IN:
+ case PERSISTENT_RESERVE_OUT:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ cmd->transport_emulate_cdb =
+ (T10_RES(su_dev)->res_type ==
+ SPC3_PERSISTENT_RESERVATIONS) ?
+ &core_scsi3_emulate_pr : NULL;
+ size = (cdb[7] << 8) + cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case GPCMD_MECHANISM_STATUS:
+ case GPCMD_READ_DVD_STRUCTURE:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[8] << 8) + cdb[9];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 1;
+ break;
+ case READ_POSITION:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = READ_POSITION_LEN;
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case 0xa4:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
+ /* MAINTENANCE_OUT from SCC-2 */
+ size = (cdb[6] << 24) | (cdb[7] << 16) |
+ (cdb[8] << 8) | cdb[9];
+ } else {
+ /* GPCMD_REPORT_KEY from multi media commands */
+ size = (cdb[8] << 8) + cdb[9];
+ }
+ CMD_ORIG_OBJ_API(cmd)->get_mem_SG(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 1;
+ break;
+ case INQUIRY:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[3] << 8) + cdb[4];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case READ_BUFFER:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case READ_CAPACITY:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = READ_CAP_LEN;
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case READ_MEDIA_SERIAL_NUMBER:
+ case SECURITY_PROTOCOL_IN:
+ case SECURITY_PROTOCOL_OUT:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case SERVICE_ACTION_IN:
+ case ACCESS_CONTROL_IN:
+ case ACCESS_CONTROL_OUT:
+ case EXTENDED_COPY:
+ case READ_ATTRIBUTE:
+ case RECEIVE_COPY_RESULTS:
+ case WRITE_ATTRIBUTE:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[10] << 24) | (cdb[11] << 16) |
+ (cdb[12] << 8) | cdb[13];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case VARIABLE_LENGTH_CMD:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[10] << 8) | cdb[11];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case RECEIVE_DIAGNOSTIC:
+ case SEND_DIAGNOSTIC:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[3] << 8) | cdb[4];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+/* #warning FIXME: Figure out correct GPCMD_READ_CD blocksize. */
+#if 0
+ case GPCMD_READ_CD:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ sectors = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
+ size = (2336 * sectors);
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+#endif
+ case READ_TOC:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case REQUEST_SENSE:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = cdb[4];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case READ_ELEMENT_STATUS:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = 65536 * cdb[7] + 256 * cdb[8] + cdb[9];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case WRITE_BUFFER:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ case RESERVE:
+ case RESERVE_10:
+ /*
+ * The SPC-2 RESERVE does not contain a size in the SCSI CDB.
+ * Assume the passthrough or $FABRIC_MOD will tell us about it.
+ */
+ if (cdb[0] == RESERVE_10)
+ size = (cdb[7] << 8) | cdb[8];
+ else
+ size = cmd->data_length;
+
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ cmd->transport_allocate_resources =
+ &transport_generic_allocate_none;
+ transport_get_maps(cmd);
+
+ ret = T10_RES(su_dev)->t10_reserve(cmd);
+ if (ret != 0) {
+ if (ret > 0)
+ return 5; /* RESERVATION_CONFILIT */
+ else
+ return 7; /* ILLEGAL_REQUEST */
+ }
+ ret = 3;
+ break;
+ case RELEASE:
+ case RELEASE_10:
+ /*
+ * The SPC-2 RELEASE does not contain a size in the SCSI CDB.
+ * Assume the passthrough or $FABRIC_MOD will tell us about it.
+ */
+ if (cdb[0] == RELEASE_10)
+ size = (cdb[7] << 8) | cdb[8];
+ else
+ size = cmd->data_length;
+
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ cmd->transport_allocate_resources =
+ &transport_generic_allocate_none;
+ transport_get_maps(cmd);
+
+ T10_RES(su_dev)->t10_release(cmd);
+ ret = 3;
+ break;
+ case ALLOW_MEDIUM_REMOVAL:
+ case GPCMD_CLOSE_TRACK:
+ case ERASE:
+ case INITIALIZE_ELEMENT_STATUS:
+ case GPCMD_LOAD_UNLOAD:
+ case REZERO_UNIT:
+ case SEEK_10:
+ case GPCMD_SET_SPEED:
+ case SPACE:
+ case START_STOP:
+ case SYNCHRONIZE_CACHE:
+ case TEST_UNIT_READY:
+ case VERIFY:
+ case WRITE_FILEMARKS:
+ case MOVE_MEDIUM:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ cmd->transport_allocate_resources =
+ &transport_generic_allocate_none;
+ transport_get_maps(cmd);
+ ret = 3;
+ break;
+ case REPORT_LUNS:
+ SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
+ cmd->transport_emulate_cdb =
+ &transport_core_report_lun_response;
+ size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
+ CMD_ORIG_OBJ_API(cmd)->get_mem_buf(cmd->se_orig_obj_ptr, cmd);
+ transport_get_maps(cmd);
+ ret = 2;
+ break;
+ default:
+ printk(KERN_WARNING "TARGET_CORE[%s]: Unsupported SCSI Opcode"
+ " 0x%02x, sending CHECK_CONDITION.\n",
+ CMD_TFO(cmd)->get_fabric_name(), cdb[0]);
+ cmd->transport_wait_for_tasks = &transport_nop_wait_for_tasks;
+ transport_get_maps(cmd);
+ return 4;
+ }
+
+ if (size != cmd->data_length) {
+ printk(KERN_WARNING "TARGET_CORE[%s]: Expected Transfer Length:"
+ " %u does not match SCSI CDB Length: %u for SAM Opcode:"
+ " 0x%02x\n", CMD_TFO(cmd)->get_fabric_name(),
+ cmd->data_length, size, cdb[0]);
+
+ cmd->cmd_spdtl = size;
+
+ if (cmd->data_direction == SE_DIRECTION_WRITE) {
+ printk(KERN_ERR "Rejecting underflow/overflow"
+ " WRITE data\n");
+ return 6;
+ }
+ /*
+ * Reject READ_* or WRITE_* with overflow/underflow for
+ * type SCF_SCSI_DATA_SG_IO_CDB.
+ */
+ if (!(ret) && (DEV_ATTRIB(dev)->block_size != 512)) {
+ printk(KERN_ERR "Failing OVERFLOW/UNDERFLOW for LBA op"
+ " CDB on non 512-byte sector setup subsystem"
+ " plugin: %s\n", TRANSPORT(dev)->name);
+ /* Returns CHECK_CONDITION + INVALID_CDB_FIELD */
+ return 6;
+ }
+
+ if (size > cmd->data_length) {
+ cmd->se_cmd_flags |= SCF_OVERFLOW_BIT;
+ cmd->residual_count = (size - cmd->data_length);
+ } else {
+ cmd->se_cmd_flags |= SCF_UNDERFLOW_BIT;
+ cmd->residual_count = (cmd->data_length - size);
+ }
+ cmd->data_length = size;
+ }
+
+ transport_set_supported_SAM_opcode(cmd);
+ return ret;
+}
+
+static inline se_cmd_t *transport_alloc_passthrough_cmd(
+ u32 data_length,
+ int data_direction)
+{
+ return __transport_alloc_se_cmd(&passthrough_fabric_ops, NULL, NULL,
+ data_length, data_direction);
+}
+
+se_cmd_t *transport_allocate_passthrough(
+ unsigned char *cdb,
+ int data_direction,
+ u32 se_cmd_flags,
+ void *mem,
+ u32 se_mem_num,
+ u32 length,
+ se_obj_lun_type_t *obj_api,
+ void *type_ptr)
+{
+ se_cmd_t *cmd;
+ se_transform_info_t ti;
+
+ cmd = transport_alloc_passthrough_cmd(length, data_direction);
+ if (!(cmd))
+ return NULL;
+ /*
+ * Simulate an SE LUN entry need for passing SCSI CDBs into
+ * se_cmd_t.
+ */
+ cmd->se_lun = kzalloc(sizeof(se_lun_t), GFP_KERNEL);
+ if (!(cmd->se_lun)) {
+ printk(KERN_ERR "Unable to allocate cmd->se_lun\n");
+ goto fail;
+ }
+
+ spin_lock_init(&cmd->se_lun->lun_sep_lock);
+ SE_LUN(cmd)->lun_type = obj_api->se_obj_type;
+ SE_LUN(cmd)->lun_type_ptr = type_ptr;
+ SE_LUN(cmd)->lun_obj_api = obj_api;
+
+ cmd->se_orig_obj_api = obj_api;
+ cmd->se_orig_obj_ptr = type_ptr;
+ cmd->se_cmd_flags = se_cmd_flags;
+ SE_LUN(cmd)->se_dev = (se_device_t *) type_ptr;
+
+ /*
+ * Double check that the passed object is currently accepting CDBs
+ */
+ if (obj_api->check_online(type_ptr) != 0) {
+ DEBUG_SO("obj_api->check_online() failed!\n");
+ goto fail;
+ }
+
+ cmd->data_length = length;
+ cmd->data_direction = data_direction;
+ cmd->se_cmd_flags |= SCF_CMD_PASSTHROUGH;
+
+ if (transport_generic_allocate_tasks(cmd, cdb) < 0)
+ goto fail;
+
+ memset(&ti, 0, sizeof(se_transform_info_t));
+ ti.ti_data_length = cmd->data_length;
+ ti.ti_dev = SE_LUN(cmd)->se_dev;
+ ti.ti_se_cmd = cmd;
+ ti.se_obj_ptr = type_ptr;
+ ti.se_obj_api = SE_LUN(cmd)->lun_obj_api;
+
+ DEBUG_SO("ti.se_obj_ptr: %p\n", ti.se_obj_ptr);
+ DEBUG_SO("ti.se_obj_api: %p\n", ti.se_obj_api);
+ DEBUG_SO("Plugin: %s\n", ti.se_obj_api->obj_plugin->plugin_name);
+
+ if (!mem) {
+ if (cmd->transport_allocate_resources(cmd, cmd->data_length,
+ PAGE_SIZE) < 0)
+ goto fail;
+ } else {
+ /*
+ * Passed *mem will contain a list_head containing preformatted
+ * se_mem_t elements...
+ */
+ T_TASK(cmd)->t_mem_list = (struct list_head *)mem;
+ T_TASK(cmd)->t_task_se_num = se_mem_num;
+ cmd->se_cmd_flags |= SCF_CMD_PASSTHROUGH_NOALLOC;
+
+#ifdef DEBUG_PASSTHROUGH
+ {
+ u32 total_se_length = 0;
+ se_mem_t *se_mem, *se_mem_tmp;
+
+ DEBUG_PT("Preallocated se_mem_list: %p se_mem_num: %d\n",
+ mem, se_mem_num);
+
+ list_for_each_entry_safe(se_mem, se_mem_tmp,
+ T_TASK(cmd)->t_mem_list, se_list) {
+ total_se_length += se_mem->se_len;
+ DEBUG_PT("se_mem: %p se_mem->se_page: %p %d:%d\n",
+ se_mem, se_mem->se_page, se_mem->se_len,
+ se_mem->se_off);
+ }
+ DEBUG_PT("Total calculated total_se_length: %u\n",
+ total_se_length);
+
+ if (total_se_length != length) {
+ printk(KERN_ERR "Passed length: %u does not equal"
+ " total_se_length: %u\n", length,
+ total_se_length);
+ BUG();
+ }
+ }
+#endif
+ }
+
+ if (transport_get_sectors(cmd, SE_LUN(cmd)->lun_obj_api, type_ptr) < 0)
+ goto fail;
+
+ if (transport_new_cmd_obj(cmd, &ti, SE_LUN(cmd)->lun_obj_api,
+ type_ptr, 0) < 0)
+ goto fail;
+
+ return cmd;
+
+fail:
+ if (T_TASK(cmd))
+ transport_release_tasks(cmd);
+ kfree(T_TASK(cmd));
+ kfree(cmd->se_lun);
+ transport_free_se_cmd(cmd);
+
+ return NULL;
+}
+
+void transport_passthrough_release(
+ se_cmd_t *cmd)
+{
+ if (!cmd) {
+ printk(KERN_ERR "transport_passthrough_release passed"
+ " NULL se_cmd_t\n");
+ return;
+ }
+
+ if (cmd->transport_wait_for_tasks)
+ cmd->transport_wait_for_tasks(cmd, 0, 0);
+
+ transport_generic_remove(cmd, 0, 0);
+}
+
+int transport_passthrough_complete(
+ se_cmd_t *cmd)
+{
+ if (cmd->se_orig_obj_api->check_shutdown(cmd->se_orig_obj_ptr) != 0)
+ return -2;
+
+ switch (cmd->scsi_status) {
+ case 0x00: /* GOOD */
+ DEBUG_PT("SCSI Status: GOOD\n");
+ return 0;
+ case 0x02: /* CHECK_CONDITION */
+ DEBUG_PT("SCSI Status: CHECK_CONDITION\n");
+/* #warning FIXME: Do some basic return values for Sense Data */
+ return -1;
+ default:
+ DEBUG_PT("SCSI Status: 0x%02x\n", cmd->scsi_status);
+ return -1;
+ }
+
+ return 0;
+}
+
+/* transport_generic_passthrough():
+ *
+ *
+ */
+int transport_generic_passthrough_async(
+ se_cmd_t *cmd,
+ void (*callback)(se_cmd_t *cmd,
+ void *callback_arg, int complete_status),
+ void *callback_arg)
+{
+ int write = (cmd->data_direction == SE_DIRECTION_WRITE);
+ int no_alloc = (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC);
+ int pt_done = (cmd->transport_passthrough_done != NULL);
+
+ if (callback) {
+ cmd->callback = callback;
+ cmd->callback_arg = callback_arg;
+ }
+
+ if (transport_generic_handle_cdb(cmd) < 0)
+ return -1;
+
+ if (write && !no_alloc) {
+ if (down_interruptible(
+ &T_TASK(cmd)->t_transport_passthrough_wsem) != 0)
+ return -1;
+
+ transport_generic_process_write(cmd);
+ }
+
+ if (callback || pt_done)
+ return 0;
+
+ down(&T_TASK(cmd)->t_transport_passthrough_sem);
+
+ return transport_passthrough_complete(cmd);
+}
+
+int transport_generic_passthrough(se_cmd_t *cmd)
+{
+ return transport_generic_passthrough_async(cmd, NULL, NULL);
+}
+
+/* transport_generic_complete_ok():
+ *
+ *
+ */
+void transport_generic_complete_ok(se_cmd_t *cmd)
+{
+ int reason = 0;
+ /*
+ * Check if we need to retrieve a sense buffer from
+ * the se_cmd_t in question.
+ */
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH) {
+ transport_lun_remove_cmd(cmd);
+ if (!(transport_cmd_check_stop(cmd, 2, 0)))
+ transport_passthrough_check_stop(cmd);
+ return;
+ } else if (cmd->se_cmd_flags & SCF_TRANSPORT_TASK_SENSE) {
+ if (transport_get_sense_data(cmd) < 0)
+ reason = NON_EXISTENT_LUN;
+
+ /*
+ * Only set when an se_task_t->task_scsi_status returned
+ * a non GOOD status.
+ */
+ if (cmd->scsi_status) {
+ transport_send_check_condition_and_sense(
+ cmd, reason, 1);
+ transport_lun_remove_cmd(cmd);
+ transport_cmd_check_stop(cmd, 2, 0);
+ return;
+ }
+ }
+
+ switch (cmd->data_direction) {
+ case SE_DIRECTION_READ:
+#ifdef SNMP_SUPPORT
+ spin_lock(&cmd->se_lun->lun_sep_lock);
+ if (SE_LUN(cmd)->lun_sep) {
+ SE_LUN(cmd)->lun_sep->sep_stats.tx_data_octets +=
+ cmd->data_length;
+ }
+ spin_unlock(&cmd->se_lun->lun_sep_lock);
+#endif
+ CMD_TFO(cmd)->queue_data_in(cmd);
+ break;
+ case SE_DIRECTION_WRITE:
+#ifdef SNMP_SUPPORT
+ spin_lock(&cmd->se_lun->lun_sep_lock);
+ if (SE_LUN(cmd)->lun_sep) {
+ SE_LUN(cmd)->lun_sep->sep_stats.rx_data_octets +=
+ cmd->data_length;
+ }
+ spin_unlock(&cmd->se_lun->lun_sep_lock);
+#endif
+ /* Fall through for SE_DIRECTION_WRITE */
+ case SE_DIRECTION_NONE:
+ CMD_TFO(cmd)->queue_status(cmd);
+ break;
+ default:
+ break;
+ }
+
+ transport_lun_remove_cmd(cmd);
+ transport_cmd_check_stop(cmd, 2, 0);
+}
+
+void transport_free_dev_tasks(se_cmd_t *cmd)
+{
+ se_task_t *task, *task_tmp;
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ list_for_each_entry_safe(task, task_tmp,
+ &T_TASK(cmd)->t_task_list, t_list) {
+ if (atomic_read(&task->task_active))
+ continue;
+
+ if (!task->transport_req)
+ continue;
+
+ kfree(task->task_sg);
+
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ if (task->se_dev)
+ TRANSPORT(task->se_dev)->free_task(task);
+ else
+ printk(KERN_ERR "task[%u] - task->se_dev is NULL\n",
+ task->task_no);
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+
+ list_del(&task->t_list);
+ kmem_cache_free(se_task_cache, task);
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+}
+
+static inline void transport_free_pages(se_cmd_t *cmd)
+{
+ se_mem_t *se_mem, *se_mem_tmp;
+
+ if (T_TASK(cmd)->t_task_buf) {
+ kfree(T_TASK(cmd)->t_task_buf);
+ T_TASK(cmd)->t_task_buf = NULL;
+ return;
+ }
+
+ if (cmd->transport_free_resources) {
+ cmd->transport_free_resources(cmd);
+ return;
+ }
+ /*
+ * Caller will handle releasing of se_mem_t.
+ */
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC)
+ return;
+
+ if (!(T_TASK(cmd)->t_task_se_num))
+ return;
+
+ list_for_each_entry_safe(se_mem, se_mem_tmp,
+ T_TASK(cmd)->t_mem_list, se_list) {
+ __free_page(se_mem->se_page);
+
+ list_del(&se_mem->se_list);
+ kfree(se_mem);
+ }
+
+ kfree(T_TASK(cmd)->t_mem_list);
+ T_TASK(cmd)->t_mem_list = NULL;
+ T_TASK(cmd)->t_task_se_num = 0;
+}
+
+void transport_release_tasks(se_cmd_t *cmd)
+{
+ CMD_ORIG_OBJ_API(cmd)->free_tasks(cmd->se_orig_obj_ptr, cmd);
+}
+
+static inline int transport_dec_and_check(se_cmd_t *cmd)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+ if (!(atomic_dec_and_test(&T_TASK(cmd)->t_fe_count))) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+ return 1;
+ }
+ }
+
+ if (atomic_read(&T_TASK(cmd)->t_se_count)) {
+ if (!(atomic_dec_and_test(&T_TASK(cmd)->t_se_count))) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+ return 1;
+ }
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ return 0;
+}
+
+void transport_release_fe_cmd(se_cmd_t *cmd)
+{
+ unsigned long flags;
+
+ if (transport_dec_and_check(cmd))
+ return;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ transport_all_task_dev_remove_state(cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ transport_release_tasks(cmd);
+ transport_free_pages(cmd);
+
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+ kfree(cmd->se_lun);
+
+ CMD_TFO(cmd)->release_cmd_direct(cmd);
+ transport_free_se_cmd(cmd);
+}
+
+/* transport_generic_remove():
+ *
+ *
+ */
+int transport_generic_remove(
+ se_cmd_t *cmd,
+ int release_to_pool,
+ int session_reinstatement)
+{
+ unsigned long flags;
+
+ if (!(T_TASK(cmd)))
+ goto release_cmd;
+
+ if (transport_dec_and_check(cmd)) {
+ if (session_reinstatement) {
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ transport_all_task_dev_remove_state(cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+ }
+ return 1;
+ }
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ transport_all_task_dev_remove_state(cmd);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ transport_release_tasks(cmd);
+ transport_free_pages(cmd);
+
+release_cmd:
+ if (release_to_pool && !(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+ transport_release_cmd_to_pool(cmd);
+ else {
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+ kfree(cmd->se_lun);
+
+ CMD_TFO(cmd)->release_cmd_direct(cmd);
+ transport_free_se_cmd(cmd);
+ }
+
+ return 0;
+}
+
+/* transport_generic_map_buffers_to_tasks():
+ *
+ * Called from transport_generic_new_cmd() in Transport Processing Thread.
+ */
+static int transport_generic_map_buffers_to_tasks(se_cmd_t *cmd)
+{
+ se_task_t *task = NULL;
+ int ret;
+
+ /*
+ * Deal with non [READ,WRITE]_XX CDBs here.
+ */
+ if (cmd->se_cmd_flags & SCF_SCSI_NON_DATA_CDB)
+ goto non_scsi_data;
+ else if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_NONSG_IO_CDB) {
+ list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+ if (atomic_read(&task->task_sent))
+ continue;
+
+ ret = task->transport_map_task(task, task->task_size);
+ if (ret < 0)
+ return ret;
+
+ DEBUG_CMD_M("Mapping SCF_SCSI_CONTROL_NONSG_IO_CDB"
+ " task_size: %u\n", task->task_size);
+ }
+ return 0;
+ }
+
+ /*
+ * Determine the scatterlist offset for each se_task_t,
+ * and segment and set pointers to storage transport buffers
+ * via task->transport_map_task().
+ */
+ list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+ if (atomic_read(&task->task_sent))
+ continue;
+
+ ret = task->transport_map_task(task, task->task_size);
+ if (ret < 0)
+ return ret;
+
+ DEBUG_CMD_M("Mapping task[%d]_se_obj_ptr[%p] %s_IO task_lba:"
+ " %llu task_size: %u task_sg_num: %d\n",
+ task->task_no, task->se_obj_ptr,
+ (cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB) ?
+ "CONTROL" : "DATA", task->task_lba, task->task_size,
+ task->task_sg_num);
+ }
+
+ return 0;
+
+non_scsi_data:
+ list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+ if (atomic_read(&task->task_sent))
+ continue;
+
+ ret = task->transport_map_task(task, task->task_size);
+ if (ret < 0)
+ return ret;
+
+ DEBUG_CMD_M("Mapping SCF_SCSI_NON_DATA_CDB task_size: %u"
+ " task->task_sg_num: %d\n", task->task_size,
+ task->task_sg_num);
+ }
+
+ return 0;
+}
+
+/* transport_generic_do_transform():
+ *
+ *
+ */
+int transport_generic_do_transform(se_cmd_t *cmd, se_transform_info_t *ti)
+{
+ if (cmd->transport_cdb_transform(cmd, ti) < 0)
+ return -1;
+
+ return 0;
+}
+
+int transport_get_sectors(
+ se_cmd_t *cmd,
+ se_obj_lun_type_t *obj_api,
+ void *obj_ptr)
+{
+ if (!(cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB))
+ return 0;
+
+ T_TASK(cmd)->t_task_sectors =
+ (cmd->data_length / obj_api->blocksize(obj_ptr));
+ if (!(T_TASK(cmd)->t_task_sectors))
+ T_TASK(cmd)->t_task_sectors = 1;
+
+ if (obj_api->get_device_type(obj_ptr) != TYPE_DISK)
+ return 0;
+
+ if ((T_TASK(cmd)->t_task_lba + T_TASK(cmd)->t_task_sectors) >
+ obj_api->total_sectors(obj_ptr)) {
+ printk(KERN_ERR "LBA: %llu Sectors: %u exceeds"
+ " obj_api->total_sectors(): %llu\n",
+ T_TASK(cmd)->t_task_lba, T_TASK(cmd)->t_task_sectors,
+ obj_api->total_sectors(obj_ptr));
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->scsi_sense_reason = SECTOR_COUNT_TOO_MANY;
+ return PYX_TRANSPORT_REQ_TOO_MANY_SECTORS;
+ }
+
+ return 0;
+}
+
+int transport_new_cmd_obj(
+ se_cmd_t *cmd,
+ se_transform_info_t *ti,
+ se_obj_lun_type_t *obj_api,
+ void *obj_ptr,
+ int post_execute)
+{
+ u32 task_cdbs = 0, task_offset = 0;
+ se_mem_t *se_mem_out = NULL;
+
+ if (!(cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)) {
+ task_cdbs++;
+ T_TASK(cmd)->t_task_cdbs++;
+ } else {
+ ti->ti_set_counts = 1;
+
+ task_cdbs = obj_api->get_cdb_count(obj_ptr, ti,
+ T_TASK(cmd)->t_task_lba,
+ T_TASK(cmd)->t_task_sectors,
+ NULL, &se_mem_out, &task_offset);
+ if (!(task_cdbs)) {
+ cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+ cmd->scsi_sense_reason =
+ LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ return PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+ }
+ T_TASK(cmd)->t_task_cdbs += task_cdbs;
+
+ cmd->transport_cdb_transform =
+ &transport_process_data_sg_transform;
+#if 0
+ printk(KERN_INFO "[%s]: api: %p ptr: %p data_length: %u, LBA:"
+ " %llu t_task_sectors: %u, t_task_cdbs: %u\n",
+ obj_api->obj_plugin->plugin_name, obj_api, obj_ptr,
+ cmd->data_length, T_TASK(cmd)->t_task_lba,
+ T_TASK(cmd)->t_task_sectors, T_TASK(cmd)->t_task_cdbs);
+#endif
+ }
+
+ cmd->transport_do_transform = &transport_generic_do_transform;
+ if (!post_execute) {
+ atomic_set(&T_TASK(cmd)->t_task_cdbs_left, task_cdbs);
+ atomic_set(&T_TASK(cmd)->t_task_cdbs_ex_left, task_cdbs);
+ atomic_set(&T_TASK(cmd)->t_task_cdbs_timeout_left, task_cdbs);
+ } else {
+ atomic_add(task_cdbs, &T_TASK(cmd)->t_task_cdbs_left);
+ atomic_add(task_cdbs, &T_TASK(cmd)->t_task_cdbs_ex_left);
+ atomic_add(task_cdbs, &T_TASK(cmd)->t_task_cdbs_timeout_left);
+ }
+
+ return 0;
+}
+
+unsigned char *transport_get_vaddr(se_mem_t *se_mem)
+{
+ return page_address(se_mem->se_page) + se_mem->se_off;
+}
+
+struct list_head *transport_init_se_mem_list(void)
+{
+ struct list_head *se_mem_list;
+
+ se_mem_list = kzalloc(sizeof(struct list_head), GFP_KERNEL);
+ if (!(se_mem_list)) {
+ printk(KERN_ERR "Unable to allocate memory for se_mem_list\n");
+ return NULL;
+ }
+ INIT_LIST_HEAD(se_mem_list);
+
+ return se_mem_list;
+}
+
+void transport_free_se_mem_list(struct list_head *se_mem_list)
+{
+ se_mem_t *se_mem, *se_mem_tmp;
+
+ if (!se_mem_list)
+ return;
+
+ list_for_each_entry_safe(se_mem, se_mem_tmp, se_mem_list, se_list) {
+ list_del(&se_mem->se_list);
+ kfree(se_mem);
+ }
+ kfree(se_mem_list);
+}
+
+int transport_generic_get_mem(se_cmd_t *cmd, u32 length, u32 dma_size)
+{
+ unsigned char *buf;
+ se_mem_t *se_mem;
+
+ T_TASK(cmd)->t_mem_list = transport_init_se_mem_list();
+ if (!(T_TASK(cmd)->t_mem_list))
+ return -1;
+
+ while (length) {
+ se_mem = kzalloc(sizeof(se_mem_t), GFP_KERNEL);
+ if (!(se_mem)) {
+ printk(KERN_ERR "Unable to allocate se_mem_t\n");
+ goto out;
+ }
+ INIT_LIST_HEAD(&se_mem->se_list);
+ se_mem->se_len = (length > dma_size) ? dma_size : length;
+
+/* #warning FIXME Allocate contigous pages for se_mem_t elements */
+ se_mem->se_page = (struct page *) alloc_pages(GFP_KERNEL, 0);
+ if (!(se_mem->se_page)) {
+ printk(KERN_ERR "alloc_pages() failed\n");
+ goto out;
+ }
+
+ buf = kmap_atomic(se_mem->se_page, KM_IRQ0);
+ if (!(buf)) {
+ printk(KERN_ERR "kmap_atomic() failed\n");
+ goto out;
+ }
+ memset(buf, 0, se_mem->se_len);
+ kunmap_atomic(buf, KM_IRQ0);
+
+ list_add_tail(&se_mem->se_list, T_TASK(cmd)->t_mem_list);
+ T_TASK(cmd)->t_task_se_num++;
+
+ DEBUG_MEM("Allocated se_mem_t page(%p) Length(%u)"
+ " Offset(%u)\n", se_mem->se_page, se_mem->se_len,
+ se_mem->se_off);
+
+ length -= se_mem->se_len;
+ }
+
+ DEBUG_MEM("Allocated total se_mem_t elements(%u)\n",
+ T_TASK(cmd)->t_task_se_num);
+
+ return 0;
+out:
+ return -1;
+}
+
+extern u32 transport_calc_sg_num (
+ se_task_t *task,
+ se_mem_t *in_se_mem,
+ u32 task_offset)
+{
+ se_mem_t *se_mem = in_se_mem;
+ u32 sg_length, sg_offset, task_size = task->task_size;
+ u32 saved_task_offset = 0;
+
+ while (task_size) {
+ DEBUG_SC("se_mem->se_page(%p) se_mem->se_len(%u)"
+ " se_mem->se_off(%u) task_offset(%u)\n",
+ se_mem->se_page, se_mem->se_len,
+ se_mem->se_off, task_offset);
+
+ if (task_offset == 0) {
+ if (task_size > se_mem->se_len)
+ sg_length = se_mem->se_len;
+ else
+ sg_length = task_size;
+
+ DEBUG_SC("sg_length(%u) task_size(%u)\n",
+ sg_length, task_size);
+
+ if (saved_task_offset)
+ task_offset = saved_task_offset;
+ } else {
+ sg_offset = task_offset;
+
+ if ((se_mem->se_len - task_offset) > task_size)
+ sg_length = task_size;
+ else
+ sg_length = (se_mem->se_len - task_offset);
+
+ DEBUG_SC("sg_length(%u) task_size(%u)\n",
+ sg_length, task_size);
+
+ saved_task_offset = task_offset;
+ task_offset = 0;
+ }
+ task_size -= sg_length;
+
+ DEBUG_SC("task[%u] - Reducing task_size to(%u)\n",
+ task->task_no, task_size);
+
+ task->task_sg_num++;
+
+ list_for_each_entry_continue(se_mem,
+ task->task_se_cmd->t_task->t_mem_list, se_list)
+ break;
+
+ if (!se_mem)
+ break;
+ }
+
+ task->task_sg = kzalloc(task->task_sg_num *
+ sizeof(struct scatterlist), GFP_KERNEL);
+ if (!(task->task_sg)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " task->task_sg\n");
+ return 0;
+ }
+
+ sg_init_table(&task->task_sg[0], task->task_sg_num);
+
+ DEBUG_SC("Successfully allocated task->task_sg_num(%u)\n",
+ task->task_sg_num);
+
+ return task->task_sg_num;
+}
+
+static inline int transport_set_task_sectors_disk(
+ se_task_t *task,
+ se_obj_lun_type_t *obj_api,
+ void *obj_ptr,
+ unsigned long long lba,
+ u32 sectors,
+ int *max_sectors_set)
+{
+ if ((lba + sectors) > obj_api->end_lba(obj_ptr, 1)) {
+ task->task_sectors = ((obj_api->end_lba(obj_ptr, 1) - lba) + 1);
+
+ if (task->task_sectors > obj_api->max_sectors(obj_ptr)) {
+ task->task_sectors = obj_api->max_sectors(obj_ptr);
+ *max_sectors_set = 1;
+ }
+ } else {
+ if (sectors > obj_api->max_sectors(obj_ptr)) {
+ task->task_sectors = obj_api->max_sectors(obj_ptr);
+ *max_sectors_set = 1;
+ } else
+ task->task_sectors = sectors;
+ }
+
+ return 0;
+}
+
+static inline int transport_set_task_sectors_non_disk(
+ se_task_t *task,
+ se_obj_lun_type_t *obj_api,
+ void *obj_ptr,
+ unsigned long long lba,
+ u32 sectors,
+ int *max_sectors_set)
+{
+ if (sectors > obj_api->max_sectors(obj_ptr)) {
+ task->task_sectors = obj_api->max_sectors(obj_ptr);
+ *max_sectors_set = 1;
+ } else
+ task->task_sectors = sectors;
+
+ return 0;
+}
+
+static inline int transport_set_task_sectors(
+ se_task_t *task,
+ se_obj_lun_type_t *obj_api,
+ void *obj_ptr,
+ unsigned long long lba,
+ u32 sectors,
+ int *max_sectors_set)
+{
+ return (obj_api->get_device_type(obj_ptr) == TYPE_DISK) ?
+ transport_set_task_sectors_disk(task, obj_api, obj_ptr,
+ lba, sectors, max_sectors_set) :
+ transport_set_task_sectors_non_disk(task, obj_api, obj_ptr,
+ lba, sectors, max_sectors_set);
+}
+
+int transport_map_sg_to_mem(
+ se_task_t *task,
+ struct list_head *se_mem_list,
+ void *in_mem,
+ se_mem_t *in_se_mem,
+ se_mem_t **out_se_mem,
+ u32 *se_mem_cnt,
+ u32 *task_offset)
+{
+ se_mem_t *se_mem;
+ struct scatterlist *sg_s;
+ u32 j = 0, saved_task_offset = 0, task_size = task->task_size;
+
+ if (!in_mem) {
+ printk(KERN_ERR "No source scatterlist\n");
+ return -1;
+ }
+ sg_s = (struct scatterlist *)in_mem;
+
+ while (task_size) {
+ se_mem = kzalloc(sizeof(se_mem_t), GFP_KERNEL);
+ if (!(se_mem)) {
+ printk(KERN_ERR "Unable to allocate se_mem_t\n");
+ return -1;
+ }
+ INIT_LIST_HEAD(&se_mem->se_list);
+
+ if (*task_offset == 0) {
+ se_mem->se_page = sg_page(&sg_s[j]);
+ se_mem->se_off = sg_s[j].offset;
+
+ if (task_size >= sg_s[j].length)
+ se_mem->se_len = sg_s[j++].length;
+ else {
+ se_mem->se_len = task_size;
+
+ task_size -= se_mem->se_len;
+ if (!(task_size)) {
+ *task_offset = (se_mem->se_len +
+ saved_task_offset);
+ goto next;
+ }
+ }
+
+ if (saved_task_offset)
+ *task_offset = saved_task_offset;
+ } else {
+ se_mem->se_page = sg_page(&sg_s[j]);
+ se_mem->se_off = (*task_offset + sg_s[j].offset);
+
+ if ((sg_s[j].length - *task_offset) > task_size) {
+ se_mem->se_len = task_size;
+
+ task_size -= se_mem->se_len;
+ if (!(task_size)) {
+ *task_offset += se_mem->se_len;
+ goto next;
+ }
+ } else
+ se_mem->se_len = (sg_s[j++].length -
+ *task_offset);
+
+ saved_task_offset = *task_offset;
+ *task_offset = 0;
+ }
+ task_size -= se_mem->se_len;
+next:
+ list_add_tail(&se_mem->se_list, se_mem_list);
+ (*se_mem_cnt)++;
+ }
+
+ DEBUG_MEM("task[%u] - Mapped(%u) struct scatterlist segments to(%u)"
+ " se_mem_t\n", task->task_no, j, *se_mem_cnt);
+
+ return 0;
+}
+
+int transport_map_mem_to_mem(
+ se_task_t *task,
+ struct list_head *se_mem_list,
+ void *in_mem,
+ se_mem_t *in_se_mem,
+ se_mem_t **out_se_mem,
+ u32 *se_mem_cnt,
+ u32 *task_offset)
+{
+ se_mem_t *se_mem = in_se_mem, *se_mem_new;
+ u32 saved_task_offset = 0, task_size = task->task_size;
+
+ if (!se_mem) {
+ printk(KERN_ERR "Invalid se_mem_t pointer\n");
+ return -1;
+ }
+
+ while (task_size) {
+ se_mem_new = kzalloc(sizeof(se_mem_t), GFP_KERNEL);
+ if (!(se_mem_new)) {
+ printk(KERN_ERR "Unable to allocate se_mem_t\n");
+ return -1;
+ }
+ INIT_LIST_HEAD(&se_mem_new->se_list);
+
+ if (*task_offset == 0) {
+ se_mem_new->se_page = se_mem->se_page;
+ se_mem_new->se_off = se_mem->se_off;
+
+ if (task_size >= se_mem->se_len) {
+ se_mem_new->se_len = se_mem->se_len;
+
+ se_mem = list_entry(se_mem->se_list.next,
+ se_mem_t, se_list);
+ if (!(se_mem)) {
+ printk(KERN_ERR "Unable to locate next"
+ " se_mem_t\n");
+ return -1;
+ }
+ } else {
+ se_mem_new->se_len = task_size;
+
+ task_size -= se_mem_new->se_len;
+ if (!(task_size)) {
+ *task_offset = (se_mem_new->se_len +
+ saved_task_offset);
+ goto next;
+ }
+ }
+
+ if (saved_task_offset)
+ *task_offset = saved_task_offset;
+ } else {
+ se_mem_new->se_page = se_mem->se_page;
+ se_mem_new->se_off = (*task_offset + se_mem->se_off);
+
+ if ((se_mem->se_len - *task_offset) > task_size) {
+ se_mem_new->se_len = task_size;
+
+ if (!(task_size -= se_mem_new->se_len)) {
+ *task_offset += se_mem_new->se_len;
+ goto next;
+ }
+ } else {
+ se_mem_new->se_len = (se_mem->se_len -
+ *task_offset);
+
+ se_mem = list_entry(se_mem->se_list.next,
+ se_mem_t, se_list);
+ if (!(se_mem)) {
+ printk(KERN_ERR "Unable to locate next"
+ " se_mem_t\n");
+ return -1;
+ }
+ }
+
+ saved_task_offset = *task_offset;
+ *task_offset = 0;
+ }
+ task_size -= se_mem_new->se_len;
+next:
+ list_add_tail(&se_mem_new->se_list, se_mem_list);
+ (*se_mem_cnt)++;
+
+ DEBUG_MEM2("task[%u] - se_mem_cnt(%u) se_page(%p) se_off(%u)"
+ " se_len(%u)\n", task->task_no, *se_mem_cnt,
+ se_mem_new->se_page, se_mem_new->se_off,
+ se_mem->se_len);
+ DEBUG_MEM2("task[%u] - Reducing task_size to(%u)\n",
+ task->task_no, task_size);
+ }
+ *out_se_mem = se_mem;
+
+ return 0;
+}
+
+/* transport_map_mem_to_sg():
+ *
+ *
+ */
+int transport_map_mem_to_sg(
+ se_task_t *task,
+ struct list_head *se_mem_list,
+ void *in_mem,
+ se_mem_t *in_se_mem,
+ se_mem_t **out_se_mem,
+ u32 *se_mem_cnt,
+ u32 *task_offset)
+{
+ se_mem_t *se_mem = in_se_mem;
+ struct scatterlist *sg = (struct scatterlist *)in_mem;
+ u32 saved_task_offset = 0, sg_no = 0;
+ u32 task_size = task->task_size;
+
+ if (!sg) {
+ printk(KERN_ERR "Unable to locate valid struct"
+ " scatterlist pointer\n");
+ return -1;
+ }
+
+ while (task_size) {
+ if (*task_offset == 0) {
+ sg_assign_page(&sg[sg_no], se_mem->se_page);
+ sg[sg_no].offset = se_mem->se_off;
+
+ if (task_size >= se_mem->se_len) {
+ sg[sg_no].length = se_mem->se_len;
+
+ se_mem = list_entry(se_mem->se_list.next,
+ se_mem_t, se_list);
+ if (!(se_mem)) {
+ printk(KERN_ERR "Unable to locate"
+ " next se_mem_t\n");
+ return -1;
+ }
+ (*se_mem_cnt)++;
+ } else {
+ sg[sg_no].length = task_size;
+ /*
+ * Determine if we need to calculate an offset
+ * into the se_mem_t on the next go around..
+ */
+ task_size -= sg[sg_no].length;
+ if (!(task_size)) {
+ *task_offset = (sg[sg_no].length +
+ saved_task_offset);
+ goto next;
+ }
+ }
+
+ if (saved_task_offset)
+ *task_offset = saved_task_offset;
+ } else {
+ sg_assign_page(&sg[sg_no], se_mem->se_page);
+ sg[sg_no].offset = (*task_offset + se_mem->se_off);
+
+ if ((se_mem->se_len - *task_offset) > task_size) {
+ sg[sg_no].length = task_size;
+ /*
+ * Determine if we need to calculate an offset
+ * into the se_mem_t on the next go around..
+ */
+ task_size -= sg[sg_no].length;
+ if (!(task_size)) {
+ *task_offset += sg[sg_no].length;
+ goto next;
+ }
+ } else {
+ sg[sg_no].length = (se_mem->se_len -
+ *task_offset);
+
+ se_mem = list_entry(se_mem->se_list.next,
+ se_mem_t, se_list);
+ if (!(se_mem)) {
+ printk(KERN_ERR "Unable to locate"
+ " next se_mem_t\n");
+ return -1;
+ }
+ (*se_mem_cnt)++;
+ }
+
+ saved_task_offset = *task_offset;
+ *task_offset = 0;
+ }
+ task_size -= sg[sg_no].length;
+next:
+ DEBUG_MEM("task[%u] - sg[%u](%p)(%u)(%u) - Reducing task_size"
+ " to(%u)\n", task->task_no, sg_no,
+ sg_page(&sg[sg_no]), sg[sg_no].length,
+ sg[sg_no].offset, task_size);
+
+ sg_no++;
+ }
+ *out_se_mem = se_mem;
+ task->task_sg_num = sg_no;
+
+ DEBUG_MEM("task[%u] - Mapped(%u) se_mem_t segments to total(%u) SGs"
+ " saved task_offset(%u)\n", task->task_no, *se_mem_cnt,
+ sg_no, *task_offset);
+
+ return 0;
+}
+
+u32 transport_generic_get_cdb_count(
+ se_cmd_t *cmd,
+ se_transform_info_t *ti,
+ se_obj_lun_type_t *head_obj_api,
+ void *head_obj_ptr,
+ unsigned long long starting_lba,
+ u32 sectors,
+ se_mem_t *se_mem_in,
+ se_mem_t **se_mem_out,
+ u32 *task_offset_in)
+{
+ unsigned char *cdb = NULL;
+ void *obj_ptr, *next_obj_ptr = NULL;
+ se_task_t *task;
+ se_mem_t *se_mem, *se_mem_lout = NULL;
+ se_obj_lun_type_t *obj_api;
+ int max_sectors_set = 0, ret;
+ u32 se_mem_cnt = 0, task_cdbs = 0;
+ unsigned long long lba;
+
+ if (!se_mem_in) {
+ list_for_each_entry(se_mem_in, T_TASK(cmd)->t_mem_list, se_list)
+ break;
+
+ if (!se_mem_in) {
+ printk(KERN_ERR "se_mem_in is NULL\n");
+ return 0;
+ }
+ }
+ se_mem = se_mem_in;
+
+ /*
+ * Locate the start volume segment in which the received LBA will be
+ * executed upon.
+ */
+ head_obj_api->obtain_obj_lock(head_obj_ptr);
+ if (head_obj_api->obj_start(head_obj_ptr, ti, starting_lba) < 0) {
+ head_obj_api->release_obj_lock(head_obj_ptr);
+ return 0;
+ }
+
+ /*
+ * Locate starting object from original starting_lba.
+ */
+ lba = ti->ti_lba;
+ obj_api = ti->ti_obj_api;
+ obj_ptr = ti->ti_obj_ptr;
+ DEBUG_VOL("Starting Physical LBA(%llu) for head_obj_api->(%p)\n",
+ lba, head_obj_api);
+
+ while (sectors) {
+ if (!obj_api) {
+ head_obj_api->release_obj_lock(head_obj_ptr);
+ printk(KERN_ERR "obj_api is NULL LBA(%llu)->Sectors"
+ "(%u)\n", lba, sectors);
+ return 0;
+ }
+
+ DEBUG_VOL("ITT[0x%08x] LBA(%llu) SectorsLeft(%u) EOBJ(%llu)\n",
+ CMD_TFO(cmd)->get_task_tag(cmd), lba, sectors,
+ obj_api->end_lba(obj_ptr, 1));
+
+ head_obj_api->release_obj_lock(head_obj_ptr);
+
+ task = cmd->transport_get_task(ti, cmd, obj_ptr, obj_api);
+ if (!(task))
+ goto out;
+
+ transport_set_task_sectors(task, obj_api, obj_ptr, lba,
+ sectors, &max_sectors_set);
+
+ task->task_lba = lba;
+ lba += task->task_sectors;
+ sectors -= task->task_sectors;
+ task->task_size = (task->task_sectors *
+ obj_api->blocksize(obj_ptr));
+ task->transport_map_task =
+ obj_api->get_map_SG(obj_ptr, cmd->data_direction);
+
+ if ((cdb = obj_api->get_cdb(obj_ptr, task))) {
+ memcpy(cdb, T_TASK(cmd)->t_task_cdb, SCSI_CDB_SIZE);
+ cmd->transport_split_cdb(task->task_lba,
+ &task->task_sectors, cdb);
+ }
+
+ /*
+ * Perform the SE OBJ plugin and/or Transport plugin specific
+ * mapping for T_TASK(cmd)->t_mem_list.
+ */
+ ret = obj_api->do_se_mem_map(obj_ptr, task,
+ T_TASK(cmd)->t_mem_list, NULL, se_mem,
+ &se_mem_lout, &se_mem_cnt, task_offset_in);
+ if (ret < 0)
+ goto out;
+
+ head_obj_api->obtain_obj_lock(head_obj_ptr);
+
+ se_mem = se_mem_lout;
+ *se_mem_out = se_mem_lout;
+ task_cdbs++;
+
+ DEBUG_VOL("Incremented task_cdbs(%u) task->task_sg_num(%u)\n",
+ task_cdbs, task->task_sg_num);
+
+ if (max_sectors_set) {
+ max_sectors_set = 0;
+ continue;
+ }
+
+ if (!sectors)
+ break;
+
+ obj_api = obj_api->get_next_obj_api(obj_ptr, &next_obj_ptr);
+ if (obj_api) {
+ obj_ptr = next_obj_ptr;
+ lba = obj_api->get_next_lba(obj_ptr, lba);
+ }
+ }
+ head_obj_api->release_obj_lock(head_obj_ptr);
+
+ if (ti->ti_set_counts) {
+ atomic_inc(&T_TASK(cmd)->t_fe_count);
+ atomic_inc(&T_TASK(cmd)->t_se_count);
+ }
+
+ DEBUG_VOL("ITT[0x%08x] total cdbs(%u)\n",
+ CMD_TFO(cmd)->get_task_tag(cmd), task_cdbs);
+
+ return task_cdbs;
+out:
+ return 0;
+}
+
+/* transport_generic_new_cmd(): Called from transport_processing_thread()
+ *
+ * Allocate storage transport resources from a set of values predefined
+ * by transport_generic_cmd_sequencer() from the iSCSI Target RX process.
+ * Any non zero return here is treated as an "out of resource' op here.
+ */
+int transport_generic_new_cmd(se_cmd_t *cmd)
+{
+ int ret = 0;
+ se_transform_info_t ti;
+ /*
+ * Generate se_task_t(s) and/or their payloads for this CDB.
+ */
+ memset((void *)&ti, 0, sizeof(se_transform_info_t));
+ ti.ti_se_cmd = cmd;
+ ti.se_obj_ptr = SE_LUN(cmd)->lun_type_ptr;
+ ti.se_obj_api = SE_LUN(cmd)->lun_obj_api;
+
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)) {
+/* #warning FIXME v2.8: Get rid of PAGE_SIZE usage */
+ ret = cmd->transport_allocate_resources(cmd, cmd->data_length,
+ PAGE_SIZE);
+ if (ret < 0)
+ goto failure;
+
+ ret = transport_get_sectors(cmd, SE_LUN(cmd)->lun_obj_api,
+ SE_LUN(cmd)->lun_type_ptr);
+ if (ret < 0)
+ goto failure;
+
+ ret = transport_new_cmd_obj(cmd, &ti, SE_LUN(cmd)->lun_obj_api,
+ SE_LUN(cmd)->lun_type_ptr, 0);
+ if (ret < 0)
+ goto failure;
+
+ /*
+ * Allocate iovecs for frontend mappings. This currently
+ * assumes traditional iSCSI going to sockets.
+ *
+ * FIXME: This should be specific to frontend protocol/hardware.
+ */
+ ret = cmd->transport_allocate_iovecs(cmd);
+ if (ret < 0) {
+ ret = PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
+ goto failure;
+ }
+ }
+ /*
+ * This is dependent upon the storage processing algorithm.
+ */
+ if (cmd->transport_do_transform(cmd, &ti) < 0) {
+ ret = PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
+ goto failure;
+ }
+ /*
+ * Set the correct (usually DMAable) buffer pointers from the master
+ * buffer list in se_cmd_t to the transport task's native
+ * buffers format.
+ */
+ ret = cmd->transport_map_buffers_to_tasks(cmd);
+ if (ret < 0)
+ goto failure;
+ /*
+ * For WRITEs, let the iSCSI Target RX Thread know its buffer is ready..
+ * This WRITE se_cmd_t (and all of its associated se_task_t's)
+ * will be added to the se_device_t execution queue after its WRITE
+ * data has arrived. (ie: It gets handled by the transport processing
+ * thread a second time)
+ */
+ if (cmd->data_direction == SE_DIRECTION_WRITE) {
+ transport_add_tasks_to_state_queue(cmd);
+ return transport_generic_write_pending(cmd);
+ }
+ /*
+ * Everything else but a WRITE, add the se_cmd_t's se_task_t's
+ * to the execution queue.
+ */
+ transport_execute_tasks(cmd);
+ return 0;
+
+failure:
+ return ret;
+}
+
+/* transport_generic_process_write():
+ *
+ *
+ */
+void transport_generic_process_write(se_cmd_t *cmd)
+{
+#if 0
+ /*
+ * Copy SCSI Presented DTL sector(s) from received buffers allocated to
+ * original EDTL
+ */
+ if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT) {
+ if (!T_TASK(cmd)->t_task_se_num) {
+ unsigned char *dst, *buf =
+ (unsigned char *)T_TASK(cmd)->t_task_buf;
+
+ dst = kzalloc(cmd->cmd_spdtl), GFP_KERNEL);
+ if (!(dst)) {
+ printk(KERN_ERR "Unable to allocate memory for"
+ " WRITE underflow\n");
+ transport_generic_request_failure(cmd, NULL,
+ PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
+ return;
+ }
+ memcpy(dst, buf, cmd->cmd_spdtl);
+
+ kfree(T_TASK(cmd)->t_task_buf);
+ T_TASK(cmd)->t_task_buf = dst;
+ } else {
+ struct scatterlist *sg =
+ (struct scatterlist *sg)T_TASK(cmd)->t_task_buf;
+ struct scatterlist *orig_sg;
+
+ orig_sg = kzalloc(sizeof(struct scatterlist) *
+ T_TASK(cmd)->t_task_se_num,
+ GFP_KERNEL))) {
+ if (!(orig_sg)) {
+ printk(KERN_ERR "Unable to allocate memory"
+ " for WRITE underflow\n");
+ transport_generic_request_failure(cmd, NULL,
+ PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
+ return;
+ }
+
+ memcpy(orig_sg, T_TASK(cmd)->t_task_buf,
+ sizeof(struct scatterlist) *
+ T_TASK(cmd)->t_task_se_num);
+
+ cmd->data_length = cmd->cmd_spdtl;
+ /*
+ * FIXME, clear out original se_task_t and state
+ * information.
+ */
+ if (transport_generic_new_cmd(cmd) < 0) {
+ transport_generic_request_failure(cmd, NULL,
+ PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
+ kfree(orig_sg);
+ return;
+ }
+
+ transport_memcpy_write_sg(cmd, orig_sg);
+ }
+ }
+#endif
+ transport_execute_tasks(cmd);
+}
+
+/* transport_generic_write_pending():
+ *
+ *
+ */
+static int transport_generic_write_pending(se_cmd_t *cmd)
+{
+ unsigned long flags;
+ int ret;
+
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH) {
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC)) {
+ up(&T_TASK(cmd)->t_transport_passthrough_wsem);
+ transport_cmd_check_stop(cmd, 1, 0);
+ return PYX_TRANSPORT_WRITE_PENDING;
+ }
+
+ transport_generic_process_write(cmd);
+ transport_cmd_check_stop(cmd, 1, 0);
+ return PYX_TRANSPORT_WRITE_PENDING;
+ }
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ cmd->t_state = TRANSPORT_WRITE_PENDING;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ /*
+ * Call the fabric write_pending function here to let the
+ * frontend know that WRITE buffers are ready.
+ */
+ ret = CMD_TFO(cmd)->write_pending(cmd);
+ if (ret < 0) {
+ transport_cmd_check_stop(cmd, 1, 0);
+ return ret;
+ }
+
+ transport_cmd_check_stop(cmd, 1, 0);
+ return PYX_TRANSPORT_WRITE_PENDING;
+}
+
+/* transport_release_cmd_to_pool():
+ *
+ *
+ */
+void transport_release_cmd_to_pool(se_cmd_t *cmd)
+{
+ if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH)
+ kfree(cmd->se_lun);
+ /*
+ * Release se_cmd_t->se_fabric_cmd_ptr in fabric
+ */
+ CMD_TFO(cmd)->release_cmd_to_pool(cmd);
+
+ transport_free_se_cmd(cmd);
+}
+EXPORT_SYMBOL(transport_release_cmd_to_pool);
+
+/* transport_generic_free_cmd():
+ *
+ * Called from processing frontend to release storage engine resources
+ */
+void transport_generic_free_cmd(
+ se_cmd_t *cmd,
+ int wait_for_tasks,
+ int release_to_pool,
+ int session_reinstatement)
+{
+ if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD) || !T_TASK(cmd))
+ transport_release_cmd_to_pool(cmd);
+ else {
+ core_dec_lacl_count(cmd->se_sess->se_node_acl, cmd);
+
+ if (SE_LUN(cmd)) {
+#if 0
+ printk(KERN_INFO "cmd: %p ITT: 0x%08x contains"
+ " SE_LUN(cmd)\n", cmd,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+#endif
+ transport_lun_remove_cmd(cmd);
+ }
+
+ if (wait_for_tasks && cmd->transport_wait_for_tasks)
+ cmd->transport_wait_for_tasks(cmd, 0, 0);
+
+ transport_generic_remove(cmd, release_to_pool,
+ session_reinstatement);
+ }
+}
+EXPORT_SYMBOL(transport_generic_free_cmd);
+
+static void transport_nop_wait_for_tasks(
+ se_cmd_t *cmd,
+ int remove_cmd,
+ int session_reinstatement)
+{
+ return;
+}
+
+/* transport_lun_wait_for_tasks():
+ *
+ * Called from ConfigFS context to stop the passed se_cmd_t to allow
+ * an se_lun_t to be successfully shutdown.
+ */
+int transport_lun_wait_for_tasks(se_cmd_t *cmd, se_lun_t *lun)
+{
+ unsigned long flags;
+ /*
+ * If the frontend has already requested this se_cmd_t to
+ * be stopped, we can safely ignore this se_cmd_t.
+ */
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (atomic_read(&T_TASK(cmd)->t_transport_stop)) {
+ atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+ DEBUG_TRANSPORT_S("ConfigFS ITT[0x%08x] - t_transport_stop =="
+ " TRUE, skipping\n", CMD_TFO(cmd)->get_task_tag(cmd));
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return -1;
+ }
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ CMD_ORIG_OBJ_API(cmd)->notify_obj(cmd->se_orig_obj_ptr);
+
+ DEBUG_TRANSPORT_S("ConfigFS: ITT[0x%08x] - stopping cmd....\n",
+ CMD_TFO(cmd)->get_task_tag(cmd));
+ down(&T_TASK(cmd)->transport_lun_stop_sem);
+ DEBUG_TRANSPORT_S("ConfigFS: ITT[0x%08x] - stopped cmd....\n",
+ CMD_TFO(cmd)->get_task_tag(cmd));
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ return 0;
+}
+EXPORT_SYMBOL(transport_lun_wait_for_tasks);
+
+/* #define DEBUG_CLEAR_LUN */
+#ifdef DEBUG_CLEAR_LUN
+#define DEBUG_CLEAR_L(x...) PYXPRINT(x)
+#else
+#define DEBUG_CLEAR_L(x...)
+#endif
+
+void transport_clear_lun_from_sessions(se_lun_t *lun)
+{
+ se_cmd_t *cmd = NULL, *cmd_p = NULL;
+ unsigned long flags;
+ /*
+ * Do exception processing and return CHECK_CONDITION status to the
+ * Initiator Port.
+ */
+ spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+ list_for_each_entry_safe(cmd, cmd_p, &lun->lun_cmd_list, se_lun_list) {
+ if (!(T_TASK(cmd))) {
+ printk(KERN_ERR "ITT: 0x%08x, T_TASK(cmd) = NULL"
+ "[i,t]_state: %u/%u\n",
+ CMD_TFO(cmd)->get_task_tag(cmd),
+ CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state);
+ BUG();
+ }
+ list_del(&cmd->se_lun_list);
+ atomic_set(&T_TASK(cmd)->transport_lun_active, 0);
+ /*
+ * This will notify iscsi_target_transport.c:
+ * transport_cmd_check_stop() that a LUN shutdown is in
+ * progress for the iscsi_cmd_t.
+ */
+ spin_lock(&T_TASK(cmd)->t_state_lock);
+ DEBUG_CLEAR_L("SE_LUN[%d] - Setting T_TASK(cmd)->transport"
+ "_lun_stop for ITT: 0x%08x\n",
+ SE_LUN(cmd)->unpacked_lun,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+ atomic_set(&T_TASK(cmd)->transport_lun_stop, 1);
+ spin_unlock(&T_TASK(cmd)->t_state_lock);
+
+ spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
+
+ if (!(SE_LUN(cmd))) {
+ printk(KERN_ERR "ITT: 0x%08x, [i,t]_state: %u/%u\n",
+ CMD_TFO(cmd)->get_task_tag(cmd),
+ CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state);
+ BUG();
+ }
+ /*
+ * If the Storage engine still owns the iscsi_cmd_t, determine
+ * and/or stop its context.
+ */
+ DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x before transport"
+ "_lun_wait_for_tasks()\n", SE_LUN(cmd)->unpacked_lun,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+
+ if (transport_lun_wait_for_tasks(cmd, SE_LUN(cmd)) < 0) {
+ spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+ continue;
+ }
+
+ DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x after transport_lun"
+ "_wait_for_tasks(): SUCCESS\n",
+ SE_LUN(cmd)->unpacked_lun,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+ /*
+ * The Storage engine stopped this se_cmd_t before it was
+ * send to the fabric frontend for delivery back to the
+ * Initiator Node. Return this SCSI CDB back with an
+ * CHECK_CONDITION status.
+ */
+ transport_send_check_condition_and_sense(cmd,
+ NON_EXISTENT_LUN, 0);
+ /*
+ * If the fabric frontend is waiting for this iscsi_cmd_t to
+ * be released, notify the waiting thread now that LU has
+ * finished accessing it.
+ */
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (atomic_read(&T_TASK(cmd)->transport_lun_fe_stop)) {
+ DEBUG_CLEAR_L("SE_LUN[%d] - Detected FE stop for"
+ " se_cmd_t: %p ITT: 0x%08x\n",
+ lun->unpacked_lun,
+ cmd, CMD_TFO(cmd)->get_task_tag(cmd));
+
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+ flags);
+ up(&T_TASK(cmd)->transport_lun_fe_stop_sem);
+ spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+ continue;
+ }
+ atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+
+ DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x finished processing\n",
+ lun->unpacked_lun, CMD_TFO(cmd)->get_task_tag(cmd));
+
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ spin_lock_irqsave(&lun->lun_cmd_lock, flags);
+ }
+ spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
+}
+EXPORT_SYMBOL(transport_clear_lun_from_sessions);
+
+/* transport_generic_wait_for_tasks():
+ *
+ * Called from frontend or passthrough context to wait for storage engine
+ * to pause and/or release frontend generated se_cmd_t.
+ */
+static void transport_generic_wait_for_tasks(
+ se_cmd_t *cmd,
+ int remove_cmd,
+ int session_reinstatement)
+{
+ unsigned long flags;
+
+ if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD))
+ return;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ /*
+ * If we are already stopped due to an external event (ie: LUN shutdown)
+ * sleep until the connection can have the passed se_cmd_t back.
+ * The T_TASK(cmd)->transport_lun_stopped_sem will be upped by
+ * transport_clear_lun_from_sessions() once the ConfigFS context caller
+ * has completed its operation on the se_cmd_t.
+ */
+ if (atomic_read(&T_TASK(cmd)->transport_lun_stop)) {
+ atomic_set(&T_TASK(cmd)->transport_lun_fe_stop, 1);
+
+ DEBUG_TRANSPORT_S("wait_for_tasks: Stopping"
+ " down(&T_TASK(cmd)transport_lun_fe_stop_sem);"
+ " for ITT: 0x%08x\n", CMD_TFO(cmd)->get_task_tag(cmd));
+ /*
+ * There is a special case for WRITES where a FE exception +
+ * LUN shutdown means ConfigFS context is still sleeping on
+ * transport_lun_stop_sem in transport_lun_wait_for_tasks().
+ * We go ahead and up transport_lun_stop_sem just to be sure
+ * here.
+ */
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ up(&T_TASK(cmd)->transport_lun_stop_sem);
+ down(&T_TASK(cmd)->transport_lun_fe_stop_sem);
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ /*
+ * At this point, the frontend who was the originator of this
+ * se_cmd_t, now owns the structure and can be released through
+ * normal means below.
+ */
+ DEBUG_TRANSPORT_S("wait_for_tasks: Stopped"
+ " down(&T_TASK(cmd)transport_lun_fe_stop_sem);"
+ " for ITT: 0x%08x\n", CMD_TFO(cmd)->get_task_tag(cmd));
+
+ atomic_set(&T_TASK(cmd)->transport_lun_fe_stop, 0);
+ atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+ }
+ if (!atomic_read(&T_TASK(cmd)->t_transport_active))
+ goto remove;
+
+ atomic_set(&T_TASK(cmd)->t_transport_stop, 1);
+
+ DEBUG_TRANSPORT_S("wait_for_tasks: Stopping %p ITT/CmdSN: 0x%08x/"
+ "0x%08x, i_state/def_i_state: %d/%d, t_state/def_t_state:"
+ " %d/%d, t_transport_stop = TRUE\n", cmd,
+ CMD_TFO(cmd)->get_task_tag(cmd), cmd->cmd_sn,
+ CMD_TFO(cmd)->get_cmd_state(cmd), cmd->deferred_i_state,
+ cmd->t_state, cmd->deferred_t_state);
+
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ CMD_ORIG_OBJ_API(cmd)->notify_obj(cmd->se_orig_obj_ptr);
+
+ down(&T_TASK(cmd)->t_transport_stop_sem);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+ atomic_set(&T_TASK(cmd)->t_transport_stop, 0);
+
+ DEBUG_TRANSPORT_S("wait_for_tasks: Stopped down(&T_TASK(cmd)->"
+ "t_transport_stop_sem) for ITT: 0x%08x\n",
+ CMD_TFO(cmd)->get_task_tag(cmd));
+remove:
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ if (!remove_cmd)
+ return;
+
+ transport_generic_free_cmd(cmd, 0, 0, session_reinstatement);
+}
+
+int transport_send_check_condition_and_sense(
+ se_cmd_t *cmd,
+ u8 reason,
+ int from_transport)
+{
+ unsigned char *buffer = cmd->sense_buffer;
+ unsigned long flags;
+ int offset;
+ u8 asc = 0, ascq = 0;
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ if (cmd->se_cmd_flags & SCF_SENT_CHECK_CONDITION) {
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+ return 0;
+ }
+ cmd->se_cmd_flags |= SCF_SENT_CHECK_CONDITION;
+ spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+
+ if (!reason && from_transport)
+ goto after_reason;
+
+ if (!from_transport)
+ cmd->se_cmd_flags |= SCF_EMULATED_TASK_SENSE;
+ /*
+ * Data Segment and SenseLength of the fabric response PDU.
+ *
+ * TRANSPORT_SENSE_BUFFER is now set to SCSI_SENSE_BUFFERSIZE
+ * from include/scsi/scsi_cmnd.h
+ */
+ offset = CMD_TFO(cmd)->set_fabric_sense_len(cmd,
+ TRANSPORT_SENSE_BUFFER);
+ /*
+ * Actual SENSE DATA, see SPC-3 7.23.2 SPC_SENSE_KEY_OFFSET uses
+ * SENSE KEY values from include/scsi/scsi.h
+ */
+ switch (reason) {
+ case NON_EXISTENT_LUN:
+ case UNSUPPORTED_SCSI_OPCODE:
+ case SECTOR_COUNT_TOO_MANY:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ILLEGAL REQUEST */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ILLEGAL_REQUEST;
+ /* INVALID COMMAND OPERATION CODE */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x20;
+ break;
+ case UNKNOWN_MODE_PAGE:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ILLEGAL REQUEST */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ILLEGAL_REQUEST;
+ /* INVALID FIELD IN CDB */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x24;
+ break;
+ case CHECK_CONDITION_ABORT_CMD:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* BUS DEVICE RESET FUNCTION OCCURRED */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x29;
+ buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x03;
+ break;
+ case INCORRECT_AMOUNT_OF_DATA:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* WRITE ERROR */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x0c;
+ /* NOT ENOUGH UNSOLICITED DATA */
+ buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x0d;
+ break;
+ case INVALID_CDB_FIELD:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* INVALID FIELD IN CDB */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x24;
+ break;
+ case INVALID_PARAMETER_LIST:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* INVALID FIELD IN PARAMETER LIST */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x26;
+ break;
+ case UNEXPECTED_UNSOLICITED_DATA:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* WRITE ERROR */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x0c;
+ /* UNEXPECTED_UNSOLICITED_DATA */
+ buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x0c;
+ break;
+ case SERVICE_CRC_ERROR:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* PROTOCOL SERVICE CRC ERROR */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x47;
+ /* N/A */
+ buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x05;
+ break;
+ case SNACK_REJECTED:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ABORTED COMMAND */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ABORTED_COMMAND;
+ /* READ ERROR */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x11;
+ /* FAILED RETRANSMISSION REQUEST */
+ buffer[offset+SPC_ASCQ_KEY_OFFSET] = 0x13;
+ break;
+ case WRITE_PROTECTED:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* DATA PROTECT */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = DATA_PROTECT;
+ /* WRITE PROTECTED */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x27;
+ break;
+ case CHECK_CONDITION_UNIT_ATTENTION:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* UNIT ATTENTION */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = UNIT_ATTENTION;
+ core_scsi3_ua_for_check_condition(cmd, &asc, &ascq);
+ buffer[offset+SPC_ASC_KEY_OFFSET] = asc;
+ buffer[offset+SPC_ASCQ_KEY_OFFSET] = ascq;
+ break;
+ case LOGICAL_UNIT_COMMUNICATION_FAILURE:
+ default:
+ /* CURRENT ERROR */
+ buffer[offset] = 0x70;
+ /* ILLEGAL REQUEST */
+ buffer[offset+SPC_SENSE_KEY_OFFSET] = ILLEGAL_REQUEST;
+ /* LOGICAL UNIT COMMUNICATION FAILURE */
+ buffer[offset+SPC_ASC_KEY_OFFSET] = 0x80;
+ break;
+ }
+ /*
+ * This code uses linux/include/scsi/scsi.h SAM status codes!
+ */
+ cmd->scsi_status = SAM_STAT_CHECK_CONDITION;
+ /*
+ * Automatically padded, this value is encoded in the fabric's
+ * data_length response PDU containing the SCSI defined sense data.
+ */
+ cmd->scsi_sense_length = TRANSPORT_SENSE_BUFFER + offset;
+
+after_reason:
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+ CMD_TFO(cmd)->queue_status(cmd);
+
+ return 0;
+}
+EXPORT_SYMBOL(transport_send_check_condition_and_sense);
+
+void transport_send_task_abort(se_cmd_t *cmd)
+{
+ cmd->scsi_status = SAM_STAT_TASK_ABORTED;
+#if 0
+ printk(KERN_INFO "Setting SAM_STAT_TASK_ABORTED status for CDB: 0x%02x,"
+ " ITT: 0x%08x\n", T_TASK(cmd)->t_task_cdb[0],
+ CMD_TFO(cmd)->get_task_tag(cmd));
+#endif
+ if (!(cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH))
+ CMD_TFO(cmd)->queue_status(cmd);
+}
+
+/* transport_generic_do_tmr():
+ *
+ *
+ */
+int transport_generic_do_tmr(se_cmd_t *cmd)
+{
+ se_cmd_t *ref_cmd;
+ se_device_t *dev = SE_DEV(cmd);
+ se_tmr_req_t *tmr = cmd->se_tmr_req;
+ int ret;
+
+ switch (tmr->function) {
+ case ABORT_TASK:
+ ref_cmd = tmr->ref_cmd;
+ tmr->response = TMR_FUNCTION_REJECTED;
+ break;
+ case ABORT_TASK_SET:
+ case CLEAR_ACA:
+ case CLEAR_TASK_SET:
+ tmr->response = TMR_TASK_MGMT_FUNCTION_NOT_SUPPORTED;
+ break;
+ case LUN_RESET:
+ ret = core_tmr_lun_reset(dev, tmr, NULL, NULL);
+ tmr->response = (!ret) ? TMR_FUNCTION_COMPLETE :
+ TMR_FUNCTION_REJECTED;
+ break;
+#if 0
+ case TARGET_WARM_RESET:
+ transport_generic_host_reset(dev->se_hba);
+ tmr->response = TMR_FUNCTION_REJECTED;
+ break;
+ case TARGET_COLD_RESET:
+ transport_generic_host_reset(dev->se_hba);
+ transport_generic_cold_reset(dev->se_hba);
+ tmr->response = TMR_FUNCTION_REJECTED;
+ break;
+#endif
+ default:
+ printk(KERN_ERR "Uknown TMR function: 0x%02x.\n",
+ tmr->function);
+ tmr->response = TMR_FUNCTION_REJECTED;
+ break;
+ }
+
+ cmd->t_state = TRANSPORT_ISTATE_PROCESSING;
+ CMD_TFO(cmd)->queue_tm_rsp(cmd);
+
+ return 0;
+}
+
+/*
+ * Called with spin_lock_irq(&dev->execute_task_lock); held
+ *
+ */
+se_task_t *transport_get_task_from_state_list(se_device_t *dev)
+{
+ se_task_t *task;
+
+ if (list_empty(&dev->state_task_list))
+ return NULL;
+
+ list_for_each_entry(task, &dev->state_task_list, t_state_list)
+ break;
+
+ list_del(&task->t_state_list);
+ atomic_set(&task->task_state_active, 0);
+
+ return task;
+}
+
+static void transport_processing_shutdown(se_device_t *dev)
+{
+ se_cmd_t *cmd;
+ se_queue_req_t *qr;
+ se_task_t *task;
+ u8 state;
+ unsigned long flags;
+ /*
+ * Empty the se_device_t's se_task_t state list.
+ */
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ while ((task = transport_get_task_from_state_list(dev))) {
+ if (!(TASK_CMD(task))) {
+ printk(KERN_ERR "TASK_CMD(task) is NULL!\n");
+ continue;
+ }
+ cmd = TASK_CMD(task);
+
+ if (!T_TASK(cmd)) {
+ printk(KERN_ERR "T_TASK(cmd) is NULL for task: %p cmd:"
+ " %p ITT: 0x%08x\n", task, cmd,
+ CMD_TFO(cmd)->get_task_tag(cmd));
+ continue;
+ }
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+
+ DEBUG_DO("PT: cmd: %p task: %p ITT/CmdSN: 0x%08x/0x%08x,"
+ " i_state/def_i_state: %d/%d, t_state/def_t_state:"
+ " %d/%d cdb: 0x%02x\n", cmd, task,
+ CMD_TFO(cmd)->get_task_tag(cmd), cmd->cmd_sn,
+ CMD_TFO(cmd)->get_cmd_state(cmd), cmd->deferred_i_state,
+ cmd->t_state, cmd->deferred_t_state,
+ T_TASK(cmd)->t_task_cdb[0]);
+ DEBUG_DO("PT: ITT[0x%08x] - t_task_cdbs: %d t_task_cdbs_left:"
+ " %d t_task_cdbs_sent: %d -- t_transport_active: %d"
+ " t_transport_stop: %d t_transport_sent: %d\n",
+ CMD_TFO(cmd)->get_task_tag(cmd),
+ T_TASK(cmd)->t_task_cdbs,
+ atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
+ atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
+ atomic_read(&T_TASK(cmd)->t_transport_active),
+ atomic_read(&T_TASK(cmd)->t_transport_stop),
+ atomic_read(&T_TASK(cmd)->t_transport_sent));
+
+ if (atomic_read(&task->task_active)) {
+ atomic_set(&task->task_stop, 1);
+ spin_unlock_irqrestore(
+ &T_TASK(cmd)->t_state_lock, flags);
+
+ DEBUG_DO("Waiting for task: %p to shutdown for dev:"
+ " %p\n", task, dev);
+ down(&task->task_stop_sem);
+ DEBUG_DO("Completed task: %p shutdown for dev: %p\n",
+ task, dev);
+
+ spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+ atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+
+ atomic_set(&task->task_active, 0);
+ atomic_set(&task->task_stop, 0);
+ }
+ __transport_stop_task_timer(task, &flags);
+
+ if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_ex_left))) {
+ spin_unlock_irqrestore(
+ &T_TASK(cmd)->t_state_lock, flags);
+
+ DEBUG_DO("Skipping task: %p, dev: %p for"
+ " t_task_cdbs_ex_left: %d\n", task, dev,
+ atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left));
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ continue;
+ }
+
+ if (atomic_read(&T_TASK(cmd)->t_transport_active)) {
+ DEBUG_DO("got t_transport_active = 1 for task: %p, dev:"
+ " %p\n", task, dev);
+
+ if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+ spin_unlock_irqrestore(
+ &T_TASK(cmd)->t_state_lock, flags);
+ transport_send_check_condition_and_sense(
+ cmd, LOGICAL_UNIT_COMMUNICATION_FAILURE,
+ 0);
+ transport_remove_cmd_from_queue(cmd,
+ CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+ cmd->se_orig_obj_ptr));
+
+ transport_lun_remove_cmd(cmd);
+ if (!(transport_cmd_check_stop(cmd, 1, 0)))
+ transport_passthrough_check_stop(cmd);
+ } else {
+ spin_unlock_irqrestore(
+ &T_TASK(cmd)->t_state_lock, flags);
+
+ transport_remove_cmd_from_queue(cmd,
+ CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+ cmd->se_orig_obj_ptr));
+
+ transport_lun_remove_cmd(cmd);
+
+ if (!(transport_cmd_check_stop(cmd, 1, 0)))
+ transport_passthrough_check_stop(cmd);
+ else
+ transport_generic_remove(cmd, 0, 0);
+ }
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ continue;
+ }
+ DEBUG_DO("Got t_transport_active = 0 for task: %p, dev: %p\n",
+ task, dev);
+
+ if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+ spin_unlock_irqrestore(
+ &T_TASK(cmd)->t_state_lock, flags);
+ transport_send_check_condition_and_sense(cmd,
+ LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
+ transport_remove_cmd_from_queue(cmd,
+ CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+ cmd->se_orig_obj_ptr));
+
+ transport_lun_remove_cmd(cmd);
+ if (!(transport_cmd_check_stop(cmd, 1, 0)))
+ transport_passthrough_check_stop(cmd);
+ } else {
+ spin_unlock_irqrestore(
+ &T_TASK(cmd)->t_state_lock, flags);
+
+ transport_remove_cmd_from_queue(cmd,
+ CMD_ORIG_OBJ_API(cmd)->get_queue_obj(
+ cmd->se_orig_obj_ptr));
+ transport_lun_remove_cmd(cmd);
+
+ if (!(transport_cmd_check_stop(cmd, 1, 0)))
+ transport_passthrough_check_stop(cmd);
+ else
+ transport_generic_remove(cmd, 0, 0);
+ }
+
+ spin_lock_irqsave(&dev->execute_task_lock, flags);
+ }
+ spin_unlock_irqrestore(&dev->execute_task_lock, flags);
+ /*
+ * Empty the se_device_t's se_cmd_t list.
+ */
+ spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+ while ((qr = __transport_get_qr_from_queue(dev->dev_queue_obj))) {
+ spin_unlock_irqrestore(
+ &dev->dev_queue_obj->cmd_queue_lock, flags);
+ cmd = (se_cmd_t *)qr->cmd;
+ state = qr->state;
+ kfree(qr);
+
+ DEBUG_DO("From Device Queue: cmd: %p t_state: %d\n",
+ cmd, state);
+
+ if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+ transport_send_check_condition_and_sense(cmd,
+ LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
+
+ transport_lun_remove_cmd(cmd);
+ if (!(transport_cmd_check_stop(cmd, 1, 0)))
+ transport_passthrough_check_stop(cmd);
+ } else {
+ transport_lun_remove_cmd(cmd);
+
+ if (!(transport_cmd_check_stop(cmd, 1, 0)))
+ transport_passthrough_check_stop(cmd);
+ else
+ transport_generic_remove(cmd, 0, 0);
+ }
+ spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+ }
+ spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock, flags);
+}
+
+/* transport_processing_thread():
+ *
+ *
+ */
+static int transport_processing_thread(void *param)
+{
+ int ret, t_state;
+ se_cmd_t *cmd;
+ se_device_t *dev = (se_device_t *) param;
+ se_queue_req_t *qr;
+
+ current->policy = SCHED_NORMAL;
+ set_user_nice(current, -20);
+ spin_lock_irq(¤t->sighand->siglock);
+ siginitsetinv(¤t->blocked, SHUTDOWN_SIGS);
+ recalc_sigpending();
+ spin_unlock_irq(¤t->sighand->siglock);
+
+ up(&dev->dev_queue_obj->thread_create_sem);
+
+ while (!(kthread_should_stop())) {
+ ret = wait_event_interruptible(dev->dev_queue_obj->thread_wq,
+ atomic_read(&dev->dev_queue_obj->queue_cnt) ||
+ kthread_should_stop());
+ if (ret < 0)
+ goto out;
+
+ spin_lock(&dev->dev_status_lock);
+ if (dev->dev_status & TRANSPORT_DEVICE_SHUTDOWN) {
+ spin_unlock(&dev->dev_status_lock);
+ transport_processing_shutdown(dev);
+ continue;
+ }
+ spin_unlock(&dev->dev_status_lock);
+
+get_cmd:
+ __transport_execute_tasks(dev);
+
+ qr = transport_get_qr_from_queue(dev->dev_queue_obj);
+ if (!(qr))
+ continue;
+
+ cmd = (se_cmd_t *)qr->cmd;
+ t_state = qr->state;
+ kfree(qr);
+
+ switch (t_state) {
+ case TRANSPORT_NEW_CMD:
+ ret = transport_generic_new_cmd(cmd);
+ if (ret < 0) {
+ cmd->transport_error_status = ret;
+ transport_generic_request_failure(cmd, NULL,
+ 0, (cmd->data_direction !=
+ SE_DIRECTION_WRITE));
+ }
+ break;
+ case TRANSPORT_PROCESS_WRITE:
+ transport_generic_process_write(cmd);
+ break;
+ case TRANSPORT_COMPLETE_OK:
+ transport_stop_all_task_timers(cmd);
+ transport_generic_complete_ok(cmd);
+ break;
+ case TRANSPORT_REMOVE:
+ transport_generic_remove(cmd, 1, 0);
+ break;
+ case TRANSPORT_PROCESS_TMR:
+ transport_generic_do_tmr(cmd);
+ break;
+ case TRANSPORT_COMPLETE_FAILURE:
+ transport_generic_request_failure(cmd, NULL, 1, 1);
+ break;
+ case TRANSPORT_COMPLETE_TIMEOUT:
+ transport_stop_all_task_timers(cmd);
+ transport_generic_request_timeout(cmd);
+ break;
+ default:
+ printk(KERN_ERR "Unknown t_state: %d deferred_t_state:"
+ " %d for ITT: 0x%08x i_state: %d on SE LUN:"
+ " %u\n", t_state, cmd->deferred_t_state,
+ CMD_TFO(cmd)->get_task_tag(cmd),
+ CMD_TFO(cmd)->get_cmd_state(cmd),
+ SE_LUN(cmd)->unpacked_lun);
+ BUG();
+ }
+
+ goto get_cmd;
+ }
+
+out:
+ transport_release_all_cmds(dev);
+ dev->process_thread = NULL;
+ up(&dev->dev_queue_obj->thread_done_sem);
+ return 0;
+}
diff --git a/include/target/target_core_transport.h b/include/target/target_core_transport.h
new file mode 100644
index 0000000..5c586cd
--- /dev/null
+++ b/include/target/target_core_transport.h
@@ -0,0 +1,519 @@
+/*******************************************************************************
+ * Filename: target_core_transport.h
+ *
+ * This file contains the iSCSI Target Generic DAS Transport Layer definitions.
+ *
+ * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
+ * Copyright (c) 2005, 2006, 2007 SBE, Inc.
+ * Copyright (c) 2007-2009 Rising Tide Software, Inc.
+ * Copyright (c) 2008-2009 Linux-iSCSI.org
+ *
+ * Nicholas A. Bellinger <nab@...nel.org>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+ *
+ ******************************************************************************/
+
+
+#ifndef TARGET_CORE_TRANSPORT_H
+#define TARGET_CORE_TRANSPORT_H
+
+#define TARGET_CORE_VERSION "v3.0.0"
+
+/* Attempts before moving from SHORT to LONG */
+#define PYX_TRANSPORT_WINDOW_CLOSED_THRESHOLD 3
+#define PYX_TRANSPORT_WINDOW_CLOSED_WAIT_SHORT 3 /* In milliseconds */
+#define PYX_TRANSPORT_WINDOW_CLOSED_WAIT_LONG 10 /* In milliseconds */
+
+#define PYX_TRANSPORT_STATUS_INTERVAL 5 /* In seconds */
+
+#define PYX_TRANSPORT_SENT_TO_TRANSPORT 0
+#define PYX_TRANSPORT_WRITE_PENDING 1
+
+#define PYX_TRANSPORT_UNKNOWN_SAM_OPCODE -1
+#define PYX_TRANSPORT_HBA_QUEUE_FULL -2
+#define PYX_TRANSPORT_REQ_TOO_MANY_SECTORS -3
+#define PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES -4
+#define PYX_TRANSPORT_INVALID_CDB_FIELD -5
+#define PYX_TRANSPORT_INVALID_PARAMETER_LIST -6
+#define PYX_TRANSPORT_LOGICAL_UNIT_COMMUNICATION_FAILURE -7
+#define PYX_TRANSPORT_UNKNOWN_MODE_PAGE -8
+#define PYX_TRANSPORT_WRITE_PROTECTED -9
+#define PYX_TRANSPORT_TASK_TIMEOUT -10
+#define PYX_TRANSPORT_RESERVATION_CONFLICT -11
+
+#ifndef SAM_STAT_RESERVATION_CONFLICT
+#define SAM_STAT_RESERVATION_CONFLICT 0x18
+#endif
+
+#define TRANSPORT_PLUGIN_FREE 0
+#define TRANSPORT_PLUGIN_REGISTERED 1
+
+#define TRANSPORT_PLUGIN_PHBA_PDEV 1
+#define TRANSPORT_PLUGIN_VHBA_PDEV 2
+#define TRANSPORT_PLUGIN_VHBA_VDEV 3
+
+/* For SE OBJ Plugins, in seconds */
+#define TRANSPORT_TIMEOUT_TUR 10
+#define TRANSPORT_TIMEOUT_TYPE_DISK 60
+#define TRANSPORT_TIMEOUT_TYPE_ROM 120
+#define TRANSPORT_TIMEOUT_TYPE_TAPE 600
+#define TRANSPORT_TIMEOUT_TYPE_OTHER 300
+
+/* For se_task->task_state_flags */
+#define TSF_EXCEPTION_CLEARED 0x01
+
+/*
+ * se_subsystem_dev_t->su_dev_flags
+*/
+#define SDF_FIRMWARE_VPD_UNIT_SERIAL 0x00000001
+#define SDF_EMULATED_VPD_UNIT_SERIAL 0x00000002
+
+/*
+ * se_device_t->dev_flags
+ */
+#define DF_READAHEAD_ACTIVE 0x00000001
+#define DF_TRANSPORT_DMA_ALLOC 0x00000002
+#define DF_TRANSPORT_BUF_ALLOC 0x00000004
+#define DF_DEV_DEBUG 0x00000008
+#define DF_CLAIMED_BLOCKDEV 0x00000010
+#define DF_PERSISTENT_CLAIMED_BLOCKDEV 0x00000020
+#define DF_READ_ONLY 0x00000040
+#define DF_SPC3_PERSISTENT_RESERVE 0x00000080
+#define DF_SPC2_RESERVATIONS 0x00000100
+
+/* se_dev_attrib_t sanity values */
+/* 10 Minutes, see transport_get_default_task_timeout() */
+#define DA_TASK_TIMEOUT_MAX 600
+/* Emulation for UNIT ATTENTION Interlock Control */
+#define DA_EMULATE_UA_INTLLCK_CTRL 0
+/* Emulation for TASK_ABORTED status (TAS) by default */
+#define DA_EMULATE_TAS 1
+/* No Emulation for PSCSI by default */
+#define DA_EMULATE_RESERVATIONS 0
+/* No Emulation for PSCSI by default */
+#define DA_EMULATE_ALUA 0
+#define DA_STATUS_MAX_SECTORS_MIN 16
+#define DA_STATUS_MAX_SECTORS_MAX 8192
+
+#define SE_MODE_PAGE_BUF 512
+/* Must match PARAM_UDEV_PATH_LEN */
+#define SE_UDEV_PATH_LEN 256
+
+#define MOD_MAX_SECTORS(ms, bs) (ms % (PAGE_SIZE / bs))
+
+struct se_mem_s;
+
+extern int init_se_global(void);
+extern void release_se_global(void);
+#ifdef DEBUG_DEV
+extern int __iscsi_debug_dev(se_device_t *);
+#endif
+extern unsigned char *transport_get_iqn_sn(void);
+extern void transport_init_queue_obj(struct se_queue_obj_s *);
+extern void transport_load_plugins(void);
+extern struct se_plugin_s *transport_core_get_plugin_by_name(const char *name);
+extern void transport_check_dev_params_delim(char *, char **);
+extern struct se_session_s *transport_init_session(void);
+extern void __transport_register_session(struct se_portal_group_s *,
+ struct se_node_acl_s *,
+ struct se_session_s *, void *);
+extern void transport_register_session(struct se_portal_group_s *,
+ struct se_node_acl_s *,
+ struct se_session_s *, void *);
+extern void transport_free_session(struct se_session_s *);
+extern void transport_deregister_session_configfs(struct se_session_s *);
+extern void transport_deregister_session(struct se_session_s *);
+extern void transport_task_dev_remove_state(struct se_task_s *,
+ struct se_device_s *);
+extern void transport_cmd_finish_abort(struct se_cmd_s *, int);
+extern void transport_cmd_finish_abort_tmr(struct se_cmd_s *);
+extern int transport_add_cmd_to_queue(struct se_cmd_s *,
+ struct se_queue_obj_s *, u8);
+extern struct se_queue_req_s *__transport_get_qr_from_queue(
+ struct se_queue_obj_s *);
+extern void transport_remove_cmd_from_queue(struct se_cmd_s *,
+ struct se_queue_obj_s *);
+extern void transport_complete_cmd(se_cmd_t *, int);
+extern void transport_complete_task(struct se_task_s *, int);
+extern void transport_add_task_to_execute_queue(struct se_task_s *,
+ struct se_device_s *);
+extern void transport_add_tasks_from_cmd(struct se_cmd_s *);
+extern struct se_task_s *transport_get_task_from_execute_queue(
+ struct se_device_s *);
+extern se_queue_req_t *transport_get_qr_from_queue(struct se_queue_obj_s *);
+extern int transport_check_device_tcq(se_device_t *, u32, u32);
+unsigned char *transport_dump_cmd_direction (struct se_cmd_s *);
+extern void transport_dump_dev_state(struct se_device_s *, char *, int *);
+extern void transport_dump_dev_info(struct se_device_s *, struct se_lun_s *,
+ unsigned long long, char *, int *);
+extern void transport_dump_vpd_proto_id(struct t10_vpd_s *,
+ unsigned char *, int);
+extern int transport_dump_vpd_assoc(struct t10_vpd_s *,
+ unsigned char *, int);
+extern int transport_dump_vpd_ident_type(struct t10_vpd_s *,
+ unsigned char *, int);
+extern int transport_dump_vpd_ident(struct t10_vpd_s *,
+ unsigned char *, int);
+extern se_device_t *transport_add_device_to_core_hba(se_hba_t *,
+ struct se_subsystem_api_s *,
+ struct se_subsystem_dev_s *, u32,
+ void *);
+extern int transport_generic_activate_device(se_device_t *);
+extern void transport_generic_deactivate_device(se_device_t *);
+extern int transport_generic_claim_phydevice(se_device_t *);
+extern void transport_generic_release_phydevice(se_device_t *, int);
+extern void transport_generic_free_device(se_device_t *);
+extern int transport_allocate_iovecs_for_cmd(struct se_cmd_s *, u32);
+extern int transport_generic_obj_start(struct se_transform_info_s *,
+ struct se_obj_lun_type_s *, void *,
+ unsigned long long);
+extern void transport_device_setup_cmd(se_cmd_t *);
+extern se_cmd_t *transport_alloc_se_cmd(struct target_core_fabric_ops *,
+ struct se_session_s *, void *,
+ u32, int);
+extern void transport_free_se_cmd(struct se_cmd_s *);
+extern int transport_generic_allocate_tasks(se_cmd_t *, unsigned char *);
+extern int transport_generic_handle_cdb(se_cmd_t *);
+extern int transport_generic_handle_data(se_cmd_t *);
+extern int transport_generic_handle_tmr(se_cmd_t *);
+extern void transport_stop_tasks_for_cmd(struct se_cmd_s *);
+extern void transport_generic_request_failure(se_cmd_t *, se_device_t *,
+ int, int);
+extern void transport_direct_request_timeout(se_cmd_t *);
+extern void transport_generic_request_timeout(se_cmd_t *);
+extern int transport_generic_allocate_buf(se_cmd_t *, u32, u32);
+extern int __transport_execute_tasks(struct se_device_s *);
+extern void transport_new_cmd_failure(struct se_cmd_s *);
+extern u32 transport_get_default_task_timeout(struct se_device_s *);
+extern void transport_set_supported_SAM_opcode(struct se_cmd_s *);
+extern void transport_start_task_timer(struct se_task_s *);
+extern void __transport_stop_task_timer(struct se_task_s *, unsigned long *);
+extern void transport_stop_task_timer(struct se_task_s *);
+extern void transport_stop_all_task_timers(struct se_cmd_s *);
+extern int transport_execute_tasks(struct se_cmd_s *);
+extern int transport_generic_emulate_inquiry(struct se_cmd_s *, unsigned char,
+ unsigned char *, unsigned char *,
+ unsigned char *);
+extern int transport_generic_emulate_readcapacity(struct se_cmd_s *, u32);
+extern int transport_generic_emulate_readcapacity_16(struct se_cmd_s *,
+ unsigned long long);
+extern int transport_generic_emulate_modesense(struct se_cmd_s *,
+ unsigned char *,
+ unsigned char *, int, int);
+extern int transport_generic_emulate_request_sense(struct se_cmd_s *,
+ unsigned char *);
+extern int transport_get_sense_data(struct se_cmd_s *);
+extern se_cmd_t *transport_allocate_passthrough(unsigned char *, int, u32,
+ void *, u32, u32,
+ struct se_obj_lun_type_s *,
+ void *);
+extern void transport_passthrough_release(se_cmd_t *);
+extern int transport_passthrough_complete(se_cmd_t *);
+extern int transport_generic_passthrough_async(se_cmd_t *cmd,
+ void(*callback)(se_cmd_t *cmd,
+ void *callback_arg, int complete_status),
+ void *callback_arg);
+extern int transport_generic_passthrough(se_cmd_t *);
+extern void transport_generic_complete_ok(se_cmd_t *);
+extern void transport_free_dev_tasks(se_cmd_t *);
+extern void transport_release_tasks(se_cmd_t *);
+extern void transport_release_fe_cmd(se_cmd_t *);
+extern int transport_generic_remove(se_cmd_t *, int, int);
+extern int transport_lun_wait_for_tasks(se_cmd_t *, se_lun_t *);
+extern void transport_clear_lun_from_sessions(se_lun_t *);
+extern int transport_send_check_condition_and_sense(se_cmd_t *, u8, int);
+extern void transport_send_task_abort(struct se_cmd_s *);
+extern void transport_release_cmd_to_pool(se_cmd_t *);
+extern void transport_generic_free_cmd(se_cmd_t *, int, int, int);
+extern void transport_generic_wait_for_cmds(se_cmd_t *, int);
+extern int transport_generic_do_transform(struct se_cmd_s *,
+ struct se_transform_info_s *);
+extern int transport_get_sectors(struct se_cmd_s *, struct se_obj_lun_type_s *,
+ void *);
+extern int transport_new_cmd_obj(struct se_cmd_s *,
+ struct se_transform_info_s *,
+ struct se_obj_lun_type_s *, void *, int);
+extern unsigned char *transport_get_vaddr(struct se_mem_s *);
+extern struct list_head *transport_init_se_mem_list(void);
+extern void transport_free_se_mem_list(struct list_head *);
+extern int transport_generic_get_mem(struct se_cmd_s *, u32, u32);
+extern u32 transport_calc_sg_num(struct se_task_s *, struct se_mem_s *, u32);
+extern int transport_map_sg_to_mem(struct se_task_s *, struct list_head *,
+ void *, struct se_mem_s *,
+ struct se_mem_s **, u32 *, u32 *);
+extern int transport_map_mem_to_mem(struct se_task_s *, struct list_head *,
+ void *, struct se_mem_s *,
+ struct se_mem_s **, u32 *, u32 *);
+extern int transport_map_mem_to_sg(struct se_task_s *, struct list_head *,
+ void *, struct se_mem_s *,
+ struct se_mem_s **, u32 *, u32 *);
+extern u32 transport_generic_get_cdb_count(struct se_cmd_s *,
+ struct se_transform_info_s *,
+ struct se_obj_lun_type_s *, void *,
+ unsigned long long, u32,
+ struct se_mem_s *, struct se_mem_s **,
+ u32 *);
+extern int transport_generic_new_cmd(se_cmd_t *);
+extern void transport_generic_process_write(se_cmd_t *);
+extern int transport_generic_do_tmr(se_cmd_t *);
+
+/*
+ * Each se_transport_task_t can have N number of possible se_task_t's
+ * for the storage transport(s) to possibly execute.
+ * Used primarily for splitting up CDBs that exceed the physical storage
+ * HBA's maximum sector count per task.
+ */
+typedef struct se_mem_s {
+ struct page *se_page;
+ u32 se_len;
+ u32 se_off;
+ struct list_head se_list;
+} ____cacheline_aligned se_mem_t;
+
+/*
+ * Each type of DAS transport that uses the generic command sequencer needs
+ * each of the following function pointers set.
+ */
+typedef struct se_subsystem_spc_s {
+ int (*inquiry)(se_task_t *, u32);
+ int (*none)(se_task_t *, u32);
+ int (*read_non_SG)(se_task_t *, u32);
+ int (*read_SG)(se_task_t *, u32);
+ int (*write_non_SG)(se_task_t *, u32);
+ int (*write_SG)(se_task_t *, u32);
+} se_subsystem_spc_t;
+
+/*
+ * Each type of disk transport supported MUST have a template defined
+ * within its .h file.
+ */
+typedef struct se_subsystem_api_s {
+ /*
+ * The Name. :-)
+ */
+ char name[16];
+ /*
+ * Plugin Type.
+ */
+ u8 type;
+ /*
+ * Transport Type.
+ */
+ u8 transport_type;
+ /*
+ * attach_hba():
+ */
+ int (*attach_hba)(se_hba_t *, u32);
+ /*
+ * detach_hba():
+ */
+ int (*detach_hba)(struct se_hba_s *);
+ /*
+ * claim_phydevice(): Only for Physical HBAs
+ */
+ int (*claim_phydevice)(struct se_hba_s *, struct se_device_s *);
+ /*
+ * allocate_virtdevice():
+ */
+ void *(*allocate_virtdevice)(struct se_hba_s *, const char *);
+ /*
+ * create_virtdevice(): Only for Virtual HBAs
+ */
+ se_device_t *(*create_virtdevice)(struct se_hba_s *,
+ struct se_subsystem_dev_s *, void *);
+ /*
+ * activate_device():
+ */
+ int (*activate_device)(struct se_device_s *);
+ /*
+ * deactivate_device():
+ */
+ void (*deactivate_device)(struct se_device_s *);
+ /*
+ * release_phydevice():
+ */
+ int (*release_phydevice)(struct se_device_s *);
+ /*
+ * free_device():
+ */
+ void (*free_device)(void *);
+ /*
+ * cmd_sequencer():
+ *
+ * Use transport_generic_cmd_sequencer() for majority of DAS transport
+ * drivers with a scsi_transport_spc_t struct as mentioned below.
+ * Provided out of convenience.
+ */
+ int (*cmd_sequencer)(se_cmd_t *cmd);
+ /*
+ * do_tmr():
+ *
+ * Use transport_do_tmr() for majority of DAS transport drivers.
+ * Provided out of convenience.
+ */
+ int (*do_tmr)(se_cmd_t *cmd);
+ /*
+ * transport_complete():
+ *
+ * Use transport_generic_complete() for majority of DAS transport
+ * drivers. Provided out of convenience.
+ */
+ int (*transport_complete)(se_task_t *task);
+ /*
+ * allocate_request():
+ */
+ void *(*allocate_request)(se_task_t *, se_device_t *);
+ /*
+ * allocate_buf():
+ */
+ int (*allocate_buf)(se_cmd_t *, u32, u32);
+ /*
+ * allocate_DMA();
+ */
+ int (*allocate_DMA)(se_cmd_t *, u32, u32);
+ /*
+ * free_buf():
+ */
+ void (*free_buf)(se_cmd_t *);
+ /*
+ * free_DMA():
+ */
+ void (*free_DMA)(se_cmd_t *);
+ /*
+ * do_task():
+ */
+ int (*do_task)(se_task_t *);
+ /*
+ * free_task():
+ */
+ void (*free_task)(se_task_t *);
+ /*
+ * check_configfs_dev_params():
+ */
+ ssize_t (*check_configfs_dev_params)(se_hba_t *, se_subsystem_dev_t *);
+ /*
+ * set_configfs_dev_params():
+ */
+ ssize_t (*set_configfs_dev_params)(se_hba_t *, se_subsystem_dev_t *,
+ const char *, ssize_t);
+ /*
+ * show_configfs_dev_params():
+ */
+ ssize_t (*show_configfs_dev_params)(se_hba_t *, se_subsystem_dev_t *,
+ char *);
+ /*
+ * create_virtdevice_from-fd():
+ */
+ se_device_t *(*create_virtdevice_from_fd)(se_subsystem_dev_t *,
+ const char *);
+ /*
+ * get_plugin_info():
+ */
+ void (*get_plugin_info)(void *, char *, int *);
+ /*
+ * get_hba_info():
+ */
+ void (*get_hba_info)(se_hba_t *, char *, int *);
+ /*
+ * get_dev_info():
+ */
+ void (*get_dev_info)(se_device_t *, char *, int *);
+ /*
+ * check_lba():
+ */
+ int (*check_lba)(unsigned long long lba, se_device_t *);
+ /*
+ * check_for_SG():
+ */
+ int (*check_for_SG)(se_task_t *);
+ /*
+ * get_cdb():
+ */
+ unsigned char *(*get_cdb)(se_task_t *);
+ /*
+ * get_blocksize():
+ */
+ u32 (*get_blocksize)(se_device_t *);
+ /*
+ * get_device_rev():
+ */
+ u32 (*get_device_rev)(se_device_t *);
+ /*
+ * get_device_type():
+ */
+ u32 (*get_device_type)(se_device_t *);
+ /*
+ * get_dma_length():
+ */
+ u32 (*get_dma_length)(u32, se_device_t *);
+ /*
+ * get_max_cdbs():
+ */
+ u32 (*get_max_cdbs)(se_device_t *);
+ /*
+ * get_max_sectors():
+ */
+ u32 (*get_max_sectors)(se_device_t *);
+ /*
+ * get_queue_depth():
+ *
+ */
+ u32 (*get_queue_depth)(se_device_t *);
+ /*
+ * get_max_queue_depth():
+ */
+ u32 (*get_max_queue_depth)(se_device_t *);
+ /*
+ * do_se_mem_map():
+ */
+ int (*do_se_mem_map)(se_task_t *, struct list_head *, void *,
+ se_mem_t *, se_mem_t **, u32 *, u32 *);
+ /*
+ * get_sense_buffer():
+ */
+ unsigned char *(*get_sense_buffer)(se_task_t *);
+ /*
+ * map_task_to_SG():
+ */
+ void (*map_task_to_SG)(se_task_t *);
+ /*
+ * set_iovec_ptrs():
+ */
+ int (*set_iovec_ptrs)(se_map_sg_t *, se_unmap_sg_t *);
+ /*
+ * shutdown_hba():
+ */
+ void (*shutdown_hba)(se_hba_t *);
+ /*
+ * write_pending():
+ */
+ int (*write_pending)(se_task_t *);
+ /*
+ * se_subsystem_spc_t structure:
+ *
+ * Contains function pointers of SPC opcodes to call from the generic
+ * command sequencer into a transport driver if the generic command
+ * sequencer is used. (ie: cmd_sequencer() is NULL)
+ */
+ se_subsystem_spc_t *spc;
+} ____cacheline_aligned se_subsystem_api_t;
+
+#define TRANSPORT(dev) ((dev)->transport)
+#define TRANSPORT_SPC(dev) ((dev)->transport->spc)
+#define HBA_TRANSPORT(hba) ((hba)->transport)
+
+#endif /* TARGET_CORE_TRANSPORT_H */
+
diff --git a/include/target/target_core_transport_plugin.h b/include/target/target_core_transport_plugin.h
new file mode 100644
index 0000000..118ee3a
--- /dev/null
+++ b/include/target/target_core_transport_plugin.h
@@ -0,0 +1,60 @@
+/*******************************************************************************
+ * Filename: target_core_transport_plugin.h
+ *
+ * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
+ * Copyright (c) 2005, 2006, 2007 SBE, Inc.
+ * Copyright (c) 2007-2009 Rising Tide Software, Inc.
+ * Copyright (c) 2008-2009 Linux-iSCSI.org
+ *
+ * Nicholas A. Bellinger <nab@...nel.org>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+ *
+ ******************************************************************************/
+
+
+#ifndef _TARGET_CORE_TRANSPORT_PLUGIN_H_
+#define _TARGET_CORE_TRANSPORT_PLUGIN_H_
+
+#ifdef PARALLEL_SCSI
+#define PSCSI_INCLUDE_STRUCTS
+#include <target/target_core_pscsi.h>
+#undef PSCSI_INCLUDE_STRUCTS
+extern se_subsystem_api_t pscsi_template;
+#endif /* PARALLEL_SCSI */
+
+#ifdef PYX_IBLOCK
+#define IBLOCK_INCLUDE_STRUCTS
+#include <target/target_core_iblock.h>
+#undef IBLOCK_INCLUDE_STRUCTS
+extern se_subsystem_api_t iblock_template;
+#endif /* PYX_IBLOCK */
+
+#ifdef PYX_RAMDISK
+#define RD_INCLUDE_STRUCTS
+#include <target/target_core_rd.h>
+#undef RD_INCLUDE_STRUCTS
+extern se_subsystem_api_t rd_dr_template;
+extern se_subsystem_api_t rd_mcp_template;
+#endif /* PYX_RAMDISK */
+
+#ifdef PYX_FILEIO
+#define FD_INCLUDE_STRUCTS
+#include <target/target_core_file.h>
+#undef FD_INCLUDE_STRUCTS
+extern se_subsystem_api_t fileio_template;
+#endif /* PYX_FILEIO */
+
+#endif /*** _TARGET_CORE_TRANSPORT_PLUGIN_H_ ***/
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists