lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4C90E2F1.4040904@panasas.com>
Date:	Wed, 15 Sep 2010 17:14:57 +0200
From:	Boaz Harrosh <bharrosh@...asas.com>
To:	"Nicholas A. Bellinger" <nab@...ux-iscsi.org>
CC:	linux-scsi <linux-scsi@...r.kernel.org>,
	linux-kernel <linux-kernel@...r.kernel.org>,
	FUJITA Tomonori <fujita.tomonori@....ntt.co.jp>,
	Mike Christie <michaelc@...wisc.edu>,
	Hannes Reinecke <hare@...e.de>,
	James Bottomley <James.Bottomley@...e.de>,
	Konrad Rzeszutek Wilk <konrad@...nok.org>,
	Douglas Gilbert <dgilbert@...erlog.com>,
	Joe Eykholt <jeykholt@...co.com>
Subject: Re: [PATCH 1/2] tcm: Add support for BIDI operation and XDWRITE_READ_10
 emulation

On 09/15/2010 02:06 PM, Nicholas A. Bellinger wrote:
> From: Nicholas Bellinger <nab@...ux-iscsi.org>
> 
> This patch series adds DMA_BIDIRECTIONAL support to TCM Core.  This
> includes the following for struct se_cmd to handle BIDI READ payloads using a new
> struct se_transport_task->t_mem_bidi_list and struct se_transport_task->t_tasks_se_bidi_num.
> The model used to keep the WRITE payload at struct se_transport_task->t_mem_list,
> and add a new READ payload memory list at struct se_transport_task->t_mem_bidi_list.
> 
> *) desciptor setup:
> 
> This patch adds support for XDWRITEREAD_10 within transport_generic_cmd_sequencer(),
> and sets up the new struct se_cmd->transport_xor_callback() completion function
> at transport_xor_callback().  It then updates transport_new_cmd_obj() to handle the
> BIDI READ case.
> 
> It also adds support for DMA_BIDIRECTIONAL to transport_generic_map_mem_to_cmd()
> to setup struct se_transport_task->t_mem_bidi_list for BIDI READ payloads
> 
> *) memory mapping:
> 
> This patch updates transport_generic_get_cdb_count() to accept a enum dma_data_direction
> parameter to handle the BIDI and the existing WRITE/READ cases for SCF_SCSI_DATA_SG_IO_CDB.
> This patch also updates transport_generic_map_mem_to_cmd() to accept a 'void *mem_bidi_in'
> and 'u32 se_mem_bidi_num' from BIDI capable TCM fabric modules.
> 
> It then updates transport_generic_get_task() and struct se_cmd->transport_get_task()
> to accept a enum dma_data_direction function parameter.
> 
> *) descriptor callback:
> 
> For the struct se_cmd callback, it updates transport_generic_complete_ok() to support
> DMA_BIDIRECTIONAL and looks for the new struct se_cmd->transport_xor_callback() in
> transport_generic_cmd_sequencer() to perform the post READ/WRITE XOR emulation.
> This also includes the additon of transport_memcpy_se_mem_read_contig() used to copy the
> WRITE scatterlists into a local contigious buffer for the XOR instructions within
> transport_xor_callback();
> 
> *) descriptor release:
> 
> Update transport_free_pages() to walk the new T_TASK(cmd)->t_mem_bidi_list (when available)
> and release struct se_mem and pages.
> 
> So far this has been tested with TCM_Loop using BSG w/ userspace code generating
> BIDI XDWRITE_READ_10 CDBs.
> 

Hi.

OK so lets see if we can manage with the scsi-ml model of:

BIDI-COMMANDS ==>  data_direction == DMA_TO_DEVICE && se_transport_task->t_mem_bidi is present

> Signed-off-by: Nicholas A. Bellinger <nab@...ux-iscsi.org>
> ---
>  drivers/target/target_core_transport.c |  282 ++++++++++++++++++++++++++++----
>  include/target/target_core_base.h      |    9 +-
>  include/target/target_core_transport.h |   10 +-
>  3 files changed, 261 insertions(+), 40 deletions(-)
> 
> diff --git a/drivers/target/target_core_transport.c b/drivers/target/target_core_transport.c
> index 517a59c..37afc39 100644
> --- a/drivers/target/target_core_transport.c
> +++ b/drivers/target/target_core_transport.c
> @@ -2529,7 +2529,8 @@ static inline int transport_check_device_cdb_sector_count(
>  static struct se_task *transport_generic_get_task(
>  	struct se_transform_info *ti,
>  	struct se_cmd *cmd,
> -	void *se_obj_ptr)
> +	void *se_obj_ptr,
> +	enum dma_data_direction data_direction)

We don't need this direction I think

>  {
>  	struct se_task *task;
>  	struct se_device *dev = SE_DEV(cmd);
> @@ -2625,7 +2626,8 @@ static int transport_process_control_sg_transform(
>  		return -1;
>  	}
>  
> -	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr);
> +	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
> +				cmd->data_direction);

I can't seem to find the changed implementation of cmd->transport_get_task
in this patch or the next one. But it seems this here is just a pass throw
to cmd->transport_get_task so the argument should go there.

>  	if (!(task))
>  		return -1;
>  
> @@ -2665,7 +2667,8 @@ static int transport_process_control_nonsg_transform(
>  	unsigned char *cdb;
>  	struct se_task *task;
>  
> -	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr);
> +	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
> +				cmd->data_direction);

Same

>  	if (!(task))
>  		return -1;
>  
> @@ -2699,7 +2702,8 @@ static int transport_process_non_data_transform(
>  	unsigned char *cdb;
>  	struct se_task *task;
>  
> -	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr);
> +	task = cmd->transport_get_task(ti, cmd, ti->se_obj_ptr,
> +				cmd->data_direction);

Same

>  	if (!(task))
>  		return -1;
>  
> @@ -5183,6 +5187,54 @@ int transport_generic_emulate_request_sense(
>  }
>  EXPORT_SYMBOL(transport_generic_emulate_request_sense);
>  
> +static void transport_xor_callback(struct se_cmd *cmd)
> +{
> +	unsigned char *buf, *addr;
> +	struct se_mem *se_mem;
> +	unsigned int offset;
> +	int i;
> +	/*
> +	 * From sbc3r22.pdf section 5.48 XDWRITEREAD (10) command
> +	 * 
> +	 * 1) read the specified logical block(s);
> +	 * 2) transfer logical blocks from the data-out buffer;
> +	 * 3) XOR the logical blocks transferred from the data-out buffer with
> +	 *    the logical blocks read, storing the resulting XOR data in a buffer;
> +	 * 4) if the DISABLE WRITE bit is set to zero, then write the logical
> +	 *    blocks transferred from the data-out buffer; and
> +	 * 5) transfer the resulting XOR data to the data-in buffer.
> +	 */
> +	buf = kmalloc(cmd->data_length, GFP_KERNEL);
> +	if (!(buf)) {
> +		printk(KERN_ERR "Unable to allocate xor_callback buf\n");
> +		return;
> +	}
> +	/*
> +	 * Copy the scatterlist WRITE buffer located at T_TASK(cmd)->t_mem_list
> +	 * into the locally allocated *buf
> +	 */
> +	transport_memcpy_se_mem_read_contig(cmd, buf, T_TASK(cmd)->t_mem_list);

This is not relevant to our discussion but did you copy the out-buffer to a
contiguous buffer so not to juggle a double-list traversal? Or am I missing
something more cardinal.

If so the a TODO: comment might be in place. Surely it's possible to walk
two sg-list lists and xor them.

> +	/*
> +	 * Now perform the XOR against the BIDI read memory located at
> +	 * T_TASK(cmd)->t_mem_bidi_list
> +	 */
> +
> +	offset = 0;
> +	list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_bidi_list, se_list) {
> +		addr = (unsigned char *)kmap_atomic(se_mem->se_page, KM_USER0);
> +		if (!(addr))
> +			goto out;
> +
> +		for (i = 0; i < se_mem->se_len; i++)
> +			*(addr + se_mem->se_off + i) ^= *(buf + offset + i);
> +
> +		offset += se_mem->se_len;
> +		kunmap_atomic(addr, KM_USER0);
> +	}
> +out:
> +	kfree(buf);
> +}
> +
>  /*
>   * Used to obtain Sense Data from underlying Linux/SCSI struct scsi_cmnd
>   */
> @@ -5472,6 +5524,25 @@ static int transport_generic_cmd_sequencer(
>  		T_TASK(cmd)->t_tasks_fua = (cdb[1] & 0x8);
>  		ret = TGCS_DATA_SG_IO_CDB;
>  		break;
> +	case XDWRITEREAD_10:
> +		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
> +		if (cmd->data_direction != DMA_BIDIRECTIONAL)
> +			return TGCS_INVALID_CDB_FIELD;

Just check for the presences of the bidi_list. Where is it?

> +		sectors = transport_get_sectors_10(cdb, cmd, &sector_ret);
> +		if (sector_ret)
> +			return TGCS_UNSUPPORTED_CDB;
> +		size = transport_get_size(sectors, cdb, cmd);
> +		transport_dev_get_mem_SG(cmd->se_orig_obj_ptr, cmd);
> +		transport_get_maps(cmd);
> +		cmd->transport_split_cdb = &split_cdb_XX_10;
> +		cmd->transport_get_lba = &transport_lba_32;
> +		/*
> +		 * Setup BIDI XOR callback to be run during transport_generic_complete_ok()
> +		 */
> +		cmd->transport_xor_callback = &transport_xor_callback;
> +		T_TASK(cmd)->t_tasks_fua = (cdb[1] & 0x8);
> +		ret = TGCS_DATA_SG_IO_CDB;
> +		break;
>  	case 0xa3:
>  		SET_GENERIC_TRANSPORT_FUNCTIONS(cmd);
>  		if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
> @@ -5842,9 +5913,10 @@ static int transport_generic_cmd_sequencer(
>  
>  		cmd->cmd_spdtl = size;
>  
> -		if (cmd->data_direction == DMA_TO_DEVICE) {
> +		if ((cmd->data_direction == DMA_TO_DEVICE) ||
> +		    (cmd->data_direction == DMA_BIDIRECTIONAL)) {
>  			printk(KERN_ERR "Rejecting underflow/overflow"
> -					" WRITE data\n");
> +					" WRITE or BIDI data\n");

The data_direction == DMA_TO_DEVICE will stay with our alternate model
so this does not change

>  			return TGCS_INVALID_CDB_FIELD;
>  		}
>  		/*
> @@ -6096,6 +6168,33 @@ void transport_memcpy_read_contig(
>  }
>  EXPORT_SYMBOL(transport_memcpy_read_contig);
>  
> +void transport_memcpy_se_mem_read_contig(
> +	struct se_cmd *cmd,
> +	unsigned char *dst,
> +	struct list_head *se_mem_list)
> +{
> +	struct se_mem *se_mem;
> +	void *src;
> +	u32 length = 0, total_length = cmd->data_length;
> +
> +	list_for_each_entry(se_mem, se_mem_list, se_list) {
> +		length = se_mem->se_len;
> +		
> +		if (length > total_length)
> +			length = total_length;
> +
> +		src = page_address(se_mem->se_page) + se_mem->se_off;
> +		
> +		memcpy(dst, src, length);
> +
> +		if (!(total_length -= length))
> +			return;
> +
> +		dst += length;
> +	}
> +}
> +
> +
>  /*     transport_generic_passthrough():
>   *
>   *
> @@ -6252,6 +6351,18 @@ void transport_generic_complete_ok(struct se_cmd *cmd)
>  
>  	switch (cmd->data_direction) {
>  	case DMA_FROM_DEVICE:
> +	case DMA_BIDIRECTIONAL:
> +		/*
> +		 * Check for the XOR BIDI callback emulation for XD_WRITEREAD_*
> +		 */
> +		if (cmd->transport_xor_callback) {

What? this one place is a bad hack. There are 200 and some BIDI commands in the
scsi-protocol. XOR is just a small group of them. Do you intend to add such an if
for every command type? I don't think so. The target in question should just be
able to proprly hook into the transport_complete mechanics and clean after itself
some how. What does the cmd->transport_xor_callback do? Why no just do a:

if (cmd->transport_complete__callback)
	cmd->transport_complete_callback(cmd);

for any kind of direction, as a target set policy. (When there are chained
processing to do)

> +			/*
> +			 * For fully emulated HBAs, this will translate to
> +			 * transport_xor_callback()
> +			 */
> +			cmd->transport_xor_callback(cmd);
> +		}
> +

>  		spin_lock(&cmd->se_lun->lun_sep_lock);
>  		if (SE_LUN(cmd)->lun_sep) {
>  			SE_LUN(cmd)->lun_sep->sep_stats.tx_data_octets +=
> @@ -6347,6 +6458,23 @@ static inline void transport_free_pages(struct se_cmd *cmd)
>  		kmem_cache_free(se_mem_cache, se_mem);
>  	}
>  
> +	if (T_TASK(cmd)->t_mem_bidi_list && T_TASK(cmd)->t_tasks_se_bidi_num) {

See here: we have a buffer we take care of it. Simple

> +		list_for_each_entry_safe(se_mem, se_mem_tmp,
> +				T_TASK(cmd)->t_mem_bidi_list, se_list) {
> +			/*
> +			 * We only release call __free_page(struct se_mem->se_page) when
> +			 * SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC is NOT in use,
> +			 */
> +			if (free_page)
> +				__free_page(se_mem->se_page);
> +
> +			list_del(&se_mem->se_list);
> +			kmem_cache_free(se_mem_cache, se_mem);
> +		}
> +	}
> +
> +	kfree(T_TASK(cmd)->t_mem_bidi_list);
> +	T_TASK(cmd)->t_mem_bidi_list = NULL;
>  	kfree(T_TASK(cmd)->t_mem_list);
>  	T_TASK(cmd)->t_mem_list = NULL;
>  	T_TASK(cmd)->t_tasks_se_num = 0;
> @@ -6477,18 +6605,34 @@ release_cmd:
>  int transport_generic_map_mem_to_cmd(
>  	struct se_cmd *cmd,
>  	void *mem,
> -	u32 se_mem_num)
> +	u32 se_mem_num,
> +	void *mem_bidi_in,
> +	u32 se_mem_bidi_num)
>  {
>  	u32 se_mem_cnt_out = 0;
>  	int ret;
>  
>  	if (!(mem) || !(se_mem_num))
>  		return 0;
> +
> +	if ((cmd->data_direction == DMA_BIDIRECTIONAL) &&
> +	    (!(mem_bidi_in) || !(se_mem_bidi_num))) {
> +		printk(KERN_ERR "Unable to process DMA_BIDIRECTIONAL with mem_bidi_in:"
> +			" %p and se_mem_bidi_num: %u\n", mem_bidi_in, se_mem_bidi_num);
> +		return -EINVAL;
> +	}
> +

In our new model this just drops. Because only one member caries information. mem_bidi_in
is the flag for bidi presence.

>  	/*
>  	 * Passed *mem will contain a list_head containing preformatted
>  	 * struct se_mem elements...
>  	 */
>  	if (!(cmd->se_cmd_flags & SCF_PASSTHROUGH_SG_TO_MEM)) {
> +		if (cmd->data_direction == DMA_BIDIRECTIONAL) {
> +			printk(KERN_ERR "SCF_CMD_PASSTHROUGH_NOALLOC not supported"
> +				" with DMA_BIDIRECTIONAL\n");
> +			return -ENOSYS;

What is the issue here. Why can't we bidi in this case?

> +		}
> +
>  		T_TASK(cmd)->t_mem_list = (struct list_head *)mem;
>  		T_TASK(cmd)->t_tasks_se_num = se_mem_num;
>  		cmd->se_cmd_flags |= SCF_CMD_PASSTHROUGH_NOALLOC;
> @@ -6507,14 +6651,35 @@ int transport_generic_map_mem_to_cmd(
>  		 */ 
>  		T_TASK(cmd)->t_mem_list = transport_init_se_mem_list();
>  		if (!(T_TASK(cmd)->t_mem_list))
> -			return -1;
> +			return -ENOMEM;
>  
>  		ret = transport_map_sg_to_mem(cmd,
>  			T_TASK(cmd)->t_mem_list, mem, &se_mem_cnt_out);
>  		if (ret < 0)
> -			return -1;
> +			return -ENOMEM;

These two belong to another patch right?

>  
>  		T_TASK(cmd)->t_tasks_se_num = se_mem_cnt_out;
> +		/*
> +		 * Setup BIDI READ list of struct se_mem elements
> +		 */
> +		if (cmd->data_direction == DMA_BIDIRECTIONAL) {

Just if (mem_bidi_in) no?

> +			T_TASK(cmd)->t_mem_bidi_list = transport_init_se_mem_list();
> +			if (!(T_TASK(cmd)->t_mem_bidi_list)) {
> +				kfree(T_TASK(cmd)->t_mem_list);
> +				return -ENOMEM;
> +			}
> +			se_mem_cnt_out = 0;
> +
> +			ret = transport_map_sg_to_mem(cmd,
> +				T_TASK(cmd)->t_mem_bidi_list, mem_bidi_in,
> +				&se_mem_cnt_out);
> +			if (ret < 0) {
> +				kfree(T_TASK(cmd)->t_mem_list);
> +				return -ENOMEM;
> +			}
> +
> +			T_TASK(cmd)->t_tasks_se_bidi_num = se_mem_cnt_out;
> +		}	
>  		cmd->se_cmd_flags |= SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC;
>  
>  	} else if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_NONSG_IO_CDB) {
> @@ -6610,6 +6775,11 @@ non_scsi_data:
>   */
>  int transport_generic_do_transform(struct se_cmd *cmd, struct se_transform_info *ti)
>  {
> +	if (!(cmd->transport_cdb_transform)) {
> +		dump_stack();
> +		return -1;
> +	}
> +	
>  	if (cmd->transport_cdb_transform(cmd, ti) < 0)
>  		return -1;
>  
> @@ -6656,9 +6826,9 @@ int transport_new_cmd_obj(
>  	struct se_transform_info *ti,
>  	int post_execute)
>  {
> -	u32 task_cdbs = 0;
> -	struct se_mem *se_mem_out = NULL;
>  	struct se_device *dev = SE_DEV(cmd);
> +	enum dma_data_direction data_direction;
> +	u32 task_cdbs = 0, rc;
>  
>  	if (!(cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)) {
>  		task_cdbs++;
> @@ -6666,11 +6836,38 @@ int transport_new_cmd_obj(
>  	} else {
>  		ti->ti_set_counts = 1;
>  		ti->ti_dev = dev;
> -
> +		/*
> +		 * Setup any BIDI READ tasks and memory from
> +		 * T_TASK(cmd)->t_mem_bidi_list so the READ struct se_tasks
> +		 * are queued first..
> +		 */
> +		if (cmd->data_direction == DMA_BIDIRECTIONAL) {

if (T_TASK(cmd)->t_mem_bidi_list)

> +			rc = transport_generic_get_cdb_count(cmd, ti,
> +				T_TASK(cmd)->t_task_lba,
> +				T_TASK(cmd)->t_tasks_sectors,
> +				DMA_FROM_DEVICE, T_TASK(cmd)->t_mem_bidi_list);
> +			if (!(rc)) {
> +				cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
> +				cmd->scsi_sense_reason =
> +					TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
> +				return PYX_TRANSPORT_LU_COMM_FAILURE;
> +			}
> +			ti->ti_set_counts = 0;
> +			/*
> +			 * Setup the DMA_TO_DEVICE direction for the next
> +			 * call to transport_generic_get_cdb_count()
> +			 */
> +			data_direction = DMA_TO_DEVICE;

If I understand this code correctly. Then this here is wrong. You are assuming a XOR operation
where we read some data then write some when that's finish. But other BIDI commands might be different
they might write first then read. Or they might might read and write in parallel. Any such decisions
should be left to the target to make. And at a per-command basis.

I know that in stgt (In user mode) the backend (what you call target, right) receives both buffers
and decides what to do with them then calls complete.

If you need any cache flushing and such. Then the DMA_TO_DEVICE main-buffer is flushed before hand
amd the bidi_read buffer if present is flushed after-hand (or is that the other way) but the generic
layer should no assume any ordering or execution.

Please explain why you had to set the direction here. What would not happen if you did not.

> +		} else
> +			data_direction = cmd->data_direction;
> +		/*
> +		 * Setup the tasks and memory from T_TASK(cmd)->t_mem_list
> +		 * Note for BIDI transfers this will contain the WRITE payload
> +		 */
>  		task_cdbs = transport_generic_get_cdb_count(cmd, ti,
>  				T_TASK(cmd)->t_task_lba,
>  				T_TASK(cmd)->t_tasks_sectors,
> -				NULL, &se_mem_out);
> +				data_direction,	T_TASK(cmd)->t_mem_list);
>  		if (!(task_cdbs)) {
>  			cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
>  			cmd->scsi_sense_reason =
> @@ -6743,7 +6940,17 @@ int transport_generic_get_mem(struct se_cmd *cmd, u32 length, u32 dma_size)
>  
>  	T_TASK(cmd)->t_mem_list = transport_init_se_mem_list();
>  	if (!(T_TASK(cmd)->t_mem_list))
> -		return -1;
> +		return -ENOMEM;
> +	/*
> +	 * Setup BIDI READ list of struct se_mem elements
> +	 */
> +	if (cmd->data_direction == DMA_BIDIRECTIONAL) {

If (T_TASK(cmd)->t_mem_bidi_list)
And the second check just drops

> +		T_TASK(cmd)->t_mem_bidi_list = transport_init_se_mem_list();
> +		if (!(T_TASK(cmd)->t_mem_bidi_list)) {
> +			kfree(T_TASK(cmd)->t_mem_list);
> +			return -ENOMEM;
> +		}
> +	}
>  
>  	while (length) {
>  		se_mem = kmem_cache_zalloc(se_mem_cache, GFP_KERNEL);
> @@ -7240,28 +7447,28 @@ u32 transport_generic_get_cdb_count(
>  	struct se_transform_info *ti,
>  	unsigned long long starting_lba,
>  	u32 sectors,
> -	struct se_mem *se_mem_in,
> -	struct se_mem **se_mem_out)
> +	enum dma_data_direction data_direction,
> +	struct list_head *mem_list)

Note for me:
You moved from an se_mem * to a struct list_head *, than also added a direction.

>  {
>  	unsigned char *cdb = NULL;
>  	struct se_task *task;
> -	struct se_mem *se_mem, *se_mem_lout = NULL;
> +	struct se_mem *se_mem = NULL, *se_mem_lout = NULL;
>  	struct se_device *dev = SE_DEV(cmd);
>  	int max_sectors_set = 0, ret;
>  	u32 task_offset_in = 0, se_mem_cnt = 0, task_cdbs = 0;
>  	unsigned long long lba;
>  
> -	if (!se_mem_in) {
> -		list_for_each_entry(se_mem_in, T_TASK(cmd)->t_mem_list, se_list)
> -			break;
> -
> -		if (!se_mem_in) {
> -			printk(KERN_ERR "se_mem_in is NULL\n");
> -			return 0;
> -		}
> +	if (!mem_list) {
> +		printk(KERN_ERR "mem_list is NULL in transport_generic_get"
> +				"_cdb_count()\n");
> +		return 0;
>  	}
> -	se_mem = se_mem_in;
> -
> +	/*
> +	 * While using RAMDISK_DR backstores is the only case where
> +	 * mem_list will ever be empty at this point.
> +	 */
> +	if (!(list_empty(mem_list)))
> +		se_mem = list_entry(mem_list->next, struct se_mem, se_list);
>  	/*
>  	 * Locate the start volume segment in which the received LBA will be
>  	 * executed upon.
> @@ -7280,7 +7487,12 @@ u32 transport_generic_get_cdb_count(
>  			CMD_TFO(cmd)->get_task_tag(cmd), lba, sectors,
>  			transport_dev_end_lba(dev));
>  
> -		task = cmd->transport_get_task(ti, cmd, dev);
> +		if (!(cmd->transport_get_task)) {
> +			dump_stack();
> +			goto out;
> +		}
> +
> +		task = cmd->transport_get_task(ti, cmd, dev, data_direction);

Again this mysterious cmd->transport_get_task(). What I don't understand is if
the transport (The transport is the fabric right? like iscsi or tcm_loop) wants
to give you the next task to preform. Then it should know better then anybody
what kind of command it is, No? what is the data_direction used for?

>  		if (!(task))
>  			goto out;
>  
> @@ -7293,7 +7505,7 @@ u32 transport_generic_get_cdb_count(
>  		task->task_size = (task->task_sectors *
>  				   DEV_ATTRIB(dev)->block_size);
>  		task->transport_map_task = transport_dev_get_map_SG(dev,
> -					cmd->data_direction);
> +					data_direction);
>  
>  		cdb = TRANSPORT(dev)->get_cdb(task);
>  		if ((cdb)) {
> @@ -7306,14 +7518,13 @@ u32 transport_generic_get_cdb_count(
>  		 * Perform the SE OBJ plugin and/or Transport plugin specific
>  		 * mapping for T_TASK(cmd)->t_mem_list.
>  		 */
> -		ret = transport_do_se_mem_map(dev, task,
> -				T_TASK(cmd)->t_mem_list, NULL, se_mem,
> -				&se_mem_lout, &se_mem_cnt, &task_offset_in);
> +		ret = transport_do_se_mem_map(dev, task, mem_list,
> +				NULL, se_mem, &se_mem_lout, &se_mem_cnt,
> +				&task_offset_in);
>  		if (ret < 0)
>  			goto out;
>  
>  		se_mem = se_mem_lout;
> -		*se_mem_out = se_mem_lout;

I don't understand these changes. I'm out of context. It looks like most of them
are not relevant to the bidi issue, and are cleanups where now you don't need
the *se_mem_out results and it was just dropped. (Perhaps because no body used it)
is that relevant to BIDI? Or you just took the chance of removing a parameter when
you had to add one. (I hope the data_direction will not be needed after all)

>  		task_cdbs++;
>  
>  		DEBUG_VOL("Incremented task_cdbs(%u) task->task_sg_num(%u)\n",
> @@ -7333,8 +7544,9 @@ u32 transport_generic_get_cdb_count(
>  		atomic_inc(&T_TASK(cmd)->t_se_count);
>  	}
>  
> -	DEBUG_VOL("ITT[0x%08x] total cdbs(%u)\n",
> -		CMD_TFO(cmd)->get_task_tag(cmd), task_cdbs);
> +	DEBUG_VOL("ITT[0x%08x] total %s cdbs(%u)\n",
> +		CMD_TFO(cmd)->get_task_tag(cmd), (data_direction == DMA_TO_DEVICE)
> +		? "DMA_TO_DEVICE" : "DMA_FROM_DEVICE", task_cdbs);
>  
>  	return task_cdbs;
>  out:
> diff --git a/include/target/target_core_base.h b/include/target/target_core_base.h
> index e9b98e9..f1b9364 100644
> --- a/include/target/target_core_base.h
> +++ b/include/target/target_core_base.h
> @@ -419,6 +419,7 @@ struct se_transport_task {
>  	u32			t_tasks_no;
>  	u32			t_tasks_sectors;
>  	u32			t_tasks_se_num;
> +	u32			t_tasks_se_bidi_num;
>  	u32			t_tasks_sg_chained_no;
>  	atomic_t		t_fe_count;
>  	atomic_t		t_se_count;
> @@ -447,8 +448,10 @@ struct se_transport_task {
>  	struct scatterlist	t_tasks_sg_bounce;
>  	void			*t_task_buf;
>  	void			*t_task_pt_buf;
> -	struct list_head	t_task_list;
>  	struct list_head	*t_mem_list;
> +	/* Used for BIDI READ */
> +	struct list_head	*t_mem_bidi_list;
> +	struct list_head	t_task_list;
>  } ____cacheline_aligned;
>  
>  struct se_task {
> @@ -598,7 +601,8 @@ struct se_cmd {
>  	u32 (*transport_get_lba)(unsigned char *);
>  	unsigned long long (*transport_get_long_lba)(unsigned char *);
>  	struct se_task *(*transport_get_task)(struct se_transform_info *,
> -					struct se_cmd *, void *);
> +					struct se_cmd *, void *,
> +					enum dma_data_direction);
>  	int (*transport_map_buffers_to_tasks)(struct se_cmd *);
>  	void (*transport_map_SG_segments)(struct se_unmap_sg *);
>  	void (*transport_passthrough_done)(struct se_cmd *);
> @@ -607,6 +611,7 @@ struct se_cmd {
>  					struct se_unmap_sg *);
>  	void (*transport_split_cdb)(unsigned long long, u32 *, unsigned char *);
>  	void (*transport_wait_for_tasks)(struct se_cmd *, int, int);
> +	void (*transport_xor_callback)(struct se_cmd *);

I see the API changes here. But where is the changed implementations? I can't
find them.

>  	void (*callback)(struct se_cmd *cmd, void *callback_arg,
>  			int complete_status);
>  	void *callback_arg;
> diff --git a/include/target/target_core_transport.h b/include/target/target_core_transport.h
> index ef4c084..20702d7 100644
> --- a/include/target/target_core_transport.h
> +++ b/include/target/target_core_transport.h
> @@ -232,6 +232,8 @@ extern void transport_memcpy_write_contig(struct se_cmd *, struct scatterlist *,
>  				unsigned char *);
>  extern void transport_memcpy_read_contig(struct se_cmd *, unsigned char *,
>  				struct scatterlist *);
> +extern void transport_memcpy_se_mem_read_contig(struct se_cmd *,
> +				unsigned char *, struct list_head *);
>  extern int transport_generic_passthrough_async(struct se_cmd *cmd,
>  				void(*callback)(struct se_cmd *cmd,
>  				void *callback_arg, int complete_status),
> @@ -242,7 +244,8 @@ extern void transport_generic_complete_ok(struct se_cmd *);
>  extern void transport_free_dev_tasks(struct se_cmd *);
>  extern void transport_release_fe_cmd(struct se_cmd *);
>  extern int transport_generic_remove(struct se_cmd *, int, int);
> -extern int transport_generic_map_mem_to_cmd(struct se_cmd *cmd, void *, u32);
> +extern int transport_generic_map_mem_to_cmd(struct se_cmd *cmd, void *, u32,
> +				void *, u32);
>  extern int transport_lun_wait_for_tasks(struct se_cmd *, struct se_lun *);
>  extern int transport_clear_lun_from_sessions(struct se_lun *);
>  extern int transport_check_aborted_status(struct se_cmd *, int);
> @@ -271,8 +274,9 @@ extern int transport_map_mem_to_sg(struct se_task *, struct list_head *,
>  extern void transport_do_task_sg_chain(struct se_cmd *);
>  extern u32 transport_generic_get_cdb_count(struct se_cmd *,
>  					struct se_transform_info *,
> -					unsigned long long, u32, struct se_mem *,
> -					struct se_mem **);
> +					unsigned long long, u32,
> +					enum dma_data_direction,
> +					struct list_head *);
>  extern int transport_generic_new_cmd(struct se_cmd *);
>  extern void transport_generic_process_write(struct se_cmd *);
>  extern int transport_generic_do_tmr(struct se_cmd *);

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ