lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <e7f7d153cfca15ad7705935b30979a26@codeaurora.org>
Date:   Thu, 04 Mar 2021 16:21:57 +0800
From:   Can Guo <cang@...eaurora.org>
To:     daejun7.park@...sung.com
Cc:     Greg KH <gregkh@...uxfoundation.org>, avri.altman@....com,
        jejb@...ux.ibm.com, martin.petersen@...cle.com,
        asutoshd@...eaurora.org, stanley.chu@...iatek.com,
        bvanassche@....org, huobean@...il.com,
        ALIM AKHTAR <alim.akhtar@...sung.com>,
        linux-scsi@...r.kernel.org, linux-kernel@...r.kernel.org,
        JinHwan Park <jh.i.park@...sung.com>,
        Javier Gonzalez <javier.gonz@...sung.com>,
        SEUNGUK SHIN <seunguk.shin@...sung.com>,
        Sung-Jun Park <sungjun07.park@...sung.com>,
        Jinyoung CHOI <j-young.choi@...sung.com>,
        BoRam Shin <boram.shin@...sung.com>
Subject: Re: [PATCH v26 4/4] scsi: ufs: Add HPB 2.0 support

On 2021-03-03 14:29, Daejun Park wrote:
> This patch supports the HPB 2.0.
> 
> The HPB 2.0 supports read of varying sizes from 4KB to 512KB.
> In the case of Read (<= 32KB) is supported as single HPB read.
> In the case of Read (36KB ~ 512KB) is supported by as a combination of
> write buffer command and HPB read command to deliver more PPN.
> The write buffer commands may not be issued immediately due to busy 
> tags.
> To use HPB read more aggressively, the driver can requeue the write 
> buffer
> command. The requeue threshold is implemented as timeout and can be
> modified with requeue_timeout_ms entry in sysfs.
> 
> Signed-off-by: Daejun Park <daejun7.park@...sung.com>
> ---
>  Documentation/ABI/testing/sysfs-driver-ufs |  35 +-
>  drivers/scsi/ufs/ufs-sysfs.c               |   2 +
>  drivers/scsi/ufs/ufs.h                     |   3 +-
>  drivers/scsi/ufs/ufshcd.c                  |  22 +-
>  drivers/scsi/ufs/ufshcd.h                  |   7 +
>  drivers/scsi/ufs/ufshpb.c                  | 622 +++++++++++++++++++--
>  drivers/scsi/ufs/ufshpb.h                  |  67 ++-
>  7 files changed, 679 insertions(+), 79 deletions(-)
> 
> diff --git a/Documentation/ABI/testing/sysfs-driver-ufs
> b/Documentation/ABI/testing/sysfs-driver-ufs
> index bf5cb8846de1..0017eaf89cbe 100644
> --- a/Documentation/ABI/testing/sysfs-driver-ufs
> +++ b/Documentation/ABI/testing/sysfs-driver-ufs
> @@ -1253,14 +1253,14 @@ Description:	This entry shows the number of
> HPB pinned regions assigned to
> 
>  		The file is read only.
> 
> -What:		/sys/class/scsi_device/*/device/hpb_sysfs/hit_cnt
> +What:		/sys/class/scsi_device/*/device/hpb_stat_sysfs/hit_cnt
>  Date:		February 2021
>  Contact:	Daejun Park <daejun7.park@...sung.com>
>  Description:	This entry shows the number of reads that changed to HPB 
> read.
> 
>  		The file is read only.
> 
> -What:		/sys/class/scsi_device/*/device/hpb_sysfs/miss_cnt
> +What:		/sys/class/scsi_device/*/device/hpb_stat_sysfs/miss_cnt
>  Date:		February 2021
>  Contact:	Daejun Park <daejun7.park@...sung.com>
>  Description:	This entry shows the number of reads that cannot be 
> changed to
> @@ -1268,7 +1268,7 @@ Description:	This entry shows the number of
> reads that cannot be changed to
> 
>  		The file is read only.
> 
> -What:		/sys/class/scsi_device/*/device/hpb_sysfs/rb_noti_cnt
> +What:		/sys/class/scsi_device/*/device/hpb_stat_sysfs/rb_noti_cnt
>  Date:		February 2021
>  Contact:	Daejun Park <daejun7.park@...sung.com>
>  Description:	This entry shows the number of response UPIUs that has
> @@ -1276,7 +1276,7 @@ Description:	This entry shows the number of
> response UPIUs that has
> 
>  		The file is read only.
> 
> -What:		/sys/class/scsi_device/*/device/hpb_sysfs/rb_active_cnt
> +What:		/sys/class/scsi_device/*/device/hpb_stat_sysfs/rb_active_cnt
>  Date:		February 2021
>  Contact:	Daejun Park <daejun7.park@...sung.com>
>  Description:	This entry shows the number of active sub-regions 
> recommended by
> @@ -1284,7 +1284,7 @@ Description:	This entry shows the number of
> active sub-regions recommended by
> 
>  		The file is read only.
> 
> -What:		/sys/class/scsi_device/*/device/hpb_sysfs/rb_inactive_cnt
> +What:		/sys/class/scsi_device/*/device/hpb_stat_sysfs/rb_inactive_cnt
>  Date:		February 2021
>  Contact:	Daejun Park <daejun7.park@...sung.com>
>  Description:	This entry shows the number of inactive regions 
> recommended by
> @@ -1292,10 +1292,33 @@ Description:	This entry shows the number of
> inactive regions recommended by
> 
>  		The file is read only.
> 
> -What:		/sys/class/scsi_device/*/device/hpb_sysfs/map_req_cnt
> +What:		/sys/class/scsi_device/*/device/hpb_stat_sysfs/map_req_cnt
>  Date:		February 2021
>  Contact:	Daejun Park <daejun7.park@...sung.com>
>  Description:	This entry shows the number of read buffer commands for
>  		activating sub-regions recommended by response UPIUs.
> 
>  		The file is read only.
> +
> +What:		/sys/class/scsi_device/*/device/hpb_param_sysfs/requeue_timeout_ms
> +Date:		February 2021
> +Contact:	Daejun Park <daejun7.park@...sung.com>
> +Description:	This entry shows the requeue timeout threshold for write 
> buffer
> +		command in ms. This value can be changed by writing proper integer 
> to
> +		this entry.
> +
> +What:		/sys/bus/platform/drivers/ufshcd/*/attributes/max_data_size_hpb_single_cmd
> +Date:		February 2021
> +Contact:	Daejun Park <daejun7.park@...sung.com>
> +Description:	This entry shows the maximum HPB data size for using 
> single HPB
> +		command.
> +
> +		===  ========
> +		00h  4KB
> +		01h  8KB
> +		02h  12KB
> +		...
> +		FFh  1024KB
> +		===  ========
> +
> +		The file is read only.
> diff --git a/drivers/scsi/ufs/ufs-sysfs.c 
> b/drivers/scsi/ufs/ufs-sysfs.c
> index 2546e7a1ac4f..00fb519406cf 100644
> --- a/drivers/scsi/ufs/ufs-sysfs.c
> +++ b/drivers/scsi/ufs/ufs-sysfs.c
> @@ -841,6 +841,7 @@ out:									\
>  static DEVICE_ATTR_RO(_name)
> 
>  UFS_ATTRIBUTE(boot_lun_enabled, _BOOT_LU_EN);
> +UFS_ATTRIBUTE(max_data_size_hpb_single_cmd, _MAX_HPB_SINGLE_CMD);
>  UFS_ATTRIBUTE(current_power_mode, _POWER_MODE);
>  UFS_ATTRIBUTE(active_icc_level, _ACTIVE_ICC_LVL);
>  UFS_ATTRIBUTE(ooo_data_enabled, _OOO_DATA_EN);
> @@ -864,6 +865,7 @@ UFS_ATTRIBUTE(wb_cur_buf, _CURR_WB_BUFF_SIZE);
> 
>  static struct attribute *ufs_sysfs_attributes[] = {
>  	&dev_attr_boot_lun_enabled.attr,
> +	&dev_attr_max_data_size_hpb_single_cmd.attr,
>  	&dev_attr_current_power_mode.attr,
>  	&dev_attr_active_icc_level.attr,
>  	&dev_attr_ooo_data_enabled.attr,
> diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h
> index 957763db1006..e0b748777a1b 100644
> --- a/drivers/scsi/ufs/ufs.h
> +++ b/drivers/scsi/ufs/ufs.h
> @@ -123,12 +123,13 @@ enum flag_idn {
>  	QUERY_FLAG_IDN_WB_BUFF_FLUSH_EN                 = 0x0F,
>  	QUERY_FLAG_IDN_WB_BUFF_FLUSH_DURING_HIBERN8     = 0x10,
>  	QUERY_FLAG_IDN_HPB_RESET                        = 0x11,
> +	QUERY_FLAG_IDN_HPB_EN				= 0x12,

Also add this flag to sysfs?

Thanks,

Can Guo.

>  };
> 
>  /* Attribute idn for Query requests */
>  enum attr_idn {
>  	QUERY_ATTR_IDN_BOOT_LU_EN		= 0x00,
> -	QUERY_ATTR_IDN_RESERVED			= 0x01,
> +	QUERY_ATTR_IDN_MAX_HPB_SINGLE_CMD	= 0x01,
>  	QUERY_ATTR_IDN_POWER_MODE		= 0x02,
>  	QUERY_ATTR_IDN_ACTIVE_ICC_LVL		= 0x03,
>  	QUERY_ATTR_IDN_OOO_DATA_EN		= 0x04,
> diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c
> index 851c01a26207..f7e491ad4fa8 100644
> --- a/drivers/scsi/ufs/ufshcd.c
> +++ b/drivers/scsi/ufs/ufshcd.c
> @@ -2656,7 +2656,12 @@ static int ufshcd_queuecommand(struct Scsi_Host
> *host, struct scsi_cmnd *cmd)
> 
>  	lrbp->req_abort_skip = false;
> 
> -	ufshpb_prep(hba, lrbp);
> +	err = ufshpb_prep(hba, lrbp);
> +	if (err == -EAGAIN) {
> +		lrbp->cmd = NULL;
> +		ufshcd_release(hba);
> +		goto out;
> +	}
> 
>  	ufshcd_comp_scsi_upiu(hba, lrbp);
> 
> @@ -3110,7 +3115,7 @@ int ufshcd_query_attr(struct ufs_hba *hba, enum
> query_opcode opcode,
>   *
>   * Returns 0 for success, non-zero in case of failure
>  */
> -static int ufshcd_query_attr_retry(struct ufs_hba *hba,
> +int ufshcd_query_attr_retry(struct ufs_hba *hba,
>  	enum query_opcode opcode, enum attr_idn idn, u8 index, u8 selector,
>  	u32 *attr_val)
>  {
> @@ -7447,8 +7452,18 @@ static int ufs_get_device_desc(struct ufs_hba 
> *hba)
> 
>  	if (dev_info->wspecversion >= UFS_DEV_HPB_SUPPORT_VERSION &&
>  	    (b_ufs_feature_sup & UFS_DEV_HPB_SUPPORT)) {
> -		dev_info->hpb_enabled = true;
> +		bool hpb_en = false;
> +
>  		ufshpb_get_dev_info(hba, desc_buf);
> +
> +		if (!ufshpb_is_legacy(hba))
> +			err = ufshcd_query_flag_retry(hba,
> +						      UPIU_QUERY_OPCODE_READ_FLAG,
> +						      QUERY_FLAG_IDN_HPB_EN, 0,
> +						      &hpb_en);
> +
> +		if (ufshpb_is_legacy(hba) || (!err && hpb_en))
> +			dev_info->hpb_enabled = true;
>  	}
> 
>  	err = ufshcd_read_string_desc(hba, model_index,
> @@ -8019,6 +8034,7 @@ static const struct attribute_group
> *ufshcd_driver_groups[] = {
>  	&ufs_sysfs_lun_attributes_group,
>  #ifdef CONFIG_SCSI_UFS_HPB
>  	&ufs_sysfs_hpb_stat_group,
> +	&ufs_sysfs_hpb_param_group,
>  #endif
>  	NULL,
>  };
> diff --git a/drivers/scsi/ufs/ufshcd.h b/drivers/scsi/ufs/ufshcd.h
> index 961fc5b77943..7d85517464ef 100644
> --- a/drivers/scsi/ufs/ufshcd.h
> +++ b/drivers/scsi/ufs/ufshcd.h
> @@ -654,6 +654,8 @@ struct ufs_hba_variant_params {
>   * @srgn_size: device reported HPB sub-region size
>   * @slave_conf_cnt: counter to check all lu finished initialization
>   * @hpb_disabled: flag to check if HPB is disabled
> + * @max_hpb_single_cmd: maximum size of single HPB command
> + * @is_legacy: flag to check HPB 1.0
>   */
>  struct ufshpb_dev_info {
>  	int num_lu;
> @@ -661,6 +663,8 @@ struct ufshpb_dev_info {
>  	int srgn_size;
>  	atomic_t slave_conf_cnt;
>  	bool hpb_disabled;
> +	int max_hpb_single_cmd;
> +	bool is_legacy;
>  };
>  #endif
> 
> @@ -1091,6 +1095,9 @@ int ufshcd_read_desc_param(struct ufs_hba *hba,
>  			   u8 param_offset,
>  			   u8 *param_read_buf,
>  			   u8 param_size);
> +int ufshcd_query_attr_retry(struct ufs_hba *hba, enum query_opcode 
> opcode,
> +			    enum attr_idn idn, u8 index, u8 selector,
> +			    u32 *attr_val);
>  int ufshcd_query_attr(struct ufs_hba *hba, enum query_opcode opcode,
>  		      enum attr_idn idn, u8 index, u8 selector, u32 *attr_val);
>  int ufshcd_query_flag(struct ufs_hba *hba, enum query_opcode opcode,
> diff --git a/drivers/scsi/ufs/ufshpb.c b/drivers/scsi/ufs/ufshpb.c
> index c75a6816a03f..ae092a442023 100644
> --- a/drivers/scsi/ufs/ufshpb.c
> +++ b/drivers/scsi/ufs/ufshpb.c
> @@ -31,6 +31,11 @@ bool ufshpb_is_allowed(struct ufs_hba *hba)
>  	return !(hba->ufshpb_dev.hpb_disabled);
>  }
> 
> +bool ufshpb_is_legacy(struct ufs_hba *hba)
> +{
> +	return hba->ufshpb_dev.is_legacy;
> +}
> +
>  static struct ufshpb_lu *ufshpb_get_hpb_data(struct scsi_device *sdev)
>  {
>  	return sdev->hostdata;
> @@ -64,9 +69,19 @@ static bool ufshpb_is_write_or_discard_cmd(struct
> scsi_cmnd *cmd)
>  	       op_is_discard(req_op(cmd->request));
>  }
> 
> -static bool ufshpb_is_support_chunk(int transfer_len)
> +static bool ufshpb_is_support_chunk(struct ufshpb_lu *hpb, int 
> transfer_len)
>  {
> -	return transfer_len <= HPB_MULTI_CHUNK_HIGH;
> +	return transfer_len <= hpb->pre_req_max_tr_len;
> +}
> +
> +/*
> + * In this driver, WRITE_BUFFER CMD support 36KB (len=9) ~ 512KB 
> (len=128) as
> + * default. It is possible to change range of transfer_len through 
> sysfs.
> + */
> +static inline bool ufshpb_is_required_wb(struct ufshpb_lu *hpb, int 
> len)
> +{
> +	return (len > hpb->pre_req_min_tr_len &&
> +		len <= hpb->pre_req_max_tr_len);
>  }
> 
>  static bool ufshpb_is_general_lun(int lun)
> @@ -74,8 +89,7 @@ static bool ufshpb_is_general_lun(int lun)
>  	return lun < UFS_UPIU_MAX_UNIT_NUM_ID;
>  }
> 
> -static bool
> -ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int rgn_idx)
> +static bool ufshpb_is_pinned_region(struct ufshpb_lu *hpb, int 
> rgn_idx)
>  {
>  	if (hpb->lu_pinned_end != PINNED_NOT_SET &&
>  	    rgn_idx >= hpb->lu_pinned_start &&
> @@ -264,7 +278,8 @@ ufshpb_get_pos_from_lpn(struct ufshpb_lu *hpb,
> unsigned long lpn, int *rgn_idx,
> 
>  static void
>  ufshpb_set_hpb_read_to_upiu(struct ufshpb_lu *hpb, struct ufshcd_lrb 
> *lrbp,
> -			    u32 lpn, u64 ppn, unsigned int transfer_len)
> +			    u32 lpn, u64 ppn, unsigned int transfer_len,
> +			    int read_id)
>  {
>  	unsigned char *cdb = lrbp->cmd->cmnd;
> 
> @@ -273,15 +288,269 @@ ufshpb_set_hpb_read_to_upiu(struct ufshpb_lu
> *hpb, struct ufshcd_lrb *lrbp,
>  	/* ppn value is stored as big-endian in the host memory */
>  	memcpy(&cdb[6], &ppn, sizeof(u64));
>  	cdb[14] = transfer_len;
> +	cdb[15] = read_id;
> 
>  	lrbp->cmd->cmd_len = UFS_CDB_SIZE;
>  }
> 
> +static inline void ufshpb_set_write_buf_cmd(unsigned char *cdb,
> +					    unsigned long lpn, unsigned int len,
> +					    int read_id)
> +{
> +	cdb[0] = UFSHPB_WRITE_BUFFER;
> +	cdb[1] = UFSHPB_WRITE_BUFFER_PREFETCH_ID;
> +
> +	put_unaligned_be32(lpn, &cdb[2]);
> +	cdb[6] = read_id;
> +	put_unaligned_be16(len * HPB_ENTRY_SIZE, &cdb[7]);
> +
> +	cdb[9] = 0x00;	/* Control = 0x00 */
> +}
> +
> +static struct ufshpb_req *ufshpb_get_pre_req(struct ufshpb_lu *hpb)
> +{
> +	struct ufshpb_req *pre_req;
> +
> +	if (hpb->num_inflight_pre_req >= hpb->throttle_pre_req) {
> +		dev_info(&hpb->sdev_ufs_lu->sdev_dev,
> +			 "pre_req throttle. inflight %d throttle %d",
> +			 hpb->num_inflight_pre_req, hpb->throttle_pre_req);
> +		return NULL;
> +	}
> +
> +	pre_req = list_first_entry_or_null(&hpb->lh_pre_req_free,
> +					   struct ufshpb_req, list_req);
> +	if (!pre_req) {
> +		dev_info(&hpb->sdev_ufs_lu->sdev_dev, "There is no pre_req");
> +		return NULL;
> +	}
> +
> +	list_del_init(&pre_req->list_req);
> +	hpb->num_inflight_pre_req++;
> +
> +	return pre_req;
> +}
> +
> +static inline void ufshpb_put_pre_req(struct ufshpb_lu *hpb,
> +				      struct ufshpb_req *pre_req)
> +{
> +	pre_req->req = NULL;
> +	pre_req->bio = NULL;
> +	list_add_tail(&pre_req->list_req, &hpb->lh_pre_req_free);
> +	hpb->num_inflight_pre_req--;
> +}
> +
> +static void ufshpb_pre_req_compl_fn(struct request *req, blk_status_t 
> error)
> +{
> +	struct ufshpb_req *pre_req = (struct ufshpb_req *)req->end_io_data;
> +	struct ufshpb_lu *hpb = pre_req->hpb;
> +	unsigned long flags;
> +	struct scsi_sense_hdr sshdr;
> +
> +	if (error) {
> +		dev_err(&hpb->sdev_ufs_lu->sdev_dev, "block status %d", error);
> +		scsi_normalize_sense(pre_req->sense, SCSI_SENSE_BUFFERSIZE,
> +				     &sshdr);
> +		dev_err(&hpb->sdev_ufs_lu->sdev_dev,
> +			"code %x sense_key %x asc %x ascq %x",
> +			sshdr.response_code,
> +			sshdr.sense_key, sshdr.asc, sshdr.ascq);
> +		dev_err(&hpb->sdev_ufs_lu->sdev_dev,
> +			"byte4 %x byte5 %x byte6 %x additional_len %x",
> +			sshdr.byte4, sshdr.byte5,
> +			sshdr.byte6, sshdr.additional_length);
> +	}
> +
> +	bio_put(pre_req->bio);
> +	blk_mq_free_request(req);
> +	spin_lock_irqsave(&hpb->rgn_state_lock, flags);
> +	ufshpb_put_pre_req(pre_req->hpb, pre_req);
> +	spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> +}
> +
> +static int ufshpb_prep_entry(struct ufshpb_req *pre_req, struct page 
> *page)
> +{
> +	struct ufshpb_lu *hpb = pre_req->hpb;
> +	struct ufshpb_region *rgn;
> +	struct ufshpb_subregion *srgn;
> +	u64 *addr;
> +	int offset = 0;
> +	int copied;
> +	unsigned long lpn = pre_req->wb.lpn;
> +	int rgn_idx, srgn_idx, srgn_offset;
> +	unsigned long flags;
> +
> +	addr = page_address(page);
> +	ufshpb_get_pos_from_lpn(hpb, lpn, &rgn_idx, &srgn_idx, &srgn_offset);
> +
> +	spin_lock_irqsave(&hpb->rgn_state_lock, flags);
> +
> +next_offset:
> +	rgn = hpb->rgn_tbl + rgn_idx;
> +	srgn = rgn->srgn_tbl + srgn_idx;
> +
> +	if (!ufshpb_is_valid_srgn(rgn, srgn))
> +		goto mctx_error;
> +
> +	if (!srgn->mctx)
> +		goto mctx_error;
> +
> +	copied = ufshpb_fill_ppn_from_page(hpb, srgn->mctx, srgn_offset,
> +					   pre_req->wb.len - offset,
> +					   &addr[offset]);
> +
> +	if (copied < 0)
> +		goto mctx_error;
> +
> +	offset += copied;
> +	srgn_offset += copied;
> +
> +	if (srgn_offset == hpb->entries_per_srgn) {
> +		srgn_offset = 0;
> +
> +		if (++srgn_idx == hpb->srgns_per_rgn) {
> +			srgn_idx = 0;
> +			rgn_idx++;
> +		}
> +	}
> +
> +	if (offset < pre_req->wb.len)
> +		goto next_offset;
> +
> +	spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> +	return 0;
> +mctx_error:
> +	spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> +	return -ENOMEM;
> +}
> +
> +static int ufshpb_pre_req_add_bio_page(struct ufshpb_lu *hpb,
> +				       struct request_queue *q,
> +				       struct ufshpb_req *pre_req)
> +{
> +	struct page *page = pre_req->wb.m_page;
> +	struct bio *bio = pre_req->bio;
> +	int entries_bytes, ret;
> +
> +	if (!page)
> +		return -ENOMEM;
> +
> +	if (ufshpb_prep_entry(pre_req, page))
> +		return -ENOMEM;
> +
> +	entries_bytes = pre_req->wb.len * sizeof(u64);
> +
> +	ret = bio_add_pc_page(q, bio, page, entries_bytes, 0);
> +	if (ret != entries_bytes) {
> +		dev_err(&hpb->sdev_ufs_lu->sdev_dev,
> +			"bio_add_pc_page fail: %d", ret);
> +		return -ENOMEM;
> +	}
> +	return 0;
> +}
> +
> +static inline int ufshpb_get_read_id(struct ufshpb_lu *hpb)
> +{
> +	if (++hpb->cur_read_id >= MAX_HPB_READ_ID)
> +		hpb->cur_read_id = 0;
> +	return hpb->cur_read_id;
> +}
> +
> +static int ufshpb_execute_pre_req(struct ufshpb_lu *hpb, struct 
> scsi_cmnd *cmd,
> +				  struct ufshpb_req *pre_req, int read_id)
> +{
> +	struct scsi_device *sdev = cmd->device;
> +	struct request_queue *q = sdev->request_queue;
> +	struct request *req;
> +	struct scsi_request *rq;
> +	struct bio *bio = pre_req->bio;
> +
> +	pre_req->hpb = hpb;
> +	pre_req->wb.lpn = sectors_to_logical(cmd->device,
> +					     blk_rq_pos(cmd->request));
> +	pre_req->wb.len = sectors_to_logical(cmd->device,
> +					     blk_rq_sectors(cmd->request));
> +	if (ufshpb_pre_req_add_bio_page(hpb, q, pre_req))
> +		return -ENOMEM;
> +
> +	req = pre_req->req;
> +
> +	/* 1. request setup */
> +	blk_rq_append_bio(req, &bio);
> +	req->rq_disk = NULL;
> +	req->end_io_data = (void *)pre_req;
> +	req->end_io = ufshpb_pre_req_compl_fn;
> +
> +	/* 2. scsi_request setup */
> +	rq = scsi_req(req);
> +	rq->retries = 1;
> +
> +	ufshpb_set_write_buf_cmd(rq->cmd, pre_req->wb.lpn, pre_req->wb.len,
> +				 read_id);
> +	rq->cmd_len = scsi_command_size(rq->cmd);
> +
> +	if (blk_insert_cloned_request(q, req) != BLK_STS_OK)
> +		return -EAGAIN;
> +
> +	hpb->stats.pre_req_cnt++;
> +
> +	return 0;
> +}
> +
> +static int ufshpb_issue_pre_req(struct ufshpb_lu *hpb, struct 
> scsi_cmnd *cmd,
> +				int *read_id)
> +{
> +	struct ufshpb_req *pre_req;
> +	struct request *req = NULL;
> +	struct bio *bio = NULL;
> +	unsigned long flags;
> +	int _read_id;
> +	int ret = 0;
> +
> +	req = blk_get_request(cmd->device->request_queue,
> +			      REQ_OP_SCSI_OUT | REQ_SYNC, BLK_MQ_REQ_NOWAIT);
> +	if (IS_ERR(req))
> +		return -EAGAIN;
> +
> +	bio = bio_alloc(GFP_ATOMIC, 1);
> +	if (!bio) {
> +		blk_put_request(req);
> +		return -EAGAIN;
> +	}
> +
> +	spin_lock_irqsave(&hpb->rgn_state_lock, flags);
> +	pre_req = ufshpb_get_pre_req(hpb);
> +	if (!pre_req) {
> +		ret = -EAGAIN;
> +		goto unlock_out;
> +	}
> +	_read_id = ufshpb_get_read_id(hpb);
> +	spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> +
> +	pre_req->req = req;
> +	pre_req->bio = bio;
> +
> +	ret = ufshpb_execute_pre_req(hpb, cmd, pre_req, _read_id);
> +	if (ret)
> +		goto free_pre_req;
> +
> +	*read_id = _read_id;
> +
> +	return ret;
> +free_pre_req:
> +	spin_lock_irqsave(&hpb->rgn_state_lock, flags);
> +	ufshpb_put_pre_req(hpb, pre_req);
> +unlock_out:
> +	spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> +	bio_put(bio);
> +	blk_put_request(req);
> +	return ret;
> +}
> +
>  /*
>   * This function will set up HPB read command using host-side L2P map 
> data.
> - * In HPB v1.0, maximum size of HPB read command is 4KB.
>   */
> -void ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp)
> +int ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp)
>  {
>  	struct ufshpb_lu *hpb;
>  	struct ufshpb_region *rgn;
> @@ -291,26 +560,27 @@ void ufshpb_prep(struct ufs_hba *hba, struct
> ufshcd_lrb *lrbp)
>  	u64 ppn;
>  	unsigned long flags;
>  	int transfer_len, rgn_idx, srgn_idx, srgn_offset;
> +	int read_id = 0;
>  	int err = 0;
> 
>  	hpb = ufshpb_get_hpb_data(cmd->device);
>  	if (!hpb)
> -		return;
> +		return -ENODEV;
> 
>  	if (ufshpb_get_state(hpb) != HPB_PRESENT) {
>  		dev_notice(&hpb->sdev_ufs_lu->sdev_dev,
>  			   "%s: ufshpb state is not PRESENT", __func__);
> -		return;
> +		return -ENODEV;
>  	}
> 
>  	if (!ufshpb_is_write_or_discard_cmd(cmd) &&
>  	    !ufshpb_is_read_cmd(cmd))
> -		return;
> +		return 0;
> 
>  	transfer_len = sectors_to_logical(cmd->device,
>  					  blk_rq_sectors(cmd->request));
>  	if (unlikely(!transfer_len))
> -		return;
> +		return 0;
> 
>  	lpn = sectors_to_logical(cmd->device, blk_rq_pos(cmd->request));
>  	ufshpb_get_pos_from_lpn(hpb, lpn, &rgn_idx, &srgn_idx, &srgn_offset);
> @@ -323,18 +593,19 @@ void ufshpb_prep(struct ufs_hba *hba, struct
> ufshcd_lrb *lrbp)
>  		ufshpb_set_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset,
>  				 transfer_len);
>  		spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> -		return;
> +		return 0;
>  	}
> 
> -	if (!ufshpb_is_support_chunk(transfer_len))
> -		return;
> +	if (!ufshpb_is_support_chunk(hpb, transfer_len) &&
> +	    (ufshpb_is_legacy(hba) && (transfer_len != 
> HPB_LEGACY_CHUNK_HIGH)))
> +		return 0;
> 
>  	spin_lock_irqsave(&hpb->rgn_state_lock, flags);
>  	if (ufshpb_test_ppn_dirty(hpb, rgn_idx, srgn_idx, srgn_offset,
>  				   transfer_len)) {
>  		hpb->stats.miss_cnt++;
>  		spin_unlock_irqrestore(&hpb->rgn_state_lock, flags);
> -		return;
> +		return 0;
>  	}
> 
>  	err = ufshpb_fill_ppn_from_page(hpb, srgn->mctx, srgn_offset, 1, 
> &ppn);
> @@ -347,28 +618,46 @@ void ufshpb_prep(struct ufs_hba *hba, struct
> ufshcd_lrb *lrbp)
>  		 * active state.
>  		 */
>  		dev_err(hba->dev, "get ppn failed. err %d\n", err);
> -		return;
> +		return err;
> +	}
> +
> +	if (!ufshpb_is_legacy(hba) &&
> +	    ufshpb_is_required_wb(hpb, transfer_len)) {
> +		err = ufshpb_issue_pre_req(hpb, cmd, &read_id);
> +		if (err) {
> +			unsigned long timeout;
> +
> +			timeout = cmd->jiffies_at_alloc + msecs_to_jiffies(
> +				  hpb->params.requeue_timeout_ms);
> +
> +			if (time_before(jiffies, timeout))
> +				return -EAGAIN;
> +
> +			hpb->stats.miss_cnt++;
> +			return 0;
> +		}
>  	}
> 
> -	ufshpb_set_hpb_read_to_upiu(hpb, lrbp, lpn, ppn, transfer_len);
> +	ufshpb_set_hpb_read_to_upiu(hpb, lrbp, lpn, ppn, transfer_len, 
> read_id);
> 
>  	hpb->stats.hit_cnt++;
> +	return 0;
>  }
> -static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb,
> -					     struct ufshpb_subregion *srgn)
> +
> +static struct ufshpb_req *ufshpb_get_req(struct ufshpb_lu *hpb,
> +					 int rgn_idx, enum req_opf dir)
>  {
> -	struct ufshpb_req *map_req;
> +	struct ufshpb_req *rq;
>  	struct request *req;
> -	struct bio *bio;
>  	int retries = HPB_MAP_REQ_RETRIES;
> 
> -	map_req = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL);
> -	if (!map_req)
> +	rq = kmem_cache_alloc(hpb->map_req_cache, GFP_KERNEL);
> +	if (!rq)
>  		return NULL;
> 
>  retry:
> -	req = blk_get_request(hpb->sdev_ufs_lu->request_queue,
> -			      REQ_OP_SCSI_IN, BLK_MQ_REQ_NOWAIT);
> +	req = blk_get_request(hpb->sdev_ufs_lu->request_queue, dir,
> +			      BLK_MQ_REQ_NOWAIT);
> 
>  	if ((PTR_ERR(req) == -EWOULDBLOCK) && (--retries > 0)) {
>  		usleep_range(3000, 3100);
> @@ -376,35 +665,54 @@ static struct ufshpb_req
> *ufshpb_get_map_req(struct ufshpb_lu *hpb,
>  	}
> 
>  	if (IS_ERR(req))
> -		goto free_map_req;
> +		goto free_rq;
> +
> +	rq->hpb = hpb;
> +	rq->req = req;
> +	rq->rb.rgn_idx = rgn_idx;
> +
> +	return rq;
> +
> +free_rq:
> +	kmem_cache_free(hpb->map_req_cache, rq);
> +	return NULL;
> +}
> +
> +static void ufshpb_put_req(struct ufshpb_lu *hpb, struct ufshpb_req 
> *rq)
> +{
> +	blk_put_request(rq->req);
> +	kmem_cache_free(hpb->map_req_cache, rq);
> +}
> +
> +static struct ufshpb_req *ufshpb_get_map_req(struct ufshpb_lu *hpb,
> +					     struct ufshpb_subregion *srgn)
> +{
> +	struct ufshpb_req *map_req;
> +	struct bio *bio;
> +
> +	map_req = ufshpb_get_req(hpb, srgn->rgn_idx, REQ_OP_SCSI_IN);
> +	if (!map_req)
> +		return NULL;
> 
>  	bio = bio_alloc(GFP_KERNEL, hpb->pages_per_srgn);
>  	if (!bio) {
> -		blk_put_request(req);
> -		goto free_map_req;
> +		ufshpb_put_req(hpb, map_req);
> +		return NULL;
>  	}
> 
> -	map_req->hpb = hpb;
> -	map_req->req = req;
>  	map_req->bio = bio;
> 
> -	map_req->rgn_idx = srgn->rgn_idx;
> -	map_req->srgn_idx = srgn->srgn_idx;
> -	map_req->mctx = srgn->mctx;
> +	map_req->rb.srgn_idx = srgn->srgn_idx;
> +	map_req->rb.mctx = srgn->mctx;
> 
>  	return map_req;
> -
> -free_map_req:
> -	kmem_cache_free(hpb->map_req_cache, map_req);
> -	return NULL;
>  }
> 
>  static void ufshpb_put_map_req(struct ufshpb_lu *hpb,
>  			       struct ufshpb_req *map_req)
>  {
>  	bio_put(map_req->bio);
> -	blk_put_request(map_req->req);
> -	kmem_cache_free(hpb->map_req_cache, map_req);
> +	ufshpb_put_req(hpb, map_req);
>  }
> 
>  static int ufshpb_clear_dirty_bitmap(struct ufshpb_lu *hpb,
> @@ -487,6 +795,13 @@ static void ufshpb_activate_subregion(struct
> ufshpb_lu *hpb,
>  	srgn->srgn_state = HPB_SRGN_VALID;
>  }
> 
> +static void ufshpb_umap_req_compl_fn(struct request *req, blk_status_t 
> error)
> +{
> +	struct ufshpb_req *umap_req = (struct ufshpb_req *)req->end_io_data;
> +
> +	ufshpb_put_req(umap_req->hpb, umap_req);
> +}
> +
>  static void ufshpb_map_req_compl_fn(struct request *req, blk_status_t 
> error)
>  {
>  	struct ufshpb_req *map_req = (struct ufshpb_req *) req->end_io_data;
> @@ -494,8 +809,8 @@ static void ufshpb_map_req_compl_fn(struct request
> *req, blk_status_t error)
>  	struct ufshpb_subregion *srgn;
>  	unsigned long flags;
> 
> -	srgn = hpb->rgn_tbl[map_req->rgn_idx].srgn_tbl +
> -		map_req->srgn_idx;
> +	srgn = hpb->rgn_tbl[map_req->rb.rgn_idx].srgn_tbl +
> +		map_req->rb.srgn_idx;
> 
>  	ufshpb_clear_dirty_bitmap(hpb, srgn);
>  	spin_lock_irqsave(&hpb->rgn_state_lock, flags);
> @@ -505,6 +820,16 @@ static void ufshpb_map_req_compl_fn(struct
> request *req, blk_status_t error)
>  	ufshpb_put_map_req(map_req->hpb, map_req);
>  }
> 
> +static void ufshpb_set_unmap_cmd(unsigned char *cdb, struct 
> ufshpb_region *rgn)
> +{
> +	cdb[0] = UFSHPB_WRITE_BUFFER;
> +	cdb[1] = rgn ? UFSHPB_WRITE_BUFFER_INACT_SINGLE_ID :
> +			  UFSHPB_WRITE_BUFFER_INACT_ALL_ID;
> +	if (rgn)
> +		put_unaligned_be16(rgn->rgn_idx, &cdb[2]);
> +	cdb[9] = 0x00;
> +}
> +
>  static void ufshpb_set_read_buf_cmd(unsigned char *cdb, int rgn_idx,
>  				    int srgn_idx, int srgn_mem_size)
>  {
> @@ -518,6 +843,27 @@ static void ufshpb_set_read_buf_cmd(unsigned char
> *cdb, int rgn_idx,
>  	cdb[9] = 0x00;
>  }
> 
> +static int ufshpb_execute_umap_req(struct ufshpb_lu *hpb,
> +				   struct ufshpb_req *umap_req,
> +				   struct ufshpb_region *rgn)
> +{
> +	struct request_queue *q;
> +	struct request *req;
> +	struct scsi_request *rq;
> +
> +	q = hpb->sdev_ufs_lu->request_queue;
> +	req = umap_req->req;
> +	req->timeout = 0;
> +	req->end_io_data = (void *)umap_req;
> +	rq = scsi_req(req);
> +	ufshpb_set_unmap_cmd(rq->cmd, rgn);
> +	rq->cmd_len = HPB_WRITE_BUFFER_CMD_LENGTH;
> +
> +	blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_umap_req_compl_fn);
> +
> +	return 0;
> +}
> +
>  static int ufshpb_execute_map_req(struct ufshpb_lu *hpb,
>  				  struct ufshpb_req *map_req, bool last)
>  {
> @@ -530,12 +876,12 @@ static int ufshpb_execute_map_req(struct 
> ufshpb_lu *hpb,
> 
>  	q = hpb->sdev_ufs_lu->request_queue;
>  	for (i = 0; i < hpb->pages_per_srgn; i++) {
> -		ret = bio_add_pc_page(q, map_req->bio, map_req->mctx->m_page[i],
> +		ret = bio_add_pc_page(q, map_req->bio, map_req->rb.mctx->m_page[i],
>  				      PAGE_SIZE, 0);
>  		if (ret != PAGE_SIZE) {
>  			dev_err(&hpb->sdev_ufs_lu->sdev_dev,
>  				   "bio_add_pc_page fail %d - %d\n",
> -				   map_req->rgn_idx, map_req->srgn_idx);
> +				   map_req->rb.rgn_idx, map_req->rb.srgn_idx);
>  			return ret;
>  		}
>  	}
> @@ -551,8 +897,8 @@ static int ufshpb_execute_map_req(struct ufshpb_lu 
> *hpb,
>  	if (unlikely(last))
>  		mem_size = hpb->last_srgn_entries * HPB_ENTRY_SIZE;
> 
> -	ufshpb_set_read_buf_cmd(rq->cmd, map_req->rgn_idx,
> -				map_req->srgn_idx, mem_size);
> +	ufshpb_set_read_buf_cmd(rq->cmd, map_req->rb.rgn_idx,
> +				map_req->rb.srgn_idx, mem_size);
>  	rq->cmd_len = HPB_READ_BUFFER_CMD_LENGTH;
> 
>  	blk_execute_rq_nowait(q, NULL, req, 1, ufshpb_map_req_compl_fn);
> @@ -684,6 +1030,31 @@ static void ufshpb_purge_active_subregion(struct
> ufshpb_lu *hpb,
>  	}
>  }
> 
> +static int ufshpb_issue_umap_req(struct ufshpb_lu *hpb,
> +				 struct ufshpb_region *rgn)
> +{
> +	struct ufshpb_req *umap_req;
> +	int rgn_idx = rgn ? rgn->rgn_idx : 0;
> +
> +	umap_req = ufshpb_get_req(hpb, rgn_idx, REQ_OP_SCSI_OUT);
> +	if (!umap_req)
> +		return -ENOMEM;
> +
> +	if (ufshpb_execute_umap_req(hpb, umap_req, rgn))
> +		goto free_umap_req;
> +
> +	return 0;
> +
> +free_umap_req:
> +	ufshpb_put_req(hpb, umap_req);
> +	return -EAGAIN;
> +}
> +
> +static int ufshpb_issue_umap_all_req(struct ufshpb_lu *hpb)
> +{
> +	return ufshpb_issue_umap_req(hpb, NULL);
> +}
> +
>  static void __ufshpb_evict_region(struct ufshpb_lu *hpb,
>  				  struct ufshpb_region *rgn)
>  {
> @@ -1209,6 +1580,13 @@ static void ufshpb_lu_parameter_init(struct 
> ufs_hba *hba,
>  	u32 entries_per_rgn;
>  	u64 rgn_mem_size, tmp;
> 
> +	/* for pre_req */
> +	hpb->pre_req_min_tr_len = hpb_dev_info->max_hpb_single_cmd + 1;
> +	hpb->pre_req_max_tr_len = max(HPB_MULTI_CHUNK_HIGH,
> +				      hpb->pre_req_min_tr_len);
> +
> +	hpb->cur_read_id = 0;
> +
>  	hpb->lu_pinned_start = hpb_lu_info->pinned_start;
>  	hpb->lu_pinned_end = hpb_lu_info->num_pinned ?
>  		(hpb_lu_info->pinned_start + hpb_lu_info->num_pinned - 1)
> @@ -1356,7 +1734,7 @@ ufshpb_sysfs_attr_show_func(rb_active_cnt);
>  ufshpb_sysfs_attr_show_func(rb_inactive_cnt);
>  ufshpb_sysfs_attr_show_func(map_req_cnt);
> 
> -static struct attribute *hpb_dev_attrs[] = {
> +static struct attribute *hpb_dev_stat_attrs[] = {
>  	&dev_attr_hit_cnt.attr,
>  	&dev_attr_miss_cnt.attr,
>  	&dev_attr_rb_noti_cnt.attr,
> @@ -1367,10 +1745,109 @@ static struct attribute *hpb_dev_attrs[] = {
>  };
> 
>  struct attribute_group ufs_sysfs_hpb_stat_group = {
> -	.name = "hpb_sysfs",
> -	.attrs = hpb_dev_attrs,
> +	.name = "hpb_stat_sysfs",
> +	.attrs = hpb_dev_stat_attrs,
> +};
> +
> +/* SYSFS functions */
> +#define ufshpb_sysfs_param_show_func(__name)				\
> +static ssize_t __name##_show(struct device *dev,			\
> +	struct device_attribute *attr, char *buf)			\
> +{									\
> +	struct scsi_device *sdev = to_scsi_device(dev);			\
> +	struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);		\
> +	if (!hpb)							\
> +		return -ENODEV;						\
> +									\
> +	return sysfs_emit(buf, "%d\n", hpb->params.__name);		\
> +}
> +
> +ufshpb_sysfs_param_show_func(requeue_timeout_ms);
> +static ssize_t
> +requeue_timeout_ms_store(struct device *dev, struct device_attribute 
> *attr,
> +			 const char *buf, size_t count)
> +{
> +	struct scsi_device *sdev = to_scsi_device(dev);
> +	struct ufshpb_lu *hpb = ufshpb_get_hpb_data(sdev);
> +	int val;
> +
> +	if (!hpb)
> +		return -ENODEV;
> +
> +	if (kstrtouint(buf, 0, &val))
> +		return -EINVAL;
> +
> +	if (val <= 0)
> +		return -EINVAL;
> +
> +	hpb->params.requeue_timeout_ms = val;
> +
> +	return count;
> +}
> +static DEVICE_ATTR_RW(requeue_timeout_ms);
> +
> +static struct attribute *hpb_dev_param_attrs[] = {
> +	&dev_attr_requeue_timeout_ms.attr,
> +	NULL,
> +};
> +
> +struct attribute_group ufs_sysfs_hpb_param_group = {
> +	.name = "hpb_param_sysfs",
> +	.attrs = hpb_dev_param_attrs,
>  };
> 
> +static int ufshpb_pre_req_mempool_init(struct ufshpb_lu *hpb)
> +{
> +	struct ufshpb_req *pre_req = NULL;
> +	int qd = hpb->sdev_ufs_lu->queue_depth / 2;
> +	int i, j;
> +
> +	INIT_LIST_HEAD(&hpb->lh_pre_req_free);
> +
> +	hpb->pre_req = kcalloc(qd, sizeof(struct ufshpb_req), GFP_KERNEL);
> +	hpb->throttle_pre_req = qd;
> +	hpb->num_inflight_pre_req = 0;
> +
> +	if (!hpb->pre_req)
> +		goto release_mem;
> +
> +	for (i = 0; i < qd; i++) {
> +		pre_req = hpb->pre_req + i;
> +		INIT_LIST_HEAD(&pre_req->list_req);
> +		pre_req->req = NULL;
> +		pre_req->bio = NULL;
> +
> +		pre_req->wb.m_page = alloc_page(GFP_KERNEL | __GFP_ZERO);
> +		if (!pre_req->wb.m_page) {
> +			for (j = 0; j < i; j++)
> +				__free_page(hpb->pre_req[j].wb.m_page);
> +
> +			goto release_mem;
> +		}
> +		list_add_tail(&pre_req->list_req, &hpb->lh_pre_req_free);
> +	}
> +
> +	return 0;
> +release_mem:
> +	kfree(hpb->pre_req);
> +	return -ENOMEM;
> +}
> +
> +static void ufshpb_pre_req_mempool_destroy(struct ufshpb_lu *hpb)
> +{
> +	struct ufshpb_req *pre_req = NULL;
> +	int i;
> +
> +	for (i = 0; i < hpb->throttle_pre_req; i++) {
> +		pre_req = hpb->pre_req + i;
> +		if (!pre_req->wb.m_page)
> +			__free_page(hpb->pre_req[i].wb.m_page);
> +		list_del_init(&pre_req->list_req);
> +	}
> +
> +	kfree(hpb->pre_req);
> +}
> +
>  static void ufshpb_stat_init(struct ufshpb_lu *hpb)
>  {
>  	hpb->stats.hit_cnt = 0;
> @@ -1381,6 +1858,11 @@ static void ufshpb_stat_init(struct ufshpb_lu 
> *hpb)
>  	hpb->stats.map_req_cnt = 0;
>  }
> 
> +static void ufshpb_param_init(struct ufshpb_lu *hpb)
> +{
> +	hpb->params.requeue_timeout_ms = HPB_REQUEUE_TIME_MS;
> +}
> +
>  static int ufshpb_lu_hpb_init(struct ufs_hba *hba, struct ufshpb_lu 
> *hpb)
>  {
>  	int ret;
> @@ -1413,14 +1895,24 @@ static int ufshpb_lu_hpb_init(struct ufs_hba
> *hba, struct ufshpb_lu *hpb)
>  		goto release_req_cache;
>  	}
> 
> +	ret = ufshpb_pre_req_mempool_init(hpb);
> +	if (ret) {
> +		dev_err(hba->dev, "ufshpb(%d) pre_req_mempool init fail",
> +			hpb->lun);
> +		goto release_m_page_cache;
> +	}
> +
>  	ret = ufshpb_alloc_region_tbl(hba, hpb);
>  	if (ret)
> -		goto release_m_page_cache;
> +		goto release_pre_req_mempool;
> 
>  	ufshpb_stat_init(hpb);
> +	ufshpb_param_init(hpb);
> 
>  	return 0;
> 
> +release_pre_req_mempool:
> +	ufshpb_pre_req_mempool_destroy(hpb);
>  release_m_page_cache:
>  	kmem_cache_destroy(hpb->m_page_cache);
>  release_req_cache:
> @@ -1429,7 +1921,7 @@ static int ufshpb_lu_hpb_init(struct ufs_hba
> *hba, struct ufshpb_lu *hpb)
>  }
> 
>  static struct ufshpb_lu *
> -ufshpb_alloc_hpb_lu(struct ufs_hba *hba, int lun,
> +ufshpb_alloc_hpb_lu(struct ufs_hba *hba, struct scsi_device *sdev,
>  		    struct ufshpb_dev_info *hpb_dev_info,
>  		    struct ufshpb_lu_info *hpb_lu_info)
>  {
> @@ -1440,7 +1932,8 @@ ufshpb_alloc_hpb_lu(struct ufs_hba *hba, int lun,
>  	if (!hpb)
>  		return NULL;
> 
> -	hpb->lun = lun;
> +	hpb->lun = sdev->lun;
> +	hpb->sdev_ufs_lu = sdev;
> 
>  	ufshpb_lu_parameter_init(hba, hpb, hpb_dev_info, hpb_lu_info);
> 
> @@ -1450,6 +1943,7 @@ ufshpb_alloc_hpb_lu(struct ufs_hba *hba, int lun,
>  		goto release_hpb;
>  	}
> 
> +	sdev->hostdata = hpb;
>  	return hpb;
> 
>  release_hpb:
> @@ -1652,6 +2146,7 @@ void ufshpb_destroy_lu(struct ufs_hba *hba,
> struct scsi_device *sdev)
> 
>  	ufshpb_cancel_jobs(hpb);
> 
> +	ufshpb_pre_req_mempool_destroy(hpb);
>  	ufshpb_destroy_region_tbl(hpb);
> 
>  	kmem_cache_destroy(hpb->map_req_cache);
> @@ -1691,6 +2186,7 @@ static void ufshpb_hpb_lu_prepared(struct ufs_hba 
> *hba)
>  			ufshpb_set_state(hpb, HPB_PRESENT);
>  			if ((hpb->lu_pinned_end - hpb->lu_pinned_start) > 0)
>  				queue_work(ufshpb_wq, &hpb->map_work);
> +			ufshpb_issue_umap_all_req(hpb);
>  		} else {
>  			dev_err(hba->dev, "destroy HPB lu %d\n", hpb->lun);
>  			ufshpb_destroy_lu(hba, sdev);
> @@ -1715,7 +2211,7 @@ void ufshpb_init_hpb_lu(struct ufs_hba *hba,
> struct scsi_device *sdev)
>  	if (ret)
>  		goto out;
> 
> -	hpb = ufshpb_alloc_hpb_lu(hba, lun, &hba->ufshpb_dev,
> +	hpb = ufshpb_alloc_hpb_lu(hba, sdev, &hba->ufshpb_dev,
>  				  &hpb_lu_info);
>  	if (!hpb)
>  		goto out;
> @@ -1723,9 +2219,6 @@ void ufshpb_init_hpb_lu(struct ufs_hba *hba,
> struct scsi_device *sdev)
>  	tot_active_srgn_pages += hpb_lu_info.max_active_rgns *
>  			hpb->srgns_per_rgn * hpb->pages_per_srgn;
> 
> -	hpb->sdev_ufs_lu = sdev;
> -	sdev->hostdata = hpb;
> -
>  out:
>  	/* All LUs are initialized */
>  	if (atomic_dec_and_test(&hba->ufshpb_dev.slave_conf_cnt))
> @@ -1812,8 +2305,9 @@ void ufshpb_get_geo_info(struct ufs_hba *hba, u8 
> *geo_buf)
>  void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf)
>  {
>  	struct ufshpb_dev_info *hpb_dev_info = &hba->ufshpb_dev;
> -	int version;
> +	int version, ret;
>  	u8 hpb_mode;
> +	u32 max_hpb_single_cmd = 0;
> 
>  	hpb_mode = desc_buf[DEVICE_DESC_PARAM_HPB_CONTROL];
>  	if (hpb_mode == HPB_HOST_CONTROL) {
> @@ -1824,13 +2318,27 @@ void ufshpb_get_dev_info(struct ufs_hba *hba,
> u8 *desc_buf)
>  	}
> 
>  	version = get_unaligned_be16(desc_buf + DEVICE_DESC_PARAM_HPB_VER);
> -	if (version != HPB_SUPPORT_VERSION) {
> +	if ((version != HPB_SUPPORT_VERSION) &&
> +	    (version != HPB_SUPPORT_LEGACY_VERSION)) {
>  		dev_err(hba->dev, "%s: HPB %x version is not supported.\n",
>  			__func__, version);
>  		hpb_dev_info->hpb_disabled = true;
>  		return;
>  	}
> 
> +	if (version == HPB_SUPPORT_LEGACY_VERSION)
> +		hpb_dev_info->is_legacy = true;
> +
> +	pm_runtime_get_sync(hba->dev);
> +	ret = ufshcd_query_attr_retry(hba, UPIU_QUERY_OPCODE_READ_ATTR,
> +		QUERY_ATTR_IDN_MAX_HPB_SINGLE_CMD, 0, 0, &max_hpb_single_cmd);
> +	pm_runtime_put_sync(hba->dev);
> +
> +	if (ret)
> +		dev_err(hba->dev, "%s: idn: read max size of single hpb cmd query
> request failed",
> +			__func__);
> +	hpb_dev_info->max_hpb_single_cmd = max_hpb_single_cmd;
> +
>  	/*
>  	 * Get the number of user logical unit to check whether all
>  	 * scsi_device finish initialization
> diff --git a/drivers/scsi/ufs/ufshpb.h b/drivers/scsi/ufs/ufshpb.h
> index c70e73546e35..c319b8a0a0d9 100644
> --- a/drivers/scsi/ufs/ufshpb.h
> +++ b/drivers/scsi/ufs/ufshpb.h
> @@ -30,19 +30,28 @@
>  #define PINNED_NOT_SET				U32_MAX
> 
>  /* hpb support chunk size */
> -#define HPB_MULTI_CHUNK_HIGH			1
> +#define HPB_LEGACY_CHUNK_HIGH			1
> +#define HPB_MULTI_CHUNK_HIGH			128
> 
>  /* hpb vender defined opcode */
>  #define UFSHPB_READ				0xF8
>  #define UFSHPB_READ_BUFFER			0xF9
>  #define UFSHPB_READ_BUFFER_ID			0x01
> +#define UFSHPB_WRITE_BUFFER			0xFA
> +#define UFSHPB_WRITE_BUFFER_INACT_SINGLE_ID	0x01
> +#define UFSHPB_WRITE_BUFFER_PREFETCH_ID		0x02
> +#define UFSHPB_WRITE_BUFFER_INACT_ALL_ID	0x03
> +#define HPB_WRITE_BUFFER_CMD_LENGTH		10
> +#define MAX_HPB_READ_ID				0x7F
>  #define HPB_READ_BUFFER_CMD_LENGTH		10
>  #define LU_ENABLED_HPB_FUNC			0x02
> 
>  #define HPB_RESET_REQ_RETRIES			10
>  #define HPB_MAP_REQ_RETRIES			5
> +#define HPB_REQUEUE_TIME_MS			0
> 
> -#define HPB_SUPPORT_VERSION			0x100
> +#define HPB_SUPPORT_VERSION			0x200
> +#define HPB_SUPPORT_LEGACY_VERSION		0x100
> 
>  enum UFSHPB_MODE {
>  	HPB_HOST_CONTROL,
> @@ -118,23 +127,39 @@ struct ufshpb_region {
>  	     (i)++)
> 
>  /**
> - * struct ufshpb_req - UFSHPB READ BUFFER (for caching map) request 
> structure
> - * @req: block layer request for READ BUFFER
> - * @bio: bio for holding map page
> - * @hpb: ufshpb_lu structure that related to the L2P map
> + * struct ufshpb_req - HPB related request structure (write/read 
> buffer)
> + * @req: block layer request structure
> + * @bio: bio for this request
> + * @hpb: ufshpb_lu structure that related to
> + * @list_req: ufshpb_req mempool list
> + * @sense: store its sense data
>   * @mctx: L2P map information
>   * @rgn_idx: target region index
>   * @srgn_idx: target sub-region index
>   * @lun: target logical unit number
> + * @m_page: L2P map information data for pre-request
> + * @len: length of host-side cached L2P map in m_page
> + * @lpn: start LPN of L2P map in m_page
>   */
>  struct ufshpb_req {
>  	struct request *req;
>  	struct bio *bio;
>  	struct ufshpb_lu *hpb;
> -	struct ufshpb_map_ctx *mctx;
> -
> -	unsigned int rgn_idx;
> -	unsigned int srgn_idx;
> +	struct list_head list_req;
> +	char sense[SCSI_SENSE_BUFFERSIZE];
> +	union {
> +		struct {
> +			struct ufshpb_map_ctx *mctx;
> +			unsigned int rgn_idx;
> +			unsigned int srgn_idx;
> +			unsigned int lun;
> +		} rb;
> +		struct {
> +			struct page *m_page;
> +			unsigned int len;
> +			unsigned long lpn;
> +		} wb;
> +	};
>  };
> 
>  struct victim_select_info {
> @@ -143,6 +168,10 @@ struct victim_select_info {
>  	atomic_t active_cnt;
>  };
> 
> +struct ufshpb_params {
> +	unsigned int requeue_timeout_ms;
> +};
> +
>  struct ufshpb_stats {
>  	u64 hit_cnt;
>  	u64 miss_cnt;
> @@ -150,6 +179,7 @@ struct ufshpb_stats {
>  	u64 rb_active_cnt;
>  	u64 rb_inactive_cnt;
>  	u64 map_req_cnt;
> +	u64 pre_req_cnt;
>  };
> 
>  struct ufshpb_lu {
> @@ -165,6 +195,15 @@ struct ufshpb_lu {
>  	struct list_head lh_act_srgn; /* hold rsp_list_lock */
>  	struct list_head lh_inact_rgn; /* hold rsp_list_lock */
> 
> +	/* pre request information */
> +	struct ufshpb_req *pre_req;
> +	int num_inflight_pre_req;
> +	int throttle_pre_req;
> +	struct list_head lh_pre_req_free;
> +	int cur_read_id;
> +	int pre_req_min_tr_len;
> +	int pre_req_max_tr_len;
> +
>  	/* cached L2P map management worker */
>  	struct work_struct map_work;
> 
> @@ -189,6 +228,7 @@ struct ufshpb_lu {
>  	u32 pages_per_srgn;
> 
>  	struct ufshpb_stats stats;
> +	struct ufshpb_params params;
> 
>  	struct kmem_cache *map_req_cache;
>  	struct kmem_cache *m_page_cache;
> @@ -200,7 +240,7 @@ struct ufs_hba;
>  struct ufshcd_lrb;
> 
>  #ifndef CONFIG_SCSI_UFS_HPB
> -static void ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp) 
> {}
> +static int ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp)
> { return 0; }
>  static void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb 
> *lrbp) {}
>  static void ufshpb_resume(struct ufs_hba *hba) {}
>  static void ufshpb_suspend(struct ufs_hba *hba) {}
> @@ -213,8 +253,9 @@ static void ufshpb_remove(struct ufs_hba *hba) {}
>  static bool ufshpb_is_allowed(struct ufs_hba *hba) { return false; }
>  static void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf) {}
>  static void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf) {}
> +static bool ufshpb_is_legacy(struct ufs_hba *hba) { return false; }
>  #else
> -void ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp);
> +int ufshpb_prep(struct ufs_hba *hba, struct ufshcd_lrb *lrbp);
>  void ufshpb_rsp_upiu(struct ufs_hba *hba, struct ufshcd_lrb *lrbp);
>  void ufshpb_resume(struct ufs_hba *hba);
>  void ufshpb_suspend(struct ufs_hba *hba);
> @@ -227,7 +268,9 @@ void ufshpb_remove(struct ufs_hba *hba);
>  bool ufshpb_is_allowed(struct ufs_hba *hba);
>  void ufshpb_get_geo_info(struct ufs_hba *hba, u8 *geo_buf);
>  void ufshpb_get_dev_info(struct ufs_hba *hba, u8 *desc_buf);
> +bool ufshpb_is_legacy(struct ufs_hba *hba);
>  extern struct attribute_group ufs_sysfs_hpb_stat_group;
> +extern struct attribute_group ufs_sysfs_hpb_param_group;
>  #endif
> 
>  #endif /* End of Header */

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ