lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Date:	Mon, 13 Jun 2016 18:16:51 +0000
From:	"Jiang, Dave" <dave.jiang@...el.com>
To:	"Allen.Hubbe@....com" <Allen.Hubbe@....com>,
	"logang@...tatee.com" <logang@...tatee.com>,
	"jdmason@...zu.us" <jdmason@...zu.us>
CC:	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	"shuahkh@....samsung.com" <shuahkh@....samsung.com>,
	"sudipm.mukherjee@...il.com" <sudipm.mukherjee@...il.com>,
	"linux-kselftest@...r.kernel.org" <linux-kselftest@...r.kernel.org>,
	"arnd@...db.de" <arnd@...db.de>,
	"linux-ntb@...glegroups.com" <linux-ntb@...glegroups.com>
Subject: Re: [PATCH 2/8] ntb_perf: Improve thread handling to increase
 robustness

On Fri, 2016-06-10 at 16:54 -0600, Logan Gunthorpe wrote:
> This commit accomplishes a few things:
> 
> 1) Properly prevent multiple sets of threads from running at once
> using
> a mutex. Lots of race issues existed with the thread_cleanup.
> 
> 2) The mutex allows us to ensure that threads are finished before
> tearing down the device or module.
> 
> 3) Don't use kthread_stop when the threads can exit by themselves, as
> this is counter-indicated by the kthread_create documentation.
> Threads
> now wait for kthread_stop to occur.
> 
> 4) Writing to the run file now blocks until the threads are complete.
> The test can then be safely interrupted by a SIGINT.
> 
> Also, while I was at it:
> 
> 5) debugfs_run_write shouldn't return 0 in the early check cases as
> this
> could cause debugfs_run_write to loop undesirably.
> 
> Signed-off-by: Logan Gunthorpe <logang@...tatee.com>
Acked-by: Dave Jiang <dave.jiang@...el.com>

> ---
>  drivers/ntb/test/ntb_perf.c | 124 +++++++++++++++++++++++++++-------
> ----------
>  1 file changed, 76 insertions(+), 48 deletions(-)
> 
> diff --git a/drivers/ntb/test/ntb_perf.c
> b/drivers/ntb/test/ntb_perf.c
> index 5008ccf..db4dc61 100644
> --- a/drivers/ntb/test/ntb_perf.c
> +++ b/drivers/ntb/test/ntb_perf.c
> @@ -58,6 +58,7 @@
>  #include <linux/delay.h>
>  #include <linux/sizes.h>
>  #include <linux/ntb.h>
> +#include <linux/mutex.h>
>  
>  #define DRIVER_NAME		"ntb_perf"
>  #define DRIVER_DESCRIPTION	"PCIe NTB Performance Measurement
> Tool"
> @@ -121,6 +122,7 @@ struct pthr_ctx {
>  	int			dma_prep_err;
>  	int			src_idx;
>  	void			*srcs[MAX_SRCS];
> +	wait_queue_head_t       *wq;
>  };
>  
>  struct perf_ctx {
> @@ -134,9 +136,11 @@ struct perf_ctx {
>  	struct dentry		*debugfs_run;
>  	struct dentry		*debugfs_threads;
>  	u8			perf_threads;
> -	bool			run;
> +	/* mutex ensures only one set of threads run at once */
> +	struct mutex		run_mutex;
>  	struct pthr_ctx		pthr_ctx[MAX_THREADS];
>  	atomic_t		tsync;
> +	atomic_t                tdone;
>  };
>  
>  enum {
> @@ -295,12 +299,18 @@ static int perf_move_data(struct pthr_ctx
> *pctx, char __iomem *dst, char *src,
>  			set_current_state(TASK_INTERRUPTIBLE);
>  			schedule_timeout(1);
>  		}
> +
> +		if (unlikely(kthread_should_stop()))
> +			break;
>  	}
>  
>  	if (use_dma) {
>  		pr_info("%s: All DMA descriptors submitted\n",
> current->comm);
> -		while (atomic_read(&pctx->dma_sync) != 0)
> +		while (atomic_read(&pctx->dma_sync) != 0) {
> +			if (kthread_should_stop())
> +				break;
>  			msleep(20);
> +		}
>  	}
>  
>  	kstop = ktime_get();
> @@ -393,7 +403,10 @@ static int ntb_perf_thread(void *data)
>  		pctx->srcs[i] = NULL;
>  	}
>  
> -	return 0;
> +	atomic_inc(&perf->tdone);
> +	wake_up(pctx->wq);
> +	rc = 0;
> +	goto done;
>  
>  err:
>  	for (i = 0; i < MAX_SRCS; i++) {
> @@ -406,6 +419,16 @@ err:
>  		pctx->dma_chan = NULL;
>  	}
>  
> +done:
> +	/* Wait until we are told to stop */
> +	for (;;) {
> +		set_current_state(TASK_INTERRUPTIBLE);
> +		if (kthread_should_stop())
> +			break;
> +		schedule();
> +	}
> +	__set_current_state(TASK_RUNNING);
> +
>  	return rc;
>  }
>  
> @@ -553,6 +576,7 @@ static ssize_t debugfs_run_read(struct file
> *filp, char __user *ubuf,
>  	struct perf_ctx *perf = filp->private_data;
>  	char *buf;
>  	ssize_t ret, out_offset;
> +	int running;
>  
>  	if (!perf)
>  		return 0;
> @@ -560,7 +584,9 @@ static ssize_t debugfs_run_read(struct file
> *filp, char __user *ubuf,
>  	buf = kmalloc(64, GFP_KERNEL);
>  	if (!buf)
>  		return -ENOMEM;
> -	out_offset = snprintf(buf, 64, "%d\n", perf->run);
> +
> +	running = mutex_is_locked(&perf->run_mutex);
> +	out_offset = snprintf(buf, 64, "%d\n", running);
>  	ret = simple_read_from_buffer(ubuf, count, offp, buf,
> out_offset);
>  	kfree(buf);
>  
> @@ -572,7 +598,6 @@ static void threads_cleanup(struct perf_ctx
> *perf)
>  	struct pthr_ctx *pctx;
>  	int i;
>  
> -	perf->run = false;
>  	for (i = 0; i < MAX_THREADS; i++) {
>  		pctx = &perf->pthr_ctx[i];
>  		if (pctx->thread) {
> @@ -587,65 +612,66 @@ static ssize_t debugfs_run_write(struct file
> *filp, const char __user *ubuf,
>  {
>  	struct perf_ctx *perf = filp->private_data;
>  	int node, i;
> +	DECLARE_WAIT_QUEUE_HEAD(wq);
>  
>  	if (!perf->link_is_up)
> -		return 0;
> +		return -ENOLINK;
>  
>  	if (perf->perf_threads == 0)
> -		return 0;
> +		return -EINVAL;
>  
> -	if (atomic_read(&perf->tsync) == 0)
> -		perf->run = false;
> +	if (!mutex_trylock(&perf->run_mutex))
> +		return -EBUSY;
>  
> -	if (perf->run)
> -		threads_cleanup(perf);
> -	else {
> -		perf->run = true;
> +	if (perf->perf_threads > MAX_THREADS) {
> +		perf->perf_threads = MAX_THREADS;
> +		pr_info("Reset total threads to: %u\n",
> MAX_THREADS);
> +	}
>  
> -		if (perf->perf_threads > MAX_THREADS) {
> -			perf->perf_threads = MAX_THREADS;
> -			pr_info("Reset total threads to: %u\n",
> MAX_THREADS);
> -		}
> +	/* no greater than 1M */
> +	if (seg_order > MAX_SEG_ORDER) {
> +		seg_order = MAX_SEG_ORDER;
> +		pr_info("Fix seg_order to %u\n", seg_order);
> +	}
>  
> -		/* no greater than 1M */
> -		if (seg_order > MAX_SEG_ORDER) {
> -			seg_order = MAX_SEG_ORDER;
> -			pr_info("Fix seg_order to %u\n", seg_order);
> -		}
> +	if (run_order < seg_order) {
> +		run_order = seg_order;
> +		pr_info("Fix run_order to %u\n", run_order);
> +	}
>  
> -		if (run_order < seg_order) {
> -			run_order = seg_order;
> -			pr_info("Fix run_order to %u\n", run_order);
> -		}
> +	node = dev_to_node(&perf->ntb->pdev->dev);
> +	atomic_set(&perf->tdone, 0);
>  
> -		node = dev_to_node(&perf->ntb->pdev->dev);
> -		/* launch kernel thread */
> -		for (i = 0; i < perf->perf_threads; i++) {
> -			struct pthr_ctx *pctx;
> -
> -			pctx = &perf->pthr_ctx[i];
> -			atomic_set(&pctx->dma_sync, 0);
> -			pctx->perf = perf;
> -			pctx->thread =
> -				kthread_create_on_node(ntb_perf_thre
> ad,
> -						       (void *)pctx,
> -						       node,
> "ntb_perf %d", i);
> -			if (IS_ERR(pctx->thread)) {
> -				pctx->thread = NULL;
> -				goto err;
> -			} else
> -				wake_up_process(pctx->thread);
> -
> -			if (perf->run == false)
> -				return -ENXIO;
> -		}
> +	/* launch kernel thread */
> +	for (i = 0; i < perf->perf_threads; i++) {
> +		struct pthr_ctx *pctx;
>  
> +		pctx = &perf->pthr_ctx[i];
> +		atomic_set(&pctx->dma_sync, 0);
> +		pctx->perf = perf;
> +		pctx->wq = &wq;
> +		pctx->thread =
> +			kthread_create_on_node(ntb_perf_thread,
> +					       (void *)pctx,
> +					       node, "ntb_perf %d",
> i);
> +		if (IS_ERR(pctx->thread)) {
> +			pctx->thread = NULL;
> +			goto err;
> +		} else {
> +			wake_up_process(pctx->thread);
> +		}
>  	}
>  
> +	wait_event_interruptible(wq,
> +		atomic_read(&perf->tdone) == perf->perf_threads);
> +
> +	threads_cleanup(perf);
> +	mutex_unlock(&perf->run_mutex);
>  	return count;
>  
>  err:
>  	threads_cleanup(perf);
> +	mutex_unlock(&perf->run_mutex);
>  	return -ENXIO;
>  }
>  
> @@ -713,7 +739,7 @@ static int perf_probe(struct ntb_client *client,
> struct ntb_dev *ntb)
>  	perf->ntb = ntb;
>  	perf->perf_threads = 1;
>  	atomic_set(&perf->tsync, 0);
> -	perf->run = false;
> +	mutex_init(&perf->run_mutex);
>  	spin_lock_init(&perf->db_lock);
>  	perf_setup_mw(ntb, perf);
>  	INIT_DELAYED_WORK(&perf->link_work, perf_link_work);
> @@ -748,6 +774,8 @@ static void perf_remove(struct ntb_client
> *client, struct ntb_dev *ntb)
>  
>  	dev_dbg(&perf->ntb->dev, "%s called\n", __func__);
>  
> +	mutex_lock(&perf->run_mutex);
> +
>  	cancel_delayed_work_sync(&perf->link_work);
>  	cancel_work_sync(&perf->link_cleanup);
>  
> -- 
> 2.1.4
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ