lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <DB39E24B-FB97-4572-BC43-5D0836FAB526@intel.com>
Date:   Fri, 27 Oct 2017 09:19:27 +0000
From:   "Dilger, Andreas" <andreas.dilger@...el.com>
To:     NeilBrown <neilb@...e.com>
CC:     "Drokin, Oleg" <oleg.drokin@...el.com>,
        James Simmons <jsimmons@...radead.org>,
        Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
        "lustre-devel@...ts.lustre.org" <lustre-devel@...ts.lustre.org>,
        "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 4/9] staging: lustre: ldlm: remove 'flags' arg from
 ldlm_process_flock_lock()

On Oct 22, 2017, at 18:53, NeilBrown <neilb@...e.com> wrote:
> 
> This is only ever set to LDLM_FL_WAIT_NOREPROC, so we can remove the arg
> and discard any code that is only run when it doesn't have that value.
> 
> Signed-off-by: NeilBrown <neilb@...e.com>

Reviewed-by: Andreas Dilger <andreas.dilger@...el.com>

> ---
> drivers/staging/lustre/lustre/ldlm/ldlm_flock.c |  110 ++++-------------------
> 1 file changed, 21 insertions(+), 89 deletions(-)
> 
> diff --git a/drivers/staging/lustre/lustre/ldlm/ldlm_flock.c b/drivers/staging/lustre/lustre/ldlm/ldlm_flock.c
> index 05e6b67b0e72..2d1fa2b33129 100644
> --- a/drivers/staging/lustre/lustre/ldlm/ldlm_flock.c
> +++ b/drivers/staging/lustre/lustre/ldlm/ldlm_flock.c
> @@ -122,7 +122,7 @@ ldlm_flock_destroy(struct ldlm_lock *lock, enum ldlm_mode mode, __u64 flags)
>  * is released.
>  *
>  */
> -static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> +static int ldlm_process_flock_lock(struct ldlm_lock *req)
> {
> 	struct ldlm_resource *res = req->l_resource;
> 	struct ldlm_namespace *ns = ldlm_res_to_ns(res);
> @@ -138,8 +138,8 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 	const struct ldlm_callback_suite null_cbs = { };
> 
> 	CDEBUG(D_DLMTRACE,
> -	       "flags %#llx owner %llu pid %u mode %u start %llu end %llu\n",
> -	       *flags, new->l_policy_data.l_flock.owner,
> +	       "owner %llu pid %u mode %u start %llu end %llu\n",
> +	       new->l_policy_data.l_flock.owner,
> 	       new->l_policy_data.l_flock.pid, mode,
> 	       req->l_policy_data.l_flock.start,
> 	       req->l_policy_data.l_flock.end);
> @@ -150,74 +150,16 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 	req->l_blocking_ast = NULL;
> 
> reprocess:
> -	if ((*flags == LDLM_FL_WAIT_NOREPROC) || (mode == LCK_NL)) {
> -		/* This loop determines where this processes locks start
> -		 * in the resource lr_granted list.
> -		 */
> -		list_for_each(tmp, &res->lr_granted) {
> -			lock = list_entry(tmp, struct ldlm_lock,
> -					  l_res_link);
> -			if (ldlm_same_flock_owner(lock, req)) {
> -				ownlocks = tmp;
> -				break;
> -			}
> -		}
> -	} else {
> -		int reprocess_failed = 0;
> -
> -		lockmode_verify(mode);
> -
> -		/* This loop determines if there are existing locks
> -		 * that conflict with the new lock request.
> -		 */
> -		list_for_each(tmp, &res->lr_granted) {
> -			lock = list_entry(tmp, struct ldlm_lock,
> -					  l_res_link);
> -
> -			if (ldlm_same_flock_owner(lock, req)) {
> -				if (!ownlocks)
> -					ownlocks = tmp;
> -				continue;
> -			}
> -
> -			/* locks are compatible, overlap doesn't matter */
> -			if (lockmode_compat(lock->l_granted_mode, mode))
> -				continue;
> -
> -			if (!ldlm_flocks_overlap(lock, req))
> -				continue;
> -
> -			if (*flags & LDLM_FL_BLOCK_NOWAIT) {
> -				ldlm_flock_destroy(req, mode, *flags);
> -				return LDLM_ITER_STOP;
> -			}
> -
> -			if (*flags & LDLM_FL_TEST_LOCK) {
> -				ldlm_flock_destroy(req, mode, *flags);
> -				req->l_req_mode = lock->l_granted_mode;
> -				req->l_policy_data.l_flock.pid =
> -					lock->l_policy_data.l_flock.pid;
> -				req->l_policy_data.l_flock.start =
> -					lock->l_policy_data.l_flock.start;
> -				req->l_policy_data.l_flock.end =
> -					lock->l_policy_data.l_flock.end;
> -				*flags |= LDLM_FL_LOCK_CHANGED;
> -				return LDLM_ITER_STOP;
> -			}
> -
> -			ldlm_resource_add_lock(res, &res->lr_waiting, req);
> -			*flags |= LDLM_FL_BLOCK_GRANTED;
> -			return LDLM_ITER_STOP;
> +	/* This loop determines where this processes locks start
> +	 * in the resource lr_granted list.
> +	 */
> +	list_for_each(tmp, &res->lr_granted) {
> +		lock = list_entry(tmp, struct ldlm_lock,
> +				  l_res_link);
> +		if (ldlm_same_flock_owner(lock, req)) {
> +			ownlocks = tmp;
> +			break;
> 		}
> -		if (reprocess_failed)
> -			return LDLM_ITER_CONTINUE;
> -	}
> -
> -	if (*flags & LDLM_FL_TEST_LOCK) {
> -		ldlm_flock_destroy(req, mode, *flags);
> -		req->l_req_mode = LCK_NL;
> -		*flags |= LDLM_FL_LOCK_CHANGED;
> -		return LDLM_ITER_STOP;
> 	}
> 
> 	/* Scan the locks owned by this process that overlap this request.
> @@ -267,7 +209,8 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 			}
> 
> 			if (added) {
> -				ldlm_flock_destroy(lock, mode, *flags);
> +				ldlm_flock_destroy(lock, mode,
> +						   LDLM_FL_WAIT_NOREPROC);
> 			} else {
> 				new = lock;
> 				added = 1;
> @@ -293,7 +236,8 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 					new->l_policy_data.l_flock.end + 1;
> 				break;
> 			}
> -			ldlm_flock_destroy(lock, lock->l_req_mode, *flags);
> +			ldlm_flock_destroy(lock, lock->l_req_mode,
> +					   LDLM_FL_WAIT_NOREPROC);
> 			continue;
> 		}
> 		if (new->l_policy_data.l_flock.end >=
> @@ -325,7 +269,7 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 			lock_res_and_lock(req);
> 			if (IS_ERR(new2)) {
> 				ldlm_flock_destroy(req, lock->l_granted_mode,
> -						   *flags);
> +						   LDLM_FL_WAIT_NOREPROC);
> 				return LDLM_ITER_STOP;
> 			}
> 			goto reprocess;
> @@ -354,9 +298,8 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 					     &new2->l_remote_handle,
> 					     &new2->l_exp_hash);
> 		}
> -		if (*flags == LDLM_FL_WAIT_NOREPROC)
> -			ldlm_lock_addref_internal_nolock(new2,
> -							 lock->l_granted_mode);
> +		ldlm_lock_addref_internal_nolock(new2,
> +						 lock->l_granted_mode);
> 
> 		/* insert new2 at lock */
> 		ldlm_resource_add_lock(res, ownlocks, new2);
> @@ -377,22 +320,13 @@ static int ldlm_process_flock_lock(struct ldlm_lock *req, __u64 *flags)
> 		ldlm_resource_add_lock(res, ownlocks, req);
> 	}
> 
> -	if (*flags != LDLM_FL_WAIT_NOREPROC) {
> -		/* The only one possible case for client-side calls flock
> -		 * policy function is ldlm_flock_completion_ast inside which
> -		 * carries LDLM_FL_WAIT_NOREPROC flag.
> -		 */
> -		CERROR("Illegal parameter for client-side-only module.\n");
> -		LBUG();
> -	}
> -
> 	/* In case we're reprocessing the requested lock we can't destroy
> 	 * it until after calling ldlm_add_ast_work_item() above so that laawi()
> 	 * can bump the reference count on \a req. Otherwise \a req
> 	 * could be freed before the completion AST can be sent.
> 	 */
> 	if (added)
> -		ldlm_flock_destroy(req, mode, *flags);
> +		ldlm_flock_destroy(req, mode, LDLM_FL_WAIT_NOREPROC);
> 
> 	ldlm_resource_dump(D_INFO, res);
> 	return LDLM_ITER_CONTINUE;
> @@ -582,12 +516,10 @@ ldlm_flock_completion_ast(struct ldlm_lock *lock, __u64 flags, void *data)
> 		getlk->fl_start = (loff_t)lock->l_policy_data.l_flock.start;
> 		getlk->fl_end = (loff_t)lock->l_policy_data.l_flock.end;
> 	} else {
> -		__u64 noreproc = LDLM_FL_WAIT_NOREPROC;
> -
> 		/* We need to reprocess the lock to do merges or splits
> 		 * with existing locks owned by this process.
> 		 */
> -		ldlm_process_flock_lock(lock, &noreproc);
> +		ldlm_process_flock_lock(lock);
> 	}
> 	unlock_res_and_lock(lock);
> 	return rc;
> 
> 

Cheers, Andreas
--
Andreas Dilger
Lustre Principal Architect
Intel Corporation







Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ