[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <AE90C24D6B3A694183C094C60CF0A2F6026B6E7E@saturn3.aculab.com>
Date: Mon, 27 Feb 2012 15:59:23 -0000
From: "David Laight" <David.Laight@...LAB.COM>
To: "Stanislav Kinsbursky" <skinsbursky@...allels.com>,
<Trond.Myklebust@...app.com>
Cc: <linux-nfs@...r.kernel.org>, <xemul@...allels.com>,
<neilb@...e.de>, <netdev@...r.kernel.org>,
<linux-kernel@...r.kernel.org>, <jbottomley@...allels.com>,
<bfields@...ldses.org>, <davem@...emloft.net>, <devel@...nvz.org>
Subject: RE: [PATCH v2 2/4] NFS: release per-net clients lock before calling PipeFS dentries creation
> spin_lock(&nn->nfs_client_lock);
> - list_for_each_entry(clp, &nn->nfs_client_list, cl_share_link) {
> + list_for_each_entry_safe(clp, tmp, &nn->nfs_client_list,
cl_share_link) {
> if (clp->rpc_ops != &nfs_v4_clientops)
> continue;
> + atomic_inc(&clp->cl_count);
> + spin_unlock(&nn->nfs_client_lock);
> error = __rpc_pipefs_event(clp, event, sb);
> + nfs_put_client(clp);
> if (error)
> break;
> + spin_lock(&nn->nfs_client_lock);
> }
> spin_unlock(&nn->nfs_client_lock);
> return error;
The locking doesn't look right if the loop breaks on error.
(Same applied to patch v2 1/4)
Although list_fo_each_entry_safe() allows the current entry
to be freed, I don't believe it allows the 'next' to be freed.
I doubt there is protection against that happening.
Do you need to use an atomic_inc() for cl_count.
I'd guess the nfs_client_lock is usually held?
David
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists