[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20121024182005.GF2465@linux.vnet.ibm.com>
Date: Wed, 24 Oct 2012 11:20:05 -0700
From: "Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>
To: Oleg Nesterov <oleg@...hat.com>
Cc: Mikulas Patocka <mpatocka@...hat.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Ingo Molnar <mingo@...e.hu>,
Peter Zijlstra <peterz@...radead.org>,
Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
Ananth N Mavinakayanahalli <ananth@...ibm.com>,
Anton Arapov <anton@...hat.com>, linux-kernel@...r.kernel.org
Subject: Re: [PATCH 2/2] percpu-rw-semaphores: use rcu_read_lock_sched
On Wed, Oct 24, 2012 at 07:18:55PM +0200, Oleg Nesterov wrote:
> On 10/24, Paul E. McKenney wrote:
> >
> > static inline void percpu_up_read(struct percpu_rw_semaphore *p)
> > {
> > /*
> > * Decrement our count, but protected by RCU-sched so that
> > * the writer can force proper serialization.
> > */
> > rcu_read_lock_sched();
> > this_cpu_dec(*p->counters);
> > rcu_read_unlock_sched();
> > }
>
> Yes, the explicit lock/unlock makes the new assumptions about
> synchronize_sched && barriers unnecessary. And iiuc this could
> even written as
>
> rcu_read_lock_sched();
> rcu_read_unlock_sched();
>
> this_cpu_dec(*p->counters);
But this would lose the memory barrier that is inserted by
synchronize_sched() after the CPU's last RCU-sched read-side critical
section.
> > Of course, it would be nice to get rid of the extra synchronize_sched().
> > One way to do this is to use SRCU, which allows blocking operations in
> > its read-side critical sections (though also increasing read-side overhead
> > a bit, and also untested):
> >
> > ------------------------------------------------------------------------
> >
> > struct percpu_rw_semaphore {
> > bool locked;
> > struct mutex mtx; /* Could also be rw_semaphore. */
> > struct srcu_struct s;
> > wait_queue_head_t wq;
> > };
>
> but in this case I don't understand
>
> > static inline void percpu_up_write(struct percpu_rw_semaphore *p)
> > {
> > /* Allow others to proceed, but not yet locklessly. */
> > mutex_unlock(&p->mtx);
> >
> > /*
> > * Ensure that all calls to percpu_down_read() that did not
> > * start unambiguously after the above mutex_unlock() still
> > * acquire the lock, forcing their critical sections to be
> > * serialized with the one terminated by this call to
> > * percpu_up_write().
> > */
> > synchronize_sched();
>
> how this synchronize_sched() can help...
Indeed it cannot! It should instead be synchronize_srcu(&p->s). I guess that
I really meant it when I said it was untested. ;-)
Thanx, Paul
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists