[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <2004310918.15554.1399982845138.JavaMail.zimbra@efficios.com>
Date: Tue, 13 May 2014 12:07:25 +0000 (UTC)
From: Mathieu Desnoyers <mathieu.desnoyers@...icios.com>
To: George Spelvin <linux@...izon.com>
Cc: john stultz <john.stultz@...aro.org>, linux-kernel@...r.kernel.org,
Thomas Gleixner <tglx@...utronix.de>,
Peter Zijlstra <peterz@...radead.org>
Subject: Re: [PATCH 4/4] timekeeping: Use printk_deferred when holding
timekeeping seqlock
----- Original Message -----
> From: "George Spelvin" <linux@...izon.com>
> To: "john stultz" <john.stultz@...aro.org>, linux@...izon.com
> Cc: linux-kernel@...r.kernel.org, "mathieu desnoyers" <mathieu.desnoyers@...icios.com>
> Sent: Tuesday, May 13, 2014 1:13:24 AM
> Subject: Re: [PATCH 4/4] timekeeping: Use printk_deferred when holding timekeeping seqlock
>
> >> 2) Using wait-free coding techniques where readers help the writer if
> >> they notice a stall. This is much hairier internal code, but makes
> >> life easier on the callers. The basic procedure is:
> >>
> >> - A conventionally locked writer decides that the frequency is to be
> >> adjusted, effective at the current time (or perhaps slightly in the
> >> future).
> >> - It publishes its idea of the effective time, and that an update is
> >> in progress.
> >> - A reader comes in, reads the hardware clock, and checks:
> >> - Is a rate update in progress, and is the effective time *before*
> >> the time I just read?
> >> - If so, I'm going to force a postponement to the time I just read,
> >> using compare-and-swap trickery.
> >> - Then it proceeds to use the old timekeeper rate information.
> >> - When the writer finishes figuring out the new timekeeper state,
> >> as part of the atomic installation of a new state, it checks to
> >> see if a reader has forced a postponement. If so, it goes back and
> >> recomputes the state with the later effective time and tries again.
> >
> > Hrm.. So basically during the update you lock readers to the counter
> > value read at the beginning of the update. So readers don't block, but
> > time doesn't progress while the update is being made. Sounds promising!
>
> Er... no. I guess my explanation didn't work. I thought of that (it's
> quite simple, after all), but I was worried that the stuttering would
> mess up fine accounting for e.g. interrupt handlers.
>
> Also, it's not implementable. The writer must pick a rate-change moment,
> then announce it. If it stalls between those two operations (and there's
> no way to make them atomic), then it might announce a moment already in
> the past, causing a later reader to return the stalled time, while an
> earlier reader that didn't see the announcement has already returned
> a later time.
Your description above summarizes well the conclusions I reached when doing
my nonblocking-read clock source implementation prototype a while back. If
we can add a new clock semantic, there is a solution I proposed last year
that might just achieve what you are looking for:
We could expose a new clock type (besides monotonic and realtime) that is
documented as non-strictly monotonic. It may return a time very slightly in
the past if readers race with clock source frequency change. The caller could
handle this situation (e.g. in user-space) by keeping its own per-cpu or
per-thread "last clock value" data structure (something we cannot do in a
vDSO) if it really cares about per-cpu/thread clock monotonicity.
This could be implemented with the scheme I proposed as a prototype here:
https://lkml.org/lkml/2013/9/14/136
The idea here would be to keep both a read seqlock (for realtime and
monotonic), as well as an API/ABI that allows reading this "latched
clock" value (documented as non-strictly-monotonic).
Thoughts ?
Thanks,
Mathieu
>
> Inseate, the writer announces a *proposed* rate-change time, but a number
> of conditions can cause that time to be postponed. (Basically, the writer
> loops back to the beginning and applies the rate change later.)
>
> One condition is enforced by the writer itself: it reads the time again
> after making the announcement, and if the announced time has already
> passed, loops.
>
> But this is also checked by all readers. If an update is in progress,
> with a proposed time earlier than the reader is in the process of reading,
> the writer is told to try again later.
>
>
> One trick that can minimize the number of retries is to add a very small
> time delta (equal to the typical write update cycle) to the writer's
> effective time. If the writer completes before that time happens (the
> common case), readers happily use the old time values during the update.
> Only if the writer is delayed more than expected will a reader notice
> a problem. "Hey, I read the hardware at tick x, but the writer is trying
> to update tick-to-seconds translations for times later than tick y < x.
> Hey, writer, recompute with a higher y!"
>
>
> There is also the issue of a reader coming along after the change-over
> and wanting to translate a time x < y. There are several ways to
> handle this. Either the old parameters have to be kept around until
> time y has passed, or the readers must wait until time y.
>
> (They may not return early reporting time y or there may be race conditions.)
>
> > Oh.. so its actually more like the update is canceled if a reader can
> > complete a read and set a flag while the update is in flight? Hrm.. I
> > worry with the number and frequency of time reads on many systems, you'd
> > end up with update starvation (something which use to be a problem back
> > when timekeeping was managed with just spinlocks, and SMP wasn't that
> > common - so I suspect it can only be worse now).
>
> The solution, mentioned above, is to have the effective time of the
> update set to the predicted write completion time. Then the readers will
> never see an effective time less than their time and need to postpone
> the writer.
>
> Only if the writer is delayed unexpectedly does that happen.
>
> > Further, the write in the read path would cause problems for VDSO time,
> > where everything is computed on read-only data in userspace.
>
> Ah! Now *that* poses a potential problem. I didn't think of that.
> Arrgh, that's going to take some work.
>
> Two possibilities:
> 1) The VDSO code could just spin. It's not holding any kernel locks,
> so there's not much problem.
> 2) Possibly after some optimistic spinning, the VDSO code could
> trap to the kernel as a fallback.
>
> > Yea. It definitely gets complex. But timekeeping is always a very
> > hotpath, so complexity for performance is a normal trade. So there may
> > be a way to make it work. But finding a way to make it easier to
> > comprehend (even just having a clear metaphor for what's being done)
> > would be excellent.
>
> I'll try.
>
--
Mathieu Desnoyers
EfficiOS Inc.
http://www.efficios.com
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists