lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <574C09E5.7050902@redhat.com>
Date:	Mon, 30 May 2016 17:37:41 +0800
From:	Xunlei Pang <xpang@...hat.com>
To:	Peter Zijlstra <peterz@...radead.org>, bsegall@...gle.com
Cc:	Xunlei Pang <xlpang@...hat.com>, linux-kernel@...r.kernel.org,
	Juri Lelli <juri.lelli@....com>,
	Ingo Molnar <mingo@...hat.com>,
	Steven Rostedt <rostedt@...dmis.org>, pjt@...gle.com
Subject: Re: [PATCH] sched/fair: Fix the wrong throttled clock time for
 cfs_rq_clock_task()

Ping

On 2016/05/12 at 11:36, Xunlei Pang wrote:
> On 2016/05/11 at 14:49, Peter Zijlstra wrote:
>> On Tue, May 10, 2016 at 11:19:44AM -0700, bsegall@...gle.com wrote:
>>> Xunlei Pang <xlpang@...hat.com> writes:
>>>
>>>> Two minor fixes for cfs_rq_clock_task().
>>>> 1) If cfs_rq is currently being throttled, we need to subtract the cfs
>>>>    throttled clock time.
>>>>
>>>> 2) Make "throttled_clock_task_time" update SMP unrelated. Now UP cases
>>>>    need it as well.
>>>>
>>>> Signed-off-by: Xunlei Pang <xlpang@...hat.com>
>>>> ---
>>>>  kernel/sched/fair.c | 4 +---
>>>>  1 file changed, 1 insertion(+), 3 deletions(-)
>>>>
>>>> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
>>>> index 1708729e..fb80a12 100644
>>>> --- a/kernel/sched/fair.c
>>>> +++ b/kernel/sched/fair.c
>>>> @@ -3655,7 +3655,7 @@ static inline struct cfs_bandwidth *tg_cfs_bandwidth(struct task_group *tg)
>>>>  static inline u64 cfs_rq_clock_task(struct cfs_rq *cfs_rq)
>>>>  {
>>>>  	if (unlikely(cfs_rq->throttle_count))
>>>> -		return cfs_rq->throttled_clock_task;
>>>> +		return cfs_rq->throttled_clock_task - cfs_rq->throttled_clock_task_time;
>>>>  
>>>>  	return rq_clock_task(rq_of(cfs_rq)) - cfs_rq->throttled_clock_task_time;
>>>>  }
>> The alternative is obviously to do the subtraction in
>> tg_throttle_down(), were we set ->throttled_clock_task.
> It is possible, but throttled_clock_task is a timestamp, I think doing it here is semantically better.
>
>>>> @@ -3793,13 +3793,11 @@ static int tg_unthrottle_up(struct task_group *tg, void *data)
>>>>  	struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
>>>>  
>>>>  	cfs_rq->throttle_count--;
>>>> -#ifdef CONFIG_SMP
>>>>  	if (!cfs_rq->throttle_count) {
>>>>  		/* adjust cfs_rq_clock_task() */
>>>>  		cfs_rq->throttled_clock_task_time += rq_clock_task(rq) -
>>>>  					     cfs_rq->throttled_clock_task;
>>>>  	}
>>>> -#endif
>>>>  
>>>>  	return 0;
>>>>  }
>>> [Cc: pjt@...gle.com]
>>>
>>> This looks reasonable to me (at least the first part; I'm not
>>> certain why the CONFIG_SMP ifdef was put in place).
>> 64660c864f46 ("sched: Prevent interactions with throttled entities")
>>
>> Introduced it, because at that time it was about updating shares, which
>> is only present on SMP. Then:
>>
>> f1b17280efbd ("sched: Maintain runnable averages across throttled periods")
>>
>> Added the clock thing inside it, and:
>>
>> 82958366cfea ("sched: Replace update_shares weight distribution with per-entity computation")
>>
>> took out the shares update and left the clock update, resulting in the
>> current code.
>>
>>
> Thanks,
> Xunlei
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ