[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20101017052454.GA21273@balbir.in.ibm.com>
Date: Sun, 17 Oct 2010 10:54:54 +0530
From: Balbir Singh <balbir@...ux.vnet.ibm.com>
To: pjt@...gle.com
Cc: linux-kernel@...r.kernel.org,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Ingo Molnar <mingo@...e.hu>,
Srivatsa Vaddagiri <vatsa@...ibm.com>,
Chris Friesen <cfriesen@...tel.com>,
Vaidyanathan Srinivasan <svaidy@...ux.vnet.ibm.com>,
Pierre Bourdon <pbourdon@...ellency.fr>,
Bharata B Rao <bharata@...ux.vnet.ibm.com>
Subject: Re: [RFC tg_shares_up improvements - v1 00/12] [RFC tg_shares_up -
v1 00/12] Reducing cost of tg->shares distribution
* pjt@...gle.com <pjt@...gle.com> [2010-10-15 21:43:49]:
> Hi all,
>
> Peter previously posted a patchset that attempted to improve the problem of
> task_group share distribution. This is something that has been a long-time
> pain point for group scheduling. The existing algorithm considers
> distributions on a per-cpu-per-domain basis and carries a fairly high update
> overhead, especially on larger machines.
>
> I was previously looking at improving this using Fenwick trees to allow a
> single sum without the exorbitant cost but then Peter's idea above was better :).
>
> The kernel is that by monitoring the average contribution to load on a
> per-cpu-per-taskgroup basis we can distribute the weight for which we are
> expected to consume.
>
> This set extends the original posting with a focus on increased fairness and
> reduced convergence (to true average) time. In particular the case of large
> over-commit in the case of a distributed wake-up is a concern which is now
> fairly well addressed.
>
> Obviously everything's experimental but it should be stable/fair.
>
> Some motivation:
>
> 24 thread intel box, 150 active cgroups, multiple threads/group, load at ~90% (10 second sample):
> tip:
> 2.64% [k] tg_shares_up <!>
> 0.15% [k] __set_se_shares
>
> patched:
> 0.02% [k] update_cfs_load
> 0.01% [k] update_cpu_load
> 0.00% [k] update_cfs_shares
>
> Some fairness coverage for the above at: http://rs5.risingnet.net/~pjt/patches/shares_data_v1.txt
>
The results do sound interesting, could I recommend that you try to
get 1000 cgroups or so, with half of them as active and half inactive.
I'd also suggest tracking system and user time at the cgroup level.
--
Three Cheers,
Balbir
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists