[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <20110901154810.cdda1d94.kamezawa.hiroyu@jp.fujitsu.com>
Date: Thu, 1 Sep 2011 15:48:10 +0900
From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
To: Glauber Costa <glommer@...allels.com>
Cc: <netdev@...r.kernel.org>,
Linux Containers <containers@...ts.osdl.org>,
<linux-mm@...ck.org>, Pavel Emelyanov <xemul@...allels.com>,
"Eric W. Biederman" <ebiederm@...ssion.com>,
David Miller <davem@...emloft.net>,
Stephen Hemminger <shemminger@...tta.com>, <penberg@...nel.org>
Subject: Re: [RFMC] per-container tcp buffer limitation
On Wed, 31 Aug 2011 23:43:27 -0300
Glauber Costa <glommer@...allels.com> wrote:
> Hello People,
>
> [ For the ones in linux-mm that are receiving this for the first time,
> this is a follow up of
> http://thread.gmane.org/gmane.linux.kernel.containers/21295 ]
>
> Here is a new, a bit more mature version of my previous RFC. Now I
> Request For More Comments from you guys in this new version of the patch.
>
> Highlights:
>
> * Although I do intend to experiment with more scenarios (suggestions
> welcome), there does not seem to be a (huge) performance hit with this
> patch applied, at least in a basic latency benchmark. That indicates
> that even if we can demonstrate a performance hit, it won't be too hard
> to optimize it away (famous last words?)
>
> Since the patch touches both rcv and snd sides, I benchmarked it with
> netperf against localhost. Command line: netperf -t TCP_RR -H localhost.
>
> Without the patch
> =================
>
> Socket Size Request Resp. Elapsed Trans.
> Send Recv Size Size Time Rate
> bytes Bytes bytes bytes secs. per sec
>
> 16384 87380 1 1 10.00 26996.35
> 16384 87380
>
> With the patch
> ===============
>
> Local /Remote
> Socket Size Request Resp. Elapsed Trans.
> Send Recv Size Size Time Rate
> bytes Bytes bytes bytes secs. per sec
>
> 16384 87380 1 1 10.00 27291.86
> 16384 87380
>
>
> As you can see, rate is a bit higher, but still under an one percent
> range, meaning it is basically unchanged. I will benchmark it with
> various levels of cgroup nesting on my next submission so we can have a
> better idea of the impact of it when enabled.
>
seems nice.
> * As nicely pointed out by Kamezawa, I dropped the sockets cgroup, and
> introduced a kmem cgroup. After careful consideration, I decided not to
> reuse the memcg. Basically, my impression is that memcg is concerned
> with user objects, with page granularity and its swap attributes.
> Because kernel objects are entirely different, I prefer to group them here.
>
I myself has no objection to this direction. Other guys ?
Thanks,
-Kame
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists