[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <m1my38lb0f.fsf@fess.ebiederm.org>
Date: Fri, 30 Oct 2009 16:25:52 -0700
From: ebiederm@...ssion.com (Eric W. Biederman)
To: Benjamin LaHaise <bcrl@...et.ca>
Cc: Eric Dumazet <eric.dumazet@...il.com>,
Octavian Purdila <opurdila@...acom.com>,
netdev@...r.kernel.org, Cosmin Ratiu <cratiu@...acom.com>
Subject: Re: [PATCH] net: allow netdev_wait_allrefs() to run faster
Benjamin LaHaise <bcrl@...et.ca> writes:
> On Thu, Oct 29, 2009 at 06:45:32PM -0700, Eric W. Biederman wrote:
>> The reason for the existence of sysfs_dirent is as things grow larger
>> we want to keep the amount of RAM consumed down. So we don't pin
>> everything in the dcache. So we try and keep the amount of memory
>> consumed down.
>
> I'm aware of that, but for users running into this sort of scaling issue,
> the amount of RAM required is a non-issue (30,000 interfaces require about
> 1GB of RAM at present), making the question more one of how to avoid the
> overhead for users who don't require it. I'd prefer a config option. The
> only way I can really see saving memory usage is to somehow tie sysfs dirent
> lookups into the network stack's own tables for looking up device entries.
> The network stack already has to cope with this kind of scaling, and that
> would save the RAM.
There is that. I'm trying to figure out how to add the improvements
without making sysfs_dirent larger. Which I think that is doable.
>> So I would like to see how much we can par down.
>
>> For dealing with seeks in the middle of readdir I expect the best way
>> to do that is to be inspired by htrees in extNfs and return a hash of
>> the filename as our position, and keep the filename list sorted by
>> that hash. Since we are optimizing for size we don't need to store
>> that hash. Then we can turn that list into a some flavor of sorted
>> binary tree.
>
> readdir() generally isn't an issue at present.
Supporting seekdir into the middle of a directory is the entire reason
I keep the entries sorted by inode. If we sort by a hash of the name.
We can use the hash to support directory position in readdir and seekdir.
And we can completely remove the linear list when the rb_tree is introduced.
>> I'm surprised sysfs_count_nlink shows up, as it is not directly on the
>> add or remove path. I think the answer there is to change s_flags
>> into a set of bitfields and make link_count one of them, perhaps
>> 16bits long. If we ever overflow our bitfield we can just set link
>> count to 0, and userspace (aka find) will know it can't optimized
>> based on link count.
>
> It shows up because of the bits of userspace (udev) touching the directory
> from things like the hotplug code path.
I realized after sending the message that s_mode in sysfs_dirent is a
real size offense. It is a 16bit field packed in between two longs.
So in practice it is possible to move the s_mode up next to s_flags
and add a s_nlink after it both unsigned short and get a cheap sysfs_nlink.
>> I was expecting someone to run into problems with the linear directory
>> of sysfs someday.
>
> Alas, sysfs isn't the only offender.
Agreed. Sysfs is probably the easiest to untangle.
Since I'm not quite ready to post my patches. I will briefly
mention what I have in my queue and hopefully get things posted.
I have changes to make it so that sysfs never has to go from
the sysfs_dirent to the sysfs inode.
I have changes to sys_sysctl() so that it becomes a filesystem lookup
under /proc/sys. Which ultimately makes the code easier to maintain
and debug.
Now back to getting things forward ported and ready to post.
Eric
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists