[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180126213943.7zwijanlbrq3y666@rob-hp-laptop>
Date: Fri, 26 Jan 2018 15:39:43 -0600
From: Rob Herring <robh@...nel.org>
To: Chintan Pandya <cpandya@...eaurora.org>
Cc: Rasmus Villemoes <rasmus.villemoes@...vas.dk>,
Frank Rowand <frowand.list@...il.com>,
"open list:OPEN FIRMWARE AND FLATTENED DEVICE TREE BINDINGS"
<devicetree@...r.kernel.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
linux-arm-msm <linux-arm-msm@...r.kernel.org>
Subject: Re: [PATCH v2] of: use hash based search in of_find_node_by_phandle
On Fri, Jan 26, 2018 at 09:34:59AM -0600, Rob Herring wrote:
> On Fri, Jan 26, 2018 at 9:14 AM, Chintan Pandya <cpandya@...eaurora.org> wrote:
> >
> >> I'm probably missing something obvious, but: Aren't phandles in practice
> >> small consecutive integers assigned by dtc? If so, why not just have a
> >> smallish static array mapping the small phandle values directly to
> >> device node, instead of adding a pointer to every struct device_node? Or
> >> one could determine the size of the array dynamically (largest seen
> >> phandle value, capping at something sensible, e.g. 1024).
>
> I do have some concerns that is a bit fragile and dependent on dtc's
> implementations. However, I guess we already kind of are with overlay
> phandles.
>
> > Haven't noticed this earlier !! If following is known or true, we can avoid
> > using hash-table and save per device_node hlish_node.
> >
> > 1. How to know max phandle value without traversing tree once? In my
> > case,
> > max is 1263.
>
> We already have to know it for overlays. Plus unflattening has to
> handle phandles specially already, so it would be easy to do there if
> we aren't already.
>
> Then the question what to do with overlays. For now, we can probably
> assume too few phandles to matter.
>
> > 2. Although, I haven't observed collision but is it like every
> > device_node
> > is associated with unique phandle value ?
>
> Yes, phandles must be unique.
>
> >> In either case, one would still need to keep the code doing the
> >> whole-tree traversal for handling large phandle values, but I think the
> >> above should make lookup O(1) in most cases.
> >
> > I would refrain doing this because that will make this API inconsistent in
> > terms
> > of time taken by different nodes. I see that people do change their device
> > placing in DT and that changes time taken in of_* APIs for them but
> > affecting
> > others.
>
> Who cares. It's the total time that matters. It's obviously not a
> problem until you have 1000s of lookups as no one cared until recently
> (though you aren't the first with a hash table lookup).
>
> >> Alternatively, one could just count the number of nodes with a phandle,
> >> allocate an array of that many pointers (so the memory use is certainly
> >> no more than if adding a pointer to each device_node), and sort it by
> >> phandle, so one can do lookup using a binary search.
> >>
> >> Rasmus
> >
> > This is certainly doable if current approach is not welcomed due to
> > addition on hlish_node in device_node.
>
> I certainly prefer an out of band approach as that's easier to turn of
> if we want to save memory.
>
> Still, I'd like to see some data on a cache based approach and reasons
> why that won't work.
I was curious, so I implemented it. It ends up being similar to Rasmus's
1st suggestion. The difference is we don't try to store all entries, but
rather implement a hash table that doesn't handle collisions. Relying on
the fact that phandles are just linearly allocated from 0, we just mask
the high bits of the phandle to get the index.
Using the unittest (modified to run twice), the time for 192 calls goes
from ~50us to ~30us. Can you try out on your setup and try different
array sizes.
8<----------------------------------------------------------------------
diff --git a/drivers/of/base.c b/drivers/of/base.c
index dd0b4201f1cc..cb88adf4f3db 100644
--- a/drivers/of/base.c
+++ b/drivers/of/base.c
@@ -986,6 +986,13 @@ int of_modalias_node(struct device_node *node, char *modalias, int len)
}
EXPORT_SYMBOL_GPL(of_modalias_node);
+static ktime_t elapsed;
+static int call_count;
+
+#define PHANDLE_CACHE_SZ 64
+#define PHANDLE_CACHE_HASH(x) ((x) & (PHANDLE_CACHE_SZ - 1))
+static struct device_node *phandle_cache[PHANDLE_CACHE_SZ];
+
/**
* of_find_node_by_phandle - Find a node given a phandle
* @handle: phandle of the node to find
@@ -997,16 +1004,30 @@ struct device_node *of_find_node_by_phandle(phandle handle)
{
struct device_node *np;
unsigned long flags;
+ ktime_t t1, t2;
if (!handle)
return NULL;
raw_spin_lock_irqsave(&devtree_lock, flags);
- for_each_of_allnodes(np)
- if (np->phandle == handle)
- break;
+ t1 = ktime_get();
+ np = phandle_cache[PHANDLE_CACHE_HASH(handle)];
+ if (!np || np->phandle != handle) {
+ for_each_of_allnodes(np)
+ if (np->phandle == handle) {
+ phandle_cache[PHANDLE_CACHE_HASH(handle)] = np;
+ break;
+ }
+
+ }
+
of_node_get(np);
+ t2 = ktime_get();
+ elapsed = ktime_add(elapsed, ktime_sub(t2, t1));
+ call_count++;
raw_spin_unlock_irqrestore(&devtree_lock, flags);
+ if (!(call_count % 32))
+ printk("of_find_node_by_phandle called %d times totalling %lld ns", call_count, elapsed);
return np;
}
EXPORT_SYMBOL(of_find_node_by_phandle);
Powered by blists - more mailing lists