[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20070817112253.e6a7cb33.akpm@linux-foundation.org>
Date: Fri, 17 Aug 2007 11:22:53 -0700
From: Andrew Morton <akpm@...ux-foundation.org>
To: Satyam Sharma <satyam@...radead.org>
Cc: Arjan van de Ven <arjan@...radead.org>,
Tim Bird <tim.bird@...sony.com>,
linux kernel <linux-kernel@...r.kernel.org>,
Christoph Lameter <clameter@....com>
Subject: Re: kfree(0) - ok?
On Wed, 15 Aug 2007 05:12:41 +0530 (IST)
Satyam Sharma <satyam@...radead.org> wrote:
> [PATCH] {slub, slob}: use unlikely() for kfree(ZERO_OR_NULL_PTR) check
>
> Considering kfree(NULL) would normally occur only in error paths and
> kfree(ZERO_SIZE_PTR) is uncommon as well, so let's use unlikely() for
> the condition check in SLUB's and SLOB's kfree() to optimize for the
> common case. SLAB has this already.
I went through my current versions of slab/slub/slub and came up with this:
diff -puN mm/slob.c~slub-slob-use-unlikely-for-kfreezero_or_null_ptr-check mm/slob.c
--- a/mm/slob.c~slub-slob-use-unlikely-for-kfreezero_or_null_ptr-check
+++ a/mm/slob.c
@@ -360,7 +360,7 @@ static void slob_free(void *block, int s
slobidx_t units;
unsigned long flags;
- if (ZERO_OR_NULL_PTR(block))
+ if (unlikely(ZERO_OR_NULL_PTR(block)))
return;
BUG_ON(!size);
@@ -466,7 +466,7 @@ void kfree(const void *block)
{
struct slob_page *sp;
- if (ZERO_OR_NULL_PTR(block))
+ if (unlikely(ZERO_OR_NULL_PTR(block)))
return;
sp = (struct slob_page *)virt_to_page(block);
@@ -484,7 +484,7 @@ size_t ksize(const void *block)
{
struct slob_page *sp;
- if (ZERO_OR_NULL_PTR(block))
+ if (unlikely(ZERO_OR_NULL_PTR(block)))
return 0;
sp = (struct slob_page *)virt_to_page(block);
diff -puN mm/slub.c~slub-slob-use-unlikely-for-kfreezero_or_null_ptr-check mm/slub.c
--- a/mm/slub.c~slub-slob-use-unlikely-for-kfreezero_or_null_ptr-check
+++ a/mm/slub.c
@@ -2434,7 +2434,7 @@ size_t ksize(const void *object)
struct page *page;
struct kmem_cache *s;
- if (ZERO_OR_NULL_PTR(object))
+ if (unlikely(ZERO_OR_NULL_PTR(object)))
return 0;
page = get_object_page(object);
@@ -2468,7 +2468,7 @@ void kfree(const void *x)
{
struct page *page;
- if (ZERO_OR_NULL_PTR(x))
+ if (unlikely(ZERO_OR_NULL_PTR(x)))
return;
page = virt_to_head_page(x);
@@ -2785,7 +2785,7 @@ void *__kmalloc_track_caller(size_t size
get_order(size));
s = get_slab(size, gfpflags);
- if (ZERO_OR_NULL_PTR(s))
+ if (unlikely(ZERO_OR_NULL_PTR(s)))
return s;
return slab_alloc(s, gfpflags, -1, caller);
@@ -2801,7 +2801,7 @@ void *__kmalloc_node_track_caller(size_t
get_order(size));
s = get_slab(size, gfpflags);
- if (ZERO_OR_NULL_PTR(s))
+ if (unlikely(ZERO_OR_NULL_PTR(s)))
return s;
return slab_alloc(s, gfpflags, node, caller);
_
Which is getting pretty idiotic:
akpm:/usr/src/25> grep ZERO_OR_NULL_PTR */*.c
mm/slab.c: BUG_ON(ZERO_OR_NULL_PTR(cachep->slabp_cache));
mm/slab.c: if (unlikely(ZERO_OR_NULL_PTR(cachep)))
mm/slab.c: if (unlikely(ZERO_OR_NULL_PTR(cachep)))
mm/slab.c: if (unlikely(ZERO_OR_NULL_PTR(objp)))
mm/slab.c: if (unlikely(ZERO_OR_NULL_PTR(objp)))
mm/slob.c: if (unlikely(ZERO_OR_NULL_PTR(block)))
mm/slob.c: if (unlikely(ZERO_OR_NULL_PTR(block)))
mm/slob.c: if (unlikely(ZERO_OR_NULL_PTR(block)))
mm/slub.c: if (unlikely(ZERO_OR_NULL_PTR(s)))
mm/slub.c: if (unlikely(ZERO_OR_NULL_PTR(s)))
mm/slub.c: if (unlikely(ZERO_OR_NULL_PTR(object)))
mm/slub.c: if (unlikely(ZERO_OR_NULL_PTR(x)))
mm/slub.c: if (unlikely(ZERO_OR_NULL_PTR(s)))
mm/slub.c: if (unlikely(ZERO_OR_NULL_PTR(s)))
are we seeing a pattern here? We could stick the unlikely inside
ZERO_OR_NULL_PTR() itself. That's a little bit sleazy though - there might
be future callsites at which it is likely, who knows?
I guess we can stick with the idiotic patch ;)
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists