lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date:	Sun, 3 May 2009 15:14:32 +0300 (EEST)
From:	Pekka J Enberg <penberg@...helsinki.fi>
To:	npiggin@...e.de
cc:	linux-kernel@...r.kernel.org
Subject: [PATCH] SLQB: Use _RET_IP_ instead of __builtin_return_address(0)

From: Pekka Enberg <penberg@...helsinki.fi>

This patch changes SLQB to use the shorter _RET_IP_ form for consistency
with SLUB and to prepare SQLB for eventual kmemtrace hooks.

Cc: Nick Piggin <npiggin@...e.de>
Signed-off-by: Pekka Enberg <penberg@...helsinki.fi>
---
 mm/slqb.c |   28 ++++++++++++++--------------
 1 files changed, 14 insertions(+), 14 deletions(-)

diff --git a/mm/slqb.c b/mm/slqb.c
index 474fd2f..f538f2c 100644
--- a/mm/slqb.c
+++ b/mm/slqb.c
@@ -246,7 +246,7 @@ static LIST_HEAD(slab_caches);
  * Tracking user of a slab.
  */
 struct track {
-	void *addr;		/* Called from address */
+	unsigned long addr;	/* Called from address */
 	int cpu;		/* Was running on cpu */
 	int pid;		/* Pid context */
 	unsigned long when;	/* When did the operation occur */
@@ -385,7 +385,7 @@ static struct track *get_track(struct kmem_cache *s, void *object,
 }
 
 static void set_track(struct kmem_cache *s, void *object,
-				enum track_item alloc, void *addr)
+				enum track_item alloc, unsigned long addr)
 {
 	struct track *p;
 
@@ -409,8 +409,8 @@ static void init_tracking(struct kmem_cache *s, void *object)
 	if (!(s->flags & SLAB_STORE_USER))
 		return;
 
-	set_track(s, object, TRACK_FREE, NULL);
-	set_track(s, object, TRACK_ALLOC, NULL);
+	set_track(s, object, TRACK_FREE, 0UL);
+	set_track(s, object, TRACK_ALLOC, 0UL);
 }
 
 static void print_track(const char *s, struct track *t)
@@ -753,7 +753,7 @@ static void setup_object_debug(struct kmem_cache *s, struct slqb_page *page,
 }
 
 static int alloc_debug_processing(struct kmem_cache *s,
-					void *object, void *addr)
+					void *object, unsigned long addr)
 {
 	struct slqb_page *page;
 	page = virt_to_head_slqb_page(object);
@@ -781,7 +781,7 @@ bad:
 }
 
 static int free_debug_processing(struct kmem_cache *s,
-					void *object, void *addr)
+					void *object, unsigned long addr)
 {
 	struct slqb_page *page;
 	page = virt_to_head_slqb_page(object);
@@ -1528,7 +1528,7 @@ try_remote:
  * (debug checking and memset()ing).
  */
 static __always_inline void *slab_alloc(struct kmem_cache *s,
-				gfp_t gfpflags, int node, void *addr)
+				gfp_t gfpflags, int node, unsigned long addr)
 {
 	void *object;
 	unsigned long flags;
@@ -1550,7 +1550,7 @@ again:
 }
 
 static __always_inline void *__kmem_cache_alloc(struct kmem_cache *s,
-				gfp_t gfpflags, void *caller)
+				gfp_t gfpflags, unsigned long caller)
 {
 	int node = -1;
 #ifdef CONFIG_NUMA
@@ -1562,14 +1562,14 @@ static __always_inline void *__kmem_cache_alloc(struct kmem_cache *s,
 
 void *kmem_cache_alloc(struct kmem_cache *s, gfp_t gfpflags)
 {
-	return __kmem_cache_alloc(s, gfpflags, __builtin_return_address(0));
+	return __kmem_cache_alloc(s, gfpflags, _RET_IP_);
 }
 EXPORT_SYMBOL(kmem_cache_alloc);
 
 #ifdef CONFIG_NUMA
 void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node)
 {
-	return slab_alloc(s, gfpflags, node, __builtin_return_address(0));
+	return slab_alloc(s, gfpflags, node, _RET_IP_);
 }
 EXPORT_SYMBOL(kmem_cache_alloc_node);
 #endif
@@ -1757,7 +1757,7 @@ static __always_inline void slab_free(struct kmem_cache *s,
 
 	debug_check_no_locks_freed(object, s->objsize);
 	if (likely(object) && unlikely(slab_debug(s))) {
-		if (unlikely(!free_debug_processing(s, object, __builtin_return_address(0))))
+		if (unlikely(!free_debug_processing(s, object, _RET_IP_)))
 			return;
 	}
 
@@ -2528,7 +2528,7 @@ void *__kmalloc(size_t size, gfp_t flags)
 	if (unlikely(ZERO_OR_NULL_PTR(s)))
 		return s;
 
-	return __kmem_cache_alloc(s, flags, __builtin_return_address(0));
+	return __kmem_cache_alloc(s, flags, _RET_IP_);
 }
 EXPORT_SYMBOL(__kmalloc);
 
@@ -3124,7 +3124,7 @@ void *__kmalloc_track_caller(size_t size, gfp_t flags, unsigned long caller)
 	if (unlikely(current->flags & (PF_SPREAD_SLAB | PF_MEMPOLICY)))
 		node = alternate_nid(s, flags, node);
 #endif
-	return slab_alloc(s, flags, node, (void *)caller);
+	return slab_alloc(s, flags, node, caller);
 }
 
 void *__kmalloc_node_track_caller(size_t size, gfp_t flags, int node,
@@ -3136,7 +3136,7 @@ void *__kmalloc_node_track_caller(size_t size, gfp_t flags, int node,
 	if (unlikely(ZERO_OR_NULL_PTR(s)))
 		return s;
 
-	return slab_alloc(s, flags, node, (void *)caller);
+	return slab_alloc(s, flags, node, caller);
 }
 #endif
 
-- 
1.5.6.3

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists