lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20150213031012.GH6592@js1304-P5Q-DELUXE>
Date:	Fri, 13 Feb 2015 12:10:12 +0900
From:	Joonsoo Kim <iamjoonsoo.kim@....com>
To:	Stefan Strogin <s.strogin@...tner.samsung.com>
Cc:	linux-mm@...ck.org, linux-kernel@...r.kernel.org,
	Andrew Morton <akpm@...ux-foundation.org>,
	Marek Szyprowski <m.szyprowski@...sung.com>,
	Michal Nazarewicz <mina86@...a86.com>,
	aneesh.kumar@...ux.vnet.ibm.com,
	Laurent Pinchart <laurent.pinchart@...asonboard.com>,
	Dmitry Safonov <d.safonov@...tner.samsung.com>,
	Pintu Kumar <pintu.k@...sung.com>,
	Weijie Yang <weijie.yang@...sung.com>,
	Laura Abbott <lauraa@...eaurora.org>,
	SeongJae Park <sj38.park@...il.com>,
	Hui Zhu <zhuhui@...omi.com>, Minchan Kim <minchan@...nel.org>,
	Dyasly Sergey <s.dyasly@...sung.com>,
	Vyacheslav Tyrtov <v.tyrtov@...sung.com>,
	gregory.0xf0@...il.com, sasha.levin@...cle.com, gioh.kim@....com,
	pavel@....cz, stefan.strogin@...il.com
Subject: Re: [PATCH 1/4] mm: cma: add currently allocated CMA buffers list to
 debugfs

On Fri, Feb 13, 2015 at 01:15:41AM +0300, Stefan Strogin wrote:
> /sys/kernel/debug/cma/cma-<N>/buffers contains a list of currently allocated
> CMA buffers for CMA region N when CONFIG_CMA_DEBUGFS is enabled.
> 
> Format is:
> 
> <base_phys_addr> - <end_phys_addr> (<size> kB), allocated by <PID> (<comm>)
>  <stack backtrace when the buffer had been allocated>
> 
> Signed-off-by: Stefan Strogin <s.strogin@...tner.samsung.com>
> ---
>  include/linux/cma.h |   9 ++++
>  mm/cma.c            |   9 ++++
>  mm/cma.h            |  16 ++++++
>  mm/cma_debug.c      | 145 +++++++++++++++++++++++++++++++++++++++++++++++++++-
>  4 files changed, 178 insertions(+), 1 deletion(-)
> 
> diff --git a/include/linux/cma.h b/include/linux/cma.h
> index 9384ba6..4c2c83c 100644
> --- a/include/linux/cma.h
> +++ b/include/linux/cma.h
> @@ -28,4 +28,13 @@ extern int cma_init_reserved_mem(phys_addr_t base,
>  					struct cma **res_cma);
>  extern struct page *cma_alloc(struct cma *cma, int count, unsigned int align);
>  extern bool cma_release(struct cma *cma, struct page *pages, int count);
> +
> +#ifdef CONFIG_CMA_DEBUGFS
> +extern int cma_buffer_list_add(struct cma *cma, unsigned long pfn, int count);
> +extern void cma_buffer_list_del(struct cma *cma, unsigned long pfn, int count);
> +#else
> +#define cma_buffer_list_add(cma, pfn, count) { }
> +#define cma_buffer_list_del(cma, pfn, count) { }
> +#endif
> +

These could be in mm/cma.h rather than include/linux/cma.h.

>  #endif
> diff --git a/mm/cma.c b/mm/cma.c
> index 2609e20..ed269b0 100644
> --- a/mm/cma.c
> +++ b/mm/cma.c
> @@ -34,6 +34,9 @@
>  #include <linux/cma.h>
>  #include <linux/highmem.h>
>  #include <linux/io.h>
> +#include <linux/list.h>
> +#include <linux/proc_fs.h>
> +#include <linux/time.h>
>  
>  #include "cma.h"
>  
> @@ -125,6 +128,8 @@ static int __init cma_activate_area(struct cma *cma)
>  #ifdef CONFIG_CMA_DEBUGFS
>  	INIT_HLIST_HEAD(&cma->mem_head);
>  	spin_lock_init(&cma->mem_head_lock);
> +	INIT_LIST_HEAD(&cma->buffers_list);
> +	mutex_init(&cma->list_lock);
>  #endif
>  
>  	return 0;
> @@ -408,6 +413,9 @@ struct page *cma_alloc(struct cma *cma, int count, unsigned int align)
>  		start = bitmap_no + mask + 1;
>  	}
>  
> +	if (page)
> +		cma_buffer_list_add(cma, pfn, count);
> +
>  	pr_debug("%s(): returned %p\n", __func__, page);
>  	return page;
>  }
> @@ -440,6 +448,7 @@ bool cma_release(struct cma *cma, struct page *pages, int count)
>  
>  	free_contig_range(pfn, count);
>  	cma_clear_bitmap(cma, pfn, count);
> +	cma_buffer_list_del(cma, pfn, count);
>  
>  	return true;
>  }
> diff --git a/mm/cma.h b/mm/cma.h
> index 1132d73..98e5f79 100644
> --- a/mm/cma.h
> +++ b/mm/cma.h
> @@ -1,6 +1,8 @@
>  #ifndef __MM_CMA_H__
>  #define __MM_CMA_H__
>  
> +#include <linux/sched.h>
> +
>  struct cma {
>  	unsigned long   base_pfn;
>  	unsigned long   count;
> @@ -10,9 +12,23 @@ struct cma {
>  #ifdef CONFIG_CMA_DEBUGFS
>  	struct hlist_head mem_head;
>  	spinlock_t mem_head_lock;
> +	struct list_head buffers_list;
> +	struct mutex	list_lock;
>  #endif
>  };
>  
> +#ifdef CONFIG_CMA_DEBUGFS
> +struct cma_buffer {
> +	unsigned long pfn;
> +	unsigned long count;
> +	pid_t pid;
> +	char comm[TASK_COMM_LEN];
> +	unsigned long trace_entries[16];
> +	unsigned int nr_entries;
> +	struct list_head list;
> +};
> +#endif
> +
>  extern struct cma cma_areas[MAX_CMA_AREAS];
>  extern unsigned cma_area_count;
>  
> diff --git a/mm/cma_debug.c b/mm/cma_debug.c
> index 7e1d325..5acd937 100644
> --- a/mm/cma_debug.c
> +++ b/mm/cma_debug.c
> @@ -2,6 +2,7 @@
>   * CMA DebugFS Interface
>   *
>   * Copyright (c) 2015 Sasha Levin <sasha.levin@...cle.com>
> + * Copyright (c) 2015 Stefan Strogin <stefan.strogin@...il.com>
>   */
>   
>  
> @@ -10,6 +11,8 @@
>  #include <linux/list.h>
>  #include <linux/kernel.h>
>  #include <linux/slab.h>
> +#include <linux/mm_types.h>
> +#include <linux/stacktrace.h>
>  
>  #include "cma.h"
>  
> @@ -21,6 +24,99 @@ struct cma_mem {
>  
>  static struct dentry *cma_debugfs_root;
>  
> +/* Must be called under cma->list_lock */
> +static int __cma_buffer_list_add(struct cma *cma, unsigned long pfn, int count)
> +{
> +	struct cma_buffer *cmabuf;
> +	struct stack_trace trace;
> +
> +	cmabuf = kmalloc(sizeof(*cmabuf), GFP_KERNEL);
> +	if (!cmabuf) {
> +		pr_warn("%s(page %p, count %d): failed to allocate buffer list entry\n",
> +			__func__, pfn_to_page(pfn), count);
> +		return -ENOMEM;
> +	}
> +
> +	trace.nr_entries = 0;
> +	trace.max_entries = ARRAY_SIZE(cmabuf->trace_entries);
> +	trace.entries = &cmabuf->trace_entries[0];
> +	trace.skip = 2;
> +	save_stack_trace(&trace);
> +
> +	cmabuf->pfn = pfn;
> +	cmabuf->count = count;
> +	cmabuf->pid = task_pid_nr(current);
> +	cmabuf->nr_entries = trace.nr_entries;
> +	get_task_comm(cmabuf->comm, current);
> +
> +	list_add_tail(&cmabuf->list, &cma->buffers_list);
> +
> +	return 0;
> +}
> +
> +/**
> + * cma_buffer_list_add() - add a new entry to a list of allocated buffers
> + * @cma:     Contiguous memory region for which the allocation is performed.
> + * @pfn:     Base PFN of the allocated buffer.
> + * @count:   Number of allocated pages.
> + *
> + * This function adds a new entry to the list of allocated contiguous memory
> + * buffers in a CMA area. It uses the CMA area specificated by the device
> + * if available or the default global one otherwise.
> + */
> +int cma_buffer_list_add(struct cma *cma, unsigned long pfn, int count)
> +{
> +	int ret;
> +
> +	mutex_lock(&cma->list_lock);
> +	ret = __cma_buffer_list_add(cma, pfn, count);
> +	mutex_unlock(&cma->list_lock);
> +
> +	return ret;
> +}
> +
> +/**
> + * cma_buffer_list_del() - delete an entry from a list of allocated buffers
> + * @cma:   Contiguous memory region for which the allocation was performed.
> + * @pfn:   Base PFN of the released buffer.
> + * @count: Number of pages.
> + *
> + * This function deletes a list entry added by cma_buffer_list_add().
> + */
> +void cma_buffer_list_del(struct cma *cma, unsigned long pfn, int count)
> +{
> +	struct cma_buffer *cmabuf, *tmp;
> +	int found = 0;
> +	unsigned long buf_end_pfn, free_end_pfn = pfn + count;
> +
> +	mutex_lock(&cma->list_lock);
> +	list_for_each_entry_safe(cmabuf, tmp, &cma->buffers_list, list) {
> +
> +		buf_end_pfn = cmabuf->pfn + cmabuf->count;
> +		if (pfn <= cmabuf->pfn && free_end_pfn >= buf_end_pfn) {
> +			list_del(&cmabuf->list);
> +			kfree(cmabuf);
> +			found = 1;
> +		} else if (pfn <= cmabuf->pfn && free_end_pfn < buf_end_pfn) {
> +			cmabuf->count -= free_end_pfn - cmabuf->pfn;
> +			cmabuf->pfn = free_end_pfn;
> +			found = 1;
> +		} else if (pfn > cmabuf->pfn && pfn < buf_end_pfn) {
> +			if (free_end_pfn < buf_end_pfn)
> +				__cma_buffer_list_add(cma, free_end_pfn,
> +						buf_end_pfn - free_end_pfn);
> +			cmabuf->count = pfn - cmabuf->pfn;
> +			found = 1;
> +		}
> +	}
> +	mutex_unlock(&cma->list_lock);

This linear searching make cma_release() slow if we have many allocated
cma buffers. It wouldn't cause any problem?

Thanks.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ