[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1424721263-25314-2-git-send-email-sj38.park@gmail.com>
Date: Tue, 24 Feb 2015 04:54:19 +0900
From: SeongJae Park <sj38.park@...il.com>
To: akpm@...ux-foundation.org
Cc: lauraa@...eaurora.org, minchan@...nel.org,
sergey.senozhatsky@...il.com, linux-mm@...ck.org,
linux-kernel@...r.kernel.org, SeongJae Park <sj38.park@...il.com>
Subject: [RFC v2 1/5] gcma: introduce contiguous memory allocator
Currently, cma reserves large contiguous memory area during early boot
and let the area could be used by others for movable pages only. Then,
if those movable pages are necessary for contiguous memory allocation,
cma migrates and / or discards them out.
This mechanism have two weakness.
1) Because any one in kernel could pin any movable pages, migration of
movable pages could be fail. It could lead to contiguous memory
allocation failure.
2) Because of migration / reclaim overhead, the latency could be
extremely high.
In short, cma doesn't guarantee success and fast latency of contiguous
memory allocation. The problem was discussed in detail from [1] and [2].
This patch introduces a simple contiguous memory allocator, namely,
GCMA(Guaranteed Contiguous Memory Allocator). It aims to guarantee
success and fast latency by reserving memory during early boot time.
However, this simple mechanism could degrade system memory space
efficiency seriously; following commits will settle the problem.
[1] https://lkml.org/lkml/2013/10/30/16
[2] http://sched.co/1qZcBAO
Signed-off-by: SeongJae Park <sj38.park@...il.com>
---
include/linux/gcma.h | 26 ++++++++
mm/gcma.c | 173 +++++++++++++++++++++++++++++++++++++++++++++++++++
2 files changed, 199 insertions(+)
create mode 100644 include/linux/gcma.h
create mode 100644 mm/gcma.c
diff --git a/include/linux/gcma.h b/include/linux/gcma.h
new file mode 100644
index 0000000..cda481f
--- /dev/null
+++ b/include/linux/gcma.h
@@ -0,0 +1,26 @@
+/*
+ * gcma.h - Guaranteed Contiguous Memory Allocator
+ *
+ * GCMA aims for contiguous memory allocation with success and fast
+ * latency guarantee.
+ * It reserves large amount of memory and let it be allocated to
+ * contiguous memory requests.
+ *
+ * Copyright (C) 2014 LG Electronics Inc.,
+ * Copyright (C) 2014 Minchan Kim <minchan@...nel.org>
+ * Copyright (C) 2014-2015 SeongJae Park <sj38.park@...il.com>
+ */
+
+#ifndef _LINUX_GCMA_H
+#define _LINUX_GCMA_H
+
+struct gcma;
+
+int gcma_init(unsigned long start_pfn, unsigned long size,
+ struct gcma **res_gcma);
+int gcma_alloc_contig(struct gcma *gcma,
+ unsigned long start_pfn, unsigned long size);
+void gcma_free_contig(struct gcma *gcma,
+ unsigned long start_pfn, unsigned long size);
+
+#endif /* _LINUX_GCMA_H */
diff --git a/mm/gcma.c b/mm/gcma.c
new file mode 100644
index 0000000..3f6a337
--- /dev/null
+++ b/mm/gcma.c
@@ -0,0 +1,173 @@
+/*
+ * gcma.c - Guaranteed Contiguous Memory Allocator
+ *
+ * GCMA aims for contiguous memory allocation with success and fast
+ * latency guarantee.
+ * It reserves large amount of memory and let it be allocated to
+ * contiguous memory requests.
+ *
+ * Copyright (C) 2014 LG Electronics Inc.,
+ * Copyright (C) 2014 Minchan Kim <minchan@...nel.org>
+ * Copyright (C) 2014-2015 SeongJae Park <sj38.park@...il.com>
+ */
+
+#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
+
+#include <linux/gcma.h>
+#include <linux/highmem.h>
+#include <linux/module.h>
+#include <linux/slab.h>
+
+struct gcma {
+ spinlock_t lock;
+ unsigned long *bitmap;
+ unsigned long base_pfn, size;
+ struct list_head list;
+};
+
+struct gcma_info {
+ spinlock_t lock; /* protect list */
+ struct list_head head;
+};
+
+static struct gcma_info ginfo = {
+ .head = LIST_HEAD_INIT(ginfo.head),
+ .lock = __SPIN_LOCK_UNLOCKED(ginfo.lock),
+};
+
+/*
+ * gcma_init - initializes a contiguous memory area
+ *
+ * @start_pfn start pfn of contiguous memory area
+ * @size number of pages in the contiguous memory area
+ * @res_gcma pointer to store the created gcma region
+ *
+ * Returns 0 on success, error code on failure.
+ */
+int gcma_init(unsigned long start_pfn, unsigned long size,
+ struct gcma **res_gcma)
+{
+ int bitmap_size = BITS_TO_LONGS(size) * sizeof(long);
+ struct gcma *gcma;
+
+ gcma = kmalloc(sizeof(*gcma), GFP_KERNEL);
+ if (!gcma)
+ goto out;
+
+ gcma->bitmap = kzalloc(bitmap_size, GFP_KERNEL);
+ if (!gcma->bitmap)
+ goto free_cma;
+
+ gcma->size = size;
+ gcma->base_pfn = start_pfn;
+ spin_lock_init(&gcma->lock);
+
+ spin_lock(&ginfo.lock);
+ list_add(&gcma->list, &ginfo.head);
+ spin_unlock(&ginfo.lock);
+
+ *res_gcma = gcma;
+ pr_info("initialized gcma area [%lu, %lu]\n",
+ start_pfn, start_pfn + size);
+ return 0;
+
+free_cma:
+ kfree(gcma);
+out:
+ return -ENOMEM;
+}
+
+static struct page *gcma_alloc_page(struct gcma *gcma)
+{
+ unsigned long bit;
+ unsigned long *bitmap = gcma->bitmap;
+ struct page *page = NULL;
+
+ spin_lock(&gcma->lock);
+ bit = bitmap_find_next_zero_area(bitmap, gcma->size, 0, 1, 0);
+ if (bit >= gcma->size) {
+ spin_unlock(&gcma->lock);
+ goto out;
+ }
+
+ bitmap_set(bitmap, bit, 1);
+ page = pfn_to_page(gcma->base_pfn + bit);
+ spin_unlock(&gcma->lock);
+
+out:
+ return page;
+}
+
+static void gcma_free_page(struct gcma *gcma, struct page *page)
+{
+ unsigned long pfn, offset;
+
+ pfn = page_to_pfn(page);
+
+ spin_lock(&gcma->lock);
+ offset = pfn - gcma->base_pfn;
+
+ bitmap_clear(gcma->bitmap, offset, 1);
+ spin_unlock(&gcma->lock);
+}
+
+/*
+ * gcma_alloc_contig - allocates contiguous pages
+ *
+ * @start_pfn start pfn of requiring contiguous memory area
+ * @size number of pages in requiring contiguous memory area
+ *
+ * Returns 0 on success, error code on failure.
+ */
+int gcma_alloc_contig(struct gcma *gcma, unsigned long start_pfn,
+ unsigned long size)
+{
+ unsigned long offset;
+
+ spin_lock(&gcma->lock);
+ offset = start_pfn - gcma->base_pfn;
+
+ if (bitmap_find_next_zero_area(gcma->bitmap, gcma->size, offset,
+ size, 0) != 0) {
+ spin_unlock(&gcma->lock);
+ pr_warn("already allocated region required: %lu, %lu",
+ start_pfn, size);
+ return -EINVAL;
+ }
+
+ bitmap_set(gcma->bitmap, offset, size);
+ spin_unlock(&gcma->lock);
+
+ return 0;
+}
+
+/*
+ * gcma_free_contig - free allocated contiguous pages
+ *
+ * @start_pfn start pfn of freeing contiguous memory area
+ * @size number of pages in freeing contiguous memory area
+ */
+void gcma_free_contig(struct gcma *gcma,
+ unsigned long start_pfn, unsigned long size)
+{
+ unsigned long offset;
+
+ spin_lock(&gcma->lock);
+ offset = start_pfn - gcma->base_pfn;
+ bitmap_clear(gcma->bitmap, offset, size);
+ spin_unlock(&gcma->lock);
+}
+
+static int __init init_gcma(void)
+{
+ pr_info("loading gcma\n");
+
+ return 0;
+}
+
+module_init(init_gcma);
+
+MODULE_LICENSE("GPL");
+MODULE_AUTHOR("Minchan Kim <minchan@...nel.org>");
+MODULE_AUTHOR("SeongJae Park <sj38.park@...il.com>");
+MODULE_DESCRIPTION("Guaranteed Contiguous Memory Allocator");
--
1.9.1
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists