lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20080819174436.db09de1c.kamezawa.hiroyu@jp.fujitsu.com>
Date:	Tue, 19 Aug 2008 17:44:36 +0900
From:	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
To:	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
Cc:	LKML <linux-kernel@...r.kernel.org>,
	"balbir@...ux.vnet.ibm.com" <balbir@...ux.vnet.ibm.com>,
	"yamamoto@...inux.co.jp" <yamamoto@...inux.co.jp>,
	"nishimura@....nes.nec.co.jp" <nishimura@....nes.nec.co.jp>,
	ryov@...inux.co.jp
Subject: [PATCH -mm][preview] memcg: a patch series for next [9/9]

Add control files to mem+swap controller.

This patch adds following 2 files.
  - memory.memsw_limit_in_bytes ..... limit for mem+swap usage.
  - memory.swap_usage_in_bytes  ..... usage for swap_entry.

Following rules must be kept.
	memory.memsw_limit_in_bytes >= memory.limit_in_bytes.

If not, -EINVAL will return.

TODO:
  - add Documentation.
  - add function/file to force swap-in for reducing swap usage.

Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>

---
 mm/memcontrol.c |  100 ++++++++++++++++++++++++++++++++++++++++++++++++++++----
 1 file changed, 93 insertions(+), 7 deletions(-)

Index: linux-2.6.27-rc1-mm1/mm/memcontrol.c
===================================================================
--- linux-2.6.27-rc1-mm1.orig/mm/memcontrol.c
+++ linux-2.6.27-rc1-mm1/mm/memcontrol.c
@@ -268,10 +268,11 @@ enum {
 	MEMCG_FILE_TYPE_PAGE_USAGE,
 	MEMCG_FILE_TYPE_FAILCNT,
 	MEMCG_FILE_TYPE_MAX_USAGE,
+	MEMCG_FILE_TYPE_MEMSW_LIMIT,
+	MEMCG_FILE_TYPE_SWAP_USAGE,
 };
 
 
-
 /*
  * Always modified under lru lock. Then, not necessary to preempt_disable()
  * "flags" passed to this function is a copy of pc->flags but flags checked
@@ -415,11 +416,11 @@ mem_counter_recharge_swapout(struct mem_
 }
 
 static inline void
-mem_counter_uncharge_swap(struct mem_cgroup *memcg, long num)
+mem_counter_uncharge_swap(struct mem_cgroup *memcg)
 {
 	unsigned long flags;
 	spin_lock_irqsave(&memcg->res.lock, flags);
-	memcg->res.swaps -= num;
+	memcg->res.swaps -= 1;
 	spin_unlock_irqrestore(&memcg->res.lock, flags);
 }
 
@@ -430,7 +431,9 @@ static int mem_counter_set_pages_limit(s
 	int ret = -EBUSY;
 
 	spin_lock_irqsave(&memcg->res.lock, flags);
-	if (memcg->res.pages < lim) {
+	if (lim > memcg->res.memsw_limit)
+		ret = -EINVAL;
+	else if (memcg->res.pages < lim) {
 		memcg->res.pages_limit = lim;
 		ret = 0;
 	}
@@ -568,6 +571,25 @@ void mem_cgroup_move_lists(struct page *
 }
 
 #ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP
+
+static int mem_cgroup_set_memsw_limit(struct mem_cgroup *memcg,
+		unsigned long lim)
+{
+	unsigned long flags;
+	int ret = -EBUSY;
+
+        spin_lock_irqsave(&memcg->res.lock, flags);
+        if (memcg->res.pages_limit > lim)
+                ret = -EINVAL;
+        else if (memcg->res.pages + memcg->res.swaps < lim) {
+		memcg->res.memsw_limit = lim;
+		ret = 0;
+	}
+        spin_unlock_irqrestore(&memcg->res.lock, flags);
+
+        return ret;
+
+}
 /*
  * Create a space for remember swap_entry.
  * Called from get_swap_page().
@@ -666,7 +688,7 @@ static void swap_cgroup_uncharge_swap(st
 
 	if (!swap_accounted(sc))
 		return;
-	mem_counter_uncharge_swap(mem, 1);
+	mem_counter_uncharge_swap(mem);
 	clear_swap_accounted(sc);
 }
 
@@ -686,7 +708,7 @@ static void swap_cgroup_delete_swap(swp_
 		list_del(&sc->list);
 		spin_unlock_irqrestore(&memcg->swap_list_lock, flags);
 		if (swap_accounted(sc))
-			mem_counter_uncharge_swap(memcg, 1);
+			mem_counter_uncharge_swap(memcg);
 		css_put(&memcg->css);
 		kfree(sc);
 	}
@@ -1294,7 +1316,10 @@ int mem_cgroup_resize_limit(struct mem_c
 	int ret = 0;
 	unsigned long pages = (unsigned long)(val >> PAGE_SHIFT);
 
-	while (mem_counter_set_pages_limit(memcg, pages)) {
+	while (1) {
+		ret = mem_counter_set_pages_limit(memcg, pages);
+		if (!ret || ret == -EINVAL)
+			break;
 		if (signal_pending(current)) {
 			ret = -EINTR;
 			break;
@@ -1310,6 +1335,43 @@ int mem_cgroup_resize_limit(struct mem_c
 	return ret;
 }
 
+#ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP
+static int
+mem_cgroup_resize_memsw_limit(struct mem_cgroup *memcg, unsigned long long val)
+{
+	int retry_count = MEM_CGROUP_RECLAIM_RETRIES;
+	int progress;
+	int ret = 0;
+	unsigned long pages = (unsigned long)(val >> PAGE_SHIFT);
+
+	while (1) {
+        	ret = mem_cgroup_set_memsw_limit(memcg, pages);
+		if (!ret || ret == -EINVAL)
+			break;
+                if (signal_pending(current)) {
+                        ret = -EINTR;
+                        break;
+                }
+                if (!retry_count) {
+                        ret = -EBUSY;
+                        break;
+                }
+                progress = try_to_free_mem_cgroup_pages(memcg, GFP_KERNEL);
+                if (!progress)
+                        retry_count--;
+        }
+        return ret;
+
+}
+#else
+static int
+mem_cgroup_resize_memsw_limit(struct mem_cgroup *memcg, unsigned long long val)
+{
+	return -EINVAL;
+}
+#endif
+
+
 
 /*
  * This routine traverse page_cgroup in given list and drop them all.
@@ -1405,6 +1467,12 @@ static u64 mem_cgroup_read(struct cgroup
 	case MEMCG_FILE_TYPE_FAILCNT:
 		ret = memcg->res.failcnt << PAGE_SHIFT;
 		break;
+	case MEMCG_FILE_TYPE_SWAP_USAGE:
+		ret = memcg->res.swaps << PAGE_SHIFT;
+		break;
+	case MEMCG_FILE_TYPE_MEMSW_LIMIT:
+		ret = memcg->res.memsw_limit << PAGE_SHIFT;
+		break;
 	default:
 		BUG();
 	}
@@ -1441,6 +1509,11 @@ static int mem_cgroup_write(struct cgrou
 		if (!ret)
 			ret = mem_cgroup_resize_limit(memcg, val);
 		break;
+	case MEMCG_FILE_TYPE_MEMSW_LIMIT:
+		ret = call_memparse(buffer, &val);
+		if (!ret)
+			ret = mem_cgroup_resize_memsw_limit(memcg, val);
+		break;
 	default:
 		ret = -EINVAL; /* should be BUG() ? */
 		break;
@@ -1552,6 +1625,19 @@ static struct cftype mem_cgroup_files[] 
 		.name = "stat",
 		.read_map = mem_control_stat_show,
 	},
+#ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP
+	{
+		.name = "memsw_limit_in_bytes",
+		.private = MEMCG_FILE_TYPE_MEMSW_LIMIT,
+		.read_u64 = mem_cgroup_read,
+		.write_string = mem_cgroup_write,
+	},
+	{
+		.name = "swap_usage_in_bytes",
+		.private = MEMCG_FILE_TYPE_SWAP_USAGE,
+		.read_u64 = mem_cgroup_read,
+	}
+#endif
 };
 
 static int alloc_mem_cgroup_per_zone_info(struct mem_cgroup *mem, int node)

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ