[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <56679FDC.1080800@huawei.com>
Date: Wed, 9 Dec 2015 11:28:28 +0800
From: Xishi Qiu <qiuxishi@...wei.com>
To: Taku Izumi <izumi.taku@...fujitsu.com>
CC: <linux-kernel@...r.kernel.org>, <linux-mm@...ck.org>,
<akpm@...ux-foundation.org>, <tony.luck@...el.com>,
<kamezawa.hiroyu@...fujitsu.com>, <mel@....ul.ie>,
<dave.hansen@...el.com>, <matt@...eblueprint.co.uk>
Subject: Re: [PATCH v3 2/2] mm: Introduce kernelcore=mirror option
On 2015/12/9 11:19, Taku Izumi wrote:
> This patch extends existing "kernelcore" option and
> introduces kernelcore=mirror option. By specifying
> "mirror" instead of specifying the amount of memory,
> non-mirrored (non-reliable) region will be arranged
> into ZONE_MOVABLE.
>
> v1 -> v2:
> - Refine so that the following case also can be
> handled properly:
>
> Node X: |MMMMMM------MMMMMM--------|
> (legend) M: mirrored -: not mirrrored
>
> In this case, ZONE_NORMAL and ZONE_MOVABLE are
> arranged like bellow:
>
> Node X: |MMMMMM------MMMMMM--------|
> |ooooooxxxxxxooooooxxxxxxxx| ZONE_NORMAL
> |ooooooxxxxxxoooooooo| ZONE_MOVABLE
> (legend) o: present x: absent
>
> v2 -> v3:
> - change the option name from kernelcore=reliable
> into kernelcore=mirror
> - documentation fix so that users can understand
> nn[KMS] and mirror are exclusive
>
> Signed-off-by: Taku Izumi <izumi.taku@...fujitsu.com>
> ---
> Documentation/kernel-parameters.txt | 11 +++-
> mm/page_alloc.c | 110 ++++++++++++++++++++++++++++++++++--
> 2 files changed, 114 insertions(+), 7 deletions(-)
>
> diff --git a/Documentation/kernel-parameters.txt b/Documentation/kernel-parameters.txt
> index f8aae63..b0ffc76 100644
> --- a/Documentation/kernel-parameters.txt
> +++ b/Documentation/kernel-parameters.txt
> @@ -1695,7 +1695,8 @@ bytes respectively. Such letter suffixes can also be entirely omitted.
>
> keepinitrd [HW,ARM]
>
> - kernelcore=nn[KMG] [KNL,X86,IA-64,PPC] This parameter
> + kernelcore= Format: nn[KMG] | "mirror"
> + [KNL,X86,IA-64,PPC] This parameter
> specifies the amount of memory usable by the kernel
> for non-movable allocations. The requested amount is
> spread evenly throughout all nodes in the system. The
> @@ -1711,6 +1712,14 @@ bytes respectively. Such letter suffixes can also be entirely omitted.
> use the HighMem zone if it exists, and the Normal
> zone if it does not.
>
> + Instead of specifying the amount of memory (nn[KMS]),
> + you can specify "mirror" option. In case "mirror"
> + option is specified, mirrored (reliable) memory is used
> + for non-movable allocations and remaining memory is used
> + for Movable pages. nn[KMS] and "mirror" are exclusive,
> + so you can NOT specify nn[KMG] and "mirror" at the same
> + time.
> +
Hi Taku,
How about add some comment, if mirrored memroy is too small, then the
normal zone is small, so it may be oom.
The mirrored memory is at least 1/64 of whole memory, because struct
pages usually take 64 bytes per page.
Thanks,
Xishi Qiu
> kgdbdbgp= [KGDB,HW] kgdb over EHCI usb debug port.
> Format: <Controller#>[,poll interval]
> The controller # is the number of the ehci usb debug
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index acb0b4e..4157476 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -251,6 +251,7 @@ static unsigned long __meminitdata arch_zone_highest_possible_pfn[MAX_NR_ZONES];
> static unsigned long __initdata required_kernelcore;
> static unsigned long __initdata required_movablecore;
> static unsigned long __meminitdata zone_movable_pfn[MAX_NUMNODES];
> +static bool mirrored_kernelcore;
>
> /* movable_zone is the "real" zone pages in ZONE_MOVABLE are taken from */
> int movable_zone;
> @@ -4472,6 +4473,7 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone,
> unsigned long pfn;
> struct zone *z;
> unsigned long nr_initialised = 0;
> + struct memblock_region *r = NULL, *tmp;
>
> if (highest_memmap_pfn < end_pfn - 1)
> highest_memmap_pfn = end_pfn - 1;
> @@ -4491,6 +4493,38 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone,
> if (!update_defer_init(pgdat, pfn, end_pfn,
> &nr_initialised))
> break;
> +
> + /*
> + * if not mirrored_kernelcore and ZONE_MOVABLE exists,
> + * range from zone_movable_pfn[nid] to end of each node
> + * should be ZONE_MOVABLE not ZONE_NORMAL. skip it.
> + */
> + if (!mirrored_kernelcore && zone_movable_pfn[nid])
> + if (zone == ZONE_NORMAL &&
> + pfn >= zone_movable_pfn[nid])
> + continue;
> +
> + /*
> + * check given memblock attribute by firmware which
> + * can affect kernel memory layout.
> + * if zone==ZONE_MOVABLE but memory is mirrored,
> + * it's an overlapped memmap init. skip it.
> + */
> + if (mirrored_kernelcore && zone == ZONE_MOVABLE) {
> + if (!r ||
> + pfn >= memblock_region_memory_end_pfn(r)) {
> + for_each_memblock(memory, tmp)
> + if (pfn < memblock_region_memory_end_pfn(tmp))
> + break;
> + r = tmp;
> + }
> + if (pfn >= memblock_region_memory_base_pfn(r) &&
> + memblock_is_mirror(r)) {
> + /* already initialized as NORMAL */
> + pfn = memblock_region_memory_end_pfn(r);
> + continue;
> + }
> + }
> }
>
> /*
> @@ -4909,11 +4943,6 @@ static void __meminit adjust_zone_range_for_zone_movable(int nid,
> *zone_end_pfn = min(node_end_pfn,
> arch_zone_highest_possible_pfn[movable_zone]);
>
> - /* Adjust for ZONE_MOVABLE starting within this range */
> - } else if (*zone_start_pfn < zone_movable_pfn[nid] &&
> - *zone_end_pfn > zone_movable_pfn[nid]) {
> - *zone_end_pfn = zone_movable_pfn[nid];
> -
> /* Check if this whole range is within ZONE_MOVABLE */
> } else if (*zone_start_pfn >= zone_movable_pfn[nid])
> *zone_start_pfn = *zone_end_pfn;
> @@ -4998,6 +5027,7 @@ static unsigned long __meminit zone_absent_pages_in_node(int nid,
> unsigned long zone_low = arch_zone_lowest_possible_pfn[zone_type];
> unsigned long zone_high = arch_zone_highest_possible_pfn[zone_type];
> unsigned long zone_start_pfn, zone_end_pfn;
> + unsigned long nr_absent;
>
> /* When hotadd a new node from cpu_up(), the node should be empty */
> if (!node_start_pfn && !node_end_pfn)
> @@ -5009,7 +5039,39 @@ static unsigned long __meminit zone_absent_pages_in_node(int nid,
> adjust_zone_range_for_zone_movable(nid, zone_type,
> node_start_pfn, node_end_pfn,
> &zone_start_pfn, &zone_end_pfn);
> - return __absent_pages_in_range(nid, zone_start_pfn, zone_end_pfn);
> + nr_absent = __absent_pages_in_range(nid, zone_start_pfn, zone_end_pfn);
> +
> + /*
> + * ZONE_MOVABLE handling.
> + * Treat pages to be ZONE_MOVABLE in ZONE_NORMAL as absent pages
> + * and vice versa.
> + */
> + if (zone_movable_pfn[nid]) {
> + if (mirrored_kernelcore) {
> + unsigned long start_pfn, end_pfn;
> + struct memblock_region *r;
> +
> + for_each_memblock(memory, r) {
> + start_pfn = clamp(memblock_region_memory_base_pfn(r),
> + zone_start_pfn, zone_end_pfn);
> + end_pfn = clamp(memblock_region_memory_end_pfn(r),
> + zone_start_pfn, zone_end_pfn);
> +
> + if (zone_type == ZONE_MOVABLE &&
> + memblock_is_mirror(r))
> + nr_absent += end_pfn - start_pfn;
> +
> + if (zone_type == ZONE_NORMAL &&
> + !memblock_is_mirror(r))
> + nr_absent += end_pfn - start_pfn;
> + }
> + } else {
> + if (zone_type == ZONE_NORMAL)
> + nr_absent += node_end_pfn - zone_movable_pfn[nid];
> + }
> + }
> +
> + return nr_absent;
> }
>
> #else /* CONFIG_HAVE_MEMBLOCK_NODE_MAP */
> @@ -5507,6 +5569,36 @@ static void __init find_zone_movable_pfns_for_nodes(void)
> }
>
> /*
> + * If kernelcore=mirror is specified, ignore movablecore option
> + */
> + if (mirrored_kernelcore) {
> + bool mem_below_4gb_not_mirrored = false;
> +
> + for_each_memblock(memory, r) {
> + if (memblock_is_mirror(r))
> + continue;
> +
> + nid = r->nid;
> +
> + usable_startpfn = memblock_region_memory_base_pfn(r);
> +
> + if (usable_startpfn < 0x100000) {
> + mem_below_4gb_not_mirrored = true;
> + continue;
> + }
> +
> + zone_movable_pfn[nid] = zone_movable_pfn[nid] ?
> + min(usable_startpfn, zone_movable_pfn[nid]) :
> + usable_startpfn;
> + }
> +
> + if (mem_below_4gb_not_mirrored)
> + pr_warn("This configuration results in unmirrored kernel memory.");
> +
> + goto out2;
> + }
> +
> + /*
> * If movablecore=nn[KMG] was specified, calculate what size of
> * kernelcore that corresponds so that memory usable for
> * any allocation type is evenly spread. If both kernelcore
> @@ -5766,6 +5858,12 @@ static int __init cmdline_parse_core(char *p, unsigned long *core)
> */
> static int __init cmdline_parse_kernelcore(char *p)
> {
> + /* parse kernelcore=mirror */
> + if (parse_option_str(p, "mirror")) {
> + mirrored_kernelcore = true;
> + return 0;
> + }
> +
> return cmdline_parse_core(p, &required_kernelcore);
> }
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists