[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20251019061631.2235405-7-xiyou.wangcong@gmail.com>
Date: Sat, 18 Oct 2025 23:16:20 -0700
From: Cong Wang <xiyou.wangcong@...il.com>
To: linux-kernel@...r.kernel.org
Cc: jiri@...nulli.us,
stefanha@...hat.com,
multikernel@...ts.linux.dev,
pasha.tatashin@...een.com,
Cong Wang <cwang@...tikernel.io>,
Andrew Morton <akpm@...ux-foundation.org>,
Baoquan He <bhe@...hat.com>,
Alexander Graf <graf@...zon.com>,
Mike Rapoport <rppt@...nel.org>,
Changyuan Lyu <changyuanl@...gle.com>,
kexec@...ts.infradead.org,
linux-mm@...ck.org
Subject: [RFC Patch v2 06/16] multikernel: Introduce physical memory reservation and allocation
From: Cong Wang <cwang@...tikernel.io>
This commit introduces:
* Memory pool reservation via mkkernel_pool= kernel parameter that
supports both fixed address reservation (mkkernel_pool=size@...r) and
dynamic allocation (mkkernel_pool=size). Uses memblock reservation
during early boot to ensure memory availability.
* Global memory pool management using gen_pool for runtime allocation
and deallocation of physical memory chunks from the reserved pool.
Provides thread-safe operations with mutex protection.
* Per-instance memory pool management that allows creating dedicated
memory pools for individual kernel instances. Each instance pool is
carved out from the main multikernel pool and provides fine-grained
allocation capabilities for IPI data, buffers, and other per-instance
resources.
* Integration with /proc/iomem resource hierarchy to provide visibility
into multikernel memory usage and prevent conflicts with other kernel
subsystems.
This memory management system uses a two-tier approach: a main pool
reserved during boot handles large allocations and spawning operations,
while per-instance pools provide efficient small allocation services
for runtime inter-kernel communication and instance-specific data
structures such as kernel image and initramfs.
Signed-off-by: Cong Wang <cwang@...tikernel.io>
---
include/linux/multikernel.h | 22 +++
kernel/multikernel/Makefile | 2 +-
kernel/multikernel/mem.c | 376 ++++++++++++++++++++++++++++++++++++
3 files changed, 399 insertions(+), 1 deletion(-)
create mode 100644 include/linux/multikernel.h
create mode 100644 kernel/multikernel/mem.c
diff --git a/include/linux/multikernel.h b/include/linux/multikernel.h
new file mode 100644
index 000000000000..51c989139a75
--- /dev/null
+++ b/include/linux/multikernel.h
@@ -0,0 +1,22 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * Copyright (C) 2025 Multikernel Technologies, Inc. All rights reserved
+ */
+#ifndef _LINUX_MULTIKERNEL_H
+#define _LINUX_MULTIKERNEL_H
+
+struct resource;
+
+extern phys_addr_t multikernel_alloc(size_t size);
+extern void multikernel_free(phys_addr_t addr, size_t size);
+extern struct resource *multikernel_get_pool_resource(void);
+extern bool multikernel_pool_available(void);
+
+/* Per-instance memory pool management */
+extern void *multikernel_create_instance_pool(int instance_id, size_t pool_size, int min_alloc_order);
+extern void multikernel_destroy_instance_pool(void *pool_handle);
+extern phys_addr_t multikernel_instance_alloc(void *pool_handle, size_t size, size_t align);
+extern void multikernel_instance_free(void *pool_handle, phys_addr_t addr, size_t size);
+extern size_t multikernel_instance_pool_avail(void *pool_handle);
+
+#endif /* _LINUX_MULTIKERNEL_H */
diff --git a/kernel/multikernel/Makefile b/kernel/multikernel/Makefile
index 950bace927a0..0dad7f2267f9 100644
--- a/kernel/multikernel/Makefile
+++ b/kernel/multikernel/Makefile
@@ -3,4 +3,4 @@
# Makefile for multikernel support
#
-obj-y += core.o
+obj-y += core.o mem.o
diff --git a/kernel/multikernel/mem.c b/kernel/multikernel/mem.c
new file mode 100644
index 000000000000..dbc3363764d7
--- /dev/null
+++ b/kernel/multikernel/mem.c
@@ -0,0 +1,376 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * Multikernel memory management
+ *
+ * Memory pool management for multikernel spawn kernels using gen_pool
+ * with mkkernel_pool= command line parameter
+ */
+
+#include <linux/memblock.h>
+#include <linux/ioport.h>
+#include <linux/kexec.h>
+#include <linux/mutex.h>
+#include <linux/genalloc.h>
+#include <linux/io.h>
+#include <asm/e820/api.h>
+#include <linux/multikernel.h>
+
+/* Global multikernel memory pool resource */
+struct resource multikernel_res = {
+ .name = "Multikernel Memory Pool",
+ .start = 0,
+ .end = 0,
+ .flags = IORESOURCE_BUSY | IORESOURCE_MEM,
+ .desc = IORES_DESC_RESERVED
+};
+
+/* Generic pool for runtime memory allocation */
+static struct gen_pool *multikernel_pool;
+
+static DEFINE_MUTEX(multikernel_mem_mutex);
+
+/**
+ * multikernel_alloc() - Allocate memory from multikernel pool
+ * @size: size to allocate
+ *
+ * Returns physical address of allocated memory, or 0 on failure
+ */
+phys_addr_t multikernel_alloc(size_t size)
+{
+ unsigned long addr;
+
+ if (!multikernel_pool)
+ return 0;
+
+ mutex_lock(&multikernel_mem_mutex);
+ addr = gen_pool_alloc(multikernel_pool, size);
+ mutex_unlock(&multikernel_mem_mutex);
+
+ return (phys_addr_t)addr;
+}
+
+/**
+ * multikernel_free() - Free memory back to multikernel pool
+ * @addr: physical address to free
+ * @size: size to free
+ */
+void multikernel_free(phys_addr_t addr, size_t size)
+{
+ if (!multikernel_pool || !addr)
+ return;
+
+ mutex_lock(&multikernel_mem_mutex);
+ gen_pool_free(multikernel_pool, (unsigned long)addr, size);
+ mutex_unlock(&multikernel_mem_mutex);
+
+ pr_debug("Multikernel freed %zu bytes at %pa\n", size, &addr);
+}
+
+/**
+ * multikernel_get_pool_resource() - Get the multikernel pool resource
+ *
+ * Returns pointer to the multikernel pool resource for memory walking
+ */
+struct resource *multikernel_get_pool_resource(void)
+{
+ if (!multikernel_res.start)
+ return NULL;
+
+ return &multikernel_res;
+}
+
+/**
+ * multikernel_pool_available() - Check if multikernel pool is available
+ *
+ * Returns true if multikernel pool is configured and available
+ */
+bool multikernel_pool_available(void)
+{
+ return multikernel_pool != NULL;
+}
+
+/**
+ * Per-instance memory pool management
+ *
+ * Each kernel instance gets its own gen_pool for fine-grained allocations
+ * (IPI data, small buffers, etc.) carved out from the main multikernel pool.
+ */
+
+/**
+ * multikernel_create_instance_pool() - Create a memory pool for a kernel instance
+ * @instance_id: Unique identifier for the instance
+ * @pool_size: Total size of memory to allocate for this instance's pool
+ * @min_alloc_order: Minimum allocation order (at least PAGE_SHIFT)
+ *
+ * Allocates multiple chunks from the main multikernel pool to reach the target
+ * pool_size and creates a gen_pool for the instance to manage smaller allocations.
+ *
+ * Returns opaque handle to the instance pool, or NULL on failure
+ */
+void *multikernel_create_instance_pool(int instance_id, size_t pool_size, int min_alloc_order)
+{
+ struct gen_pool *instance_pool;
+ size_t remaining_size = pool_size;
+ size_t chunk_size;
+ phys_addr_t chunk_base;
+ int chunks_added = 0;
+
+ if (!multikernel_pool_available()) {
+ pr_err("Multikernel main pool not available for instance %d\n", instance_id);
+ return NULL;
+ }
+
+ if (min_alloc_order < PAGE_SHIFT) {
+ pr_err("Invalid min_alloc_order %d for instance %d (must be >= PAGE_SHIFT %d)\n",
+ min_alloc_order, instance_id, PAGE_SHIFT);
+ return NULL;
+ }
+
+ instance_pool = gen_pool_create(min_alloc_order, -1);
+ if (!instance_pool) {
+ pr_err("Failed to create gen_pool for instance %d\n", instance_id);
+ return NULL;
+ }
+
+ /* Allocate memory in chunks and add to the pool */
+ while (remaining_size > 0) {
+ /* Try to allocate the remaining size, but be flexible */
+ chunk_size = remaining_size;
+ chunk_base = multikernel_alloc(chunk_size);
+
+ if (!chunk_base) {
+ /* If we can't get the full remaining size, try smaller chunks */
+ if (chunk_size > (1024 * 1024)) {
+ /* Try 1MB chunks */
+ chunk_size = 1024 * 1024;
+ chunk_base = multikernel_alloc(chunk_size);
+ }
+
+ if (!chunk_base && chunk_size > (256 * 1024)) {
+ /* Try 256KB chunks */
+ chunk_size = 256 * 1024;
+ chunk_base = multikernel_alloc(chunk_size);
+ }
+
+ if (!chunk_base && chunk_size > (1 << min_alloc_order)) {
+ /* Try minimum allocation size */
+ chunk_size = 1 << min_alloc_order;
+ chunk_base = multikernel_alloc(chunk_size);
+ }
+
+ if (!chunk_base) {
+ pr_err("Failed to allocate chunk %d for instance %d (remaining: %zu bytes)\n",
+ chunks_added + 1, instance_id, remaining_size);
+ goto cleanup;
+ }
+ }
+
+ /* Add the allocated chunk to the instance pool */
+ if (gen_pool_add(instance_pool, chunk_base, chunk_size, -1)) {
+ pr_err("Failed to add chunk %d to instance pool %d\n",
+ chunks_added + 1, instance_id);
+ multikernel_free(chunk_base, chunk_size);
+ goto cleanup;
+ }
+
+ chunks_added++;
+ remaining_size -= chunk_size;
+
+ pr_debug("Added chunk %d to instance pool %d: base=0x%llx, size=%zu bytes (remaining: %zu)\n",
+ chunks_added, instance_id, (unsigned long long)chunk_base,
+ chunk_size, remaining_size);
+ }
+
+ pr_info("Created instance pool %d: %d chunks, total size=%zu bytes\n",
+ instance_id, chunks_added, pool_size);
+
+ return instance_pool;
+
+cleanup:
+ /* Free all chunks that were successfully added */
+ multikernel_destroy_instance_pool(instance_pool);
+ return NULL;
+}
+
+/**
+ * multikernel_destroy_instance_pool() - Destroy an instance memory pool
+ * @pool_handle: Handle returned by multikernel_create_instance_pool()
+ *
+ * Frees all memory associated with the instance pool back to the main pool
+ */
+void multikernel_destroy_instance_pool(void *pool_handle)
+{
+ struct gen_pool *instance_pool = (struct gen_pool *)pool_handle;
+ struct gen_pool_chunk *chunk;
+
+ if (!instance_pool)
+ return;
+
+ /* Free all chunks back to main pool */
+ list_for_each_entry(chunk, &instance_pool->chunks, next_chunk) {
+ multikernel_free(chunk->start_addr, chunk->end_addr - chunk->start_addr + 1);
+ pr_debug("Freed instance pool chunk: 0x%lx-0x%lx\n",
+ chunk->start_addr, chunk->end_addr);
+ }
+
+ gen_pool_destroy(instance_pool);
+}
+
+/**
+ * multikernel_instance_alloc() - Allocate from an instance pool
+ * @pool_handle: Handle returned by multikernel_create_instance_pool()
+ * @size: Size to allocate
+ * @align: Alignment requirement (must be power of 2)
+ *
+ * Returns physical address of allocated memory, or 0 on failure
+ */
+phys_addr_t multikernel_instance_alloc(void *pool_handle, size_t size, size_t align)
+{
+ struct gen_pool *instance_pool = (struct gen_pool *)pool_handle;
+ unsigned long addr;
+
+ if (!instance_pool)
+ return 0;
+
+ if (align <= 1) {
+ addr = gen_pool_alloc(instance_pool, size);
+ } else {
+ /* Ensure alignment is at least the pool's minimum allocation order */
+ size_t a = max_t(size_t, align, BIT(instance_pool->min_alloc_order));
+ struct genpool_data_align data = { .align = a };
+ addr = gen_pool_alloc_algo(instance_pool, size, gen_pool_first_fit_align, &data);
+ }
+
+ return (phys_addr_t)addr;
+}
+
+/**
+ * multikernel_instance_free() - Free memory back to instance pool
+ * @pool_handle: Handle returned by multikernel_create_instance_pool()
+ * @addr: Physical address to free
+ * @size: Size to free
+ */
+void multikernel_instance_free(void *pool_handle, phys_addr_t addr, size_t size)
+{
+ struct gen_pool *instance_pool = (struct gen_pool *)pool_handle;
+
+ if (!instance_pool || !addr)
+ return;
+
+ gen_pool_free(instance_pool, (unsigned long)addr, size);
+ pr_debug("Instance pool freed %zu bytes at 0x%llx\n", size, (unsigned long long)addr);
+}
+
+/**
+ * multikernel_instance_pool_avail() - Get available space in instance pool
+ * @pool_handle: Handle returned by multikernel_create_instance_pool()
+ *
+ * Returns available bytes in the instance pool
+ */
+size_t multikernel_instance_pool_avail(void *pool_handle)
+{
+ struct gen_pool *instance_pool = (struct gen_pool *)pool_handle;
+
+ if (!instance_pool)
+ return 0;
+
+ return gen_pool_avail(instance_pool);
+}
+
+static int __init mkkernel_pool_setup(char *str)
+{
+ char *cur = str;
+ unsigned long long size, start;
+
+ if (!str)
+ return -EINVAL;
+
+ size = memparse(cur, &cur);
+ if (size == 0) {
+ pr_err("mkkernel_pool: invalid size\n");
+ return -EINVAL;
+ }
+
+ /* Expect '@' separator, or end of string for dynamic allocation */
+ if (*cur == '@') {
+ cur++;
+ /* Parse start address */
+ start = memparse(cur, &cur);
+ if (start == 0) {
+ pr_err("mkkernel_pool: invalid start address\n");
+ return -EINVAL;
+ }
+ } else if (*cur == '\0') {
+ /* No address specified, use dynamic allocation */
+ start = 0;
+ } else {
+ pr_err("mkkernel_pool: expected '@' or end of string after size\n");
+ return -EINVAL;
+ }
+
+ /* Reserve the memory using the proper memblock reservation approach */
+ phys_addr_t reserved_addr;
+ if (start != 0) {
+ /* Reserve at the user-specified address */
+ pr_info("mkkernel_pool: trying to reserve at specific address %llx\n", start);
+ if (memblock_reserve(start, size)) {
+ pr_err("mkkernel_pool: failed to reserve at specified address %llx\n", start);
+ return -ENOMEM;
+ }
+ reserved_addr = start;
+ pr_info("mkkernel_pool: successfully reserved at requested address %llx\n", start);
+ } else {
+ /* Dynamic allocation */
+ pr_info("mkkernel_pool: trying dynamic allocation\n");
+ reserved_addr = memblock_phys_alloc(size, PAGE_SIZE);
+ if (!reserved_addr) {
+ pr_err("mkkernel_pool: failed to allocate %llu bytes\n", size);
+ return -ENOMEM;
+ }
+ pr_info("mkkernel_pool: dynamic allocation succeeded at %pa\n", &reserved_addr);
+ }
+
+ multikernel_res.start = reserved_addr;
+ multikernel_res.end = reserved_addr + size - 1;
+
+ pr_info("Multikernel pool: %pa-%pa (%lluMB) allocated\n",
+ &multikernel_res.start, &multikernel_res.end, (unsigned long long)size >> 20);
+
+ return 0;
+}
+early_param("mkkernel_pool", mkkernel_pool_setup);
+
+static int __init multikernel_mem_init(void)
+{
+ if (multikernel_res.start) {
+ /* Create the generic pool */
+ multikernel_pool = gen_pool_create(PAGE_SHIFT, -1);
+ if (!multikernel_pool) {
+ pr_err("Failed to create multikernel memory pool\n");
+ return -ENOMEM;
+ }
+
+ /* Add the reserved memory to the pool */
+ if (gen_pool_add(multikernel_pool, multikernel_res.start,
+ multikernel_res.end - multikernel_res.start + 1, -1)) {
+ pr_err("Failed to add memory to multikernel pool\n");
+ gen_pool_destroy(multikernel_pool);
+ multikernel_pool = NULL;
+ return -ENOMEM;
+ }
+
+ if (insert_resource(&iomem_resource, &multikernel_res)) {
+ pr_warn("mkkernel_pool: failed to register in /proc/iomem\n");
+ } else {
+ pr_info("mkkernel_pool: successfully registered in /proc/iomem\n");
+ }
+
+ pr_info("Multikernel memory pool initialized: %pa-%pa\n",
+ &multikernel_res.start, &multikernel_res.end);
+ } else {
+ pr_info("No multikernel pool found - multikernel support disabled\n");
+ }
+
+ return 0;
+}
+core_initcall(multikernel_mem_init);
--
2.34.1
Powered by blists - more mailing lists