[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20251127141117.87420-4-luxu.kernel@bytedance.com>
Date: Thu, 27 Nov 2025 22:11:11 +0800
From: Xu Lu <luxu.kernel@...edance.com>
To: pjw@...nel.org,
palmer@...belt.com,
aou@...s.berkeley.edu,
alex@...ti.fr,
kees@...nel.org,
mingo@...hat.com,
peterz@...radead.org,
juri.lelli@...hat.com,
vincent.guittot@...aro.org,
akpm@...ux-foundation.org,
david@...hat.com,
apatel@...tanamicro.com,
guoren@...nel.org
Cc: linux-riscv@...ts.infradead.org,
linux-kernel@...r.kernel.org,
linux-mm@...ck.org,
Xu Lu <luxu.kernel@...edance.com>
Subject: [RFC PATCH v2 3/9] riscv: mm: Grab mm_count to avoid mm getting released
We maintain an array of mm_structs whose ASIDs are active on the current
CPU. To avoid these mm_structs getting released, we grab their mm_count
before loaded them into the array. And drop their mm_count via tasklet
when they are evicted out of the array.
Signed-off-by: Xu Lu <luxu.kernel@...edance.com>
---
arch/riscv/include/asm/mmu.h | 4 +++
arch/riscv/mm/tlbflush.c | 47 ++++++++++++++++++++++++++++++++++++
2 files changed, 51 insertions(+)
diff --git a/arch/riscv/include/asm/mmu.h b/arch/riscv/include/asm/mmu.h
index cf8e6eac77d52..913fa535b3d19 100644
--- a/arch/riscv/include/asm/mmu.h
+++ b/arch/riscv/include/asm/mmu.h
@@ -30,6 +30,10 @@ typedef struct {
#ifdef CONFIG_RISCV_ISA_SUPM
u8 pmlen;
#endif
+#ifdef CONFIG_RISCV_LAZY_TLB_FLUSH
+ atomic_t lazy_tlb_cnt;
+ void *next;
+#endif
} mm_context_t;
/* Lock the pointer masking mode because this mm is multithreaded */
diff --git a/arch/riscv/mm/tlbflush.c b/arch/riscv/mm/tlbflush.c
index 0b1c21c7aafb8..4b2ce06cbe6bd 100644
--- a/arch/riscv/mm/tlbflush.c
+++ b/arch/riscv/mm/tlbflush.c
@@ -104,12 +104,57 @@ struct flush_tlb_range_data {
};
#ifdef CONFIG_RISCV_LAZY_TLB_FLUSH
+
DEFINE_PER_CPU_SHARED_ALIGNED(struct tlb_info, tlbinfo) = {
.rwlock = __RW_LOCK_UNLOCKED(tlbinfo.rwlock),
.active_mm = NULL,
.next_gen = 1,
.contexts = { { NULL, 0, }, },
};
+
+static DEFINE_PER_CPU(mm_context_t *, mmdrop_victims);
+
+static void mmdrop_lazy_mms(struct tasklet_struct *tasklet)
+{
+ mm_context_t *victim = xchg_relaxed(this_cpu_ptr(&mmdrop_victims), NULL);
+ struct mm_struct *mm = NULL;
+
+ while (victim) {
+ mm = container_of(victim, struct mm_struct, context);
+ while (atomic_dec_return_relaxed(&victim->lazy_tlb_cnt) != 0)
+ mmdrop_lazy_tlb(mm);
+ victim = victim->next;
+ }
+}
+
+static DEFINE_PER_CPU(struct tasklet_struct, mmdrop_tasklets) = {
+ .count = ATOMIC_INIT(0),
+ .callback = mmdrop_lazy_mms,
+ .use_callback = true,
+};
+
+static inline void mmgrab_lazy_mm(struct mm_struct *mm)
+{
+ mmgrab_lazy_tlb(mm);
+ atomic_inc(&mm->context.lazy_tlb_cnt);
+}
+
+static inline void mmdrop_lazy_mm(struct mm_struct *mm)
+{
+ mm_context_t **head, *list, *context = &mm->context;
+
+ if (atomic_inc_return_relaxed(&context->lazy_tlb_cnt) == 1) {
+ head = this_cpu_ptr(&mmdrop_victims);
+
+ do {
+ list = *head;
+ context->next = list;
+ } while (cmpxchg_relaxed(head, list, context) != list);
+
+ tasklet_schedule(this_cpu_ptr(&mmdrop_tasklets));
+ }
+}
+
#endif /* CONFIG_RISCV_LAZY_TLB_FLUSH */
static void __ipi_flush_tlb_range_asid(void *info)
@@ -292,6 +337,7 @@ void local_load_tlb_mm(struct mm_struct *mm)
info->active_mm = mm;
if (contexts[pos].mm != mm) {
+ mmgrab_lazy_mm(mm);
victim = contexts[pos].mm;
contexts[pos].mm = mm;
}
@@ -302,6 +348,7 @@ void local_load_tlb_mm(struct mm_struct *mm)
if (victim) {
cpumask_clear_cpu(raw_smp_processor_id(), mm_cpumask(victim));
local_flush_tlb_all_asid(get_mm_asid(victim));
+ mmdrop_lazy_mm(victim);
}
}
--
2.20.1
Powered by blists - more mailing lists