[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20241222110509.GS11133@noisy.programming.kicks-ass.net>
Date: Sun, 22 Dec 2024 12:05:09 +0100
From: Peter Zijlstra <peterz@...radead.org>
To: Rik van Riel <riel@...riel.com>
Cc: x86@...nel.org, linux-kernel@...r.kernel.org, kernel-team@...a.com,
dave.hansen@...ux.intel.com, luto@...nel.org, tglx@...utronix.de,
mingo@...hat.com, bp@...en8.de, hpa@...or.com,
akpm@...ux-foundation.org
Subject: Re: [PATCH 03/10] x86,mm: add INVLPGB support code
On Sat, Dec 21, 2024 at 11:06:35PM -0500, Rik van Riel wrote:
> +static inline void __invlpgb(unsigned long asid, unsigned long pcid, unsigned long addr,
> + int extra_count, bool pmd_stride, unsigned long flags)
> +{
> + u64 rax = addr | flags;
> + u32 ecx = (pmd_stride << 31) | extra_count;
> + u32 edx = (pcid << 16) | asid;
> +
> + /*
> + * The memory clobber is because the whole point is to invalidate
> + * stale TLB entries and, especially if we're flushing global
> + * mappings, we don't want the compiler to reorder any subsequent
> + * memory accesses before the TLB flush.
> + */
> + asm volatile("invlpgb" : : "a" (rax), "c" (ecx), "d" (edx));
What memory clobber? Is "memory" gone missing?
> +}
Powered by blists - more mailing lists