[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <90f8a4e1-aa71-0c10-1a91-495ba0cb329b@linux.ee>
Date: Tue, 21 May 2019 00:36:22 +0300
From: Meelis Roos <mroos@...ux.ee>
To: Rick Edgecombe <rick.p.edgecombe@...el.com>,
linux-kernel@...r.kernel.org, peterz@...radead.org,
sparclinux@...r.kernel.org, linux-mm@...ck.org,
netdev@...r.kernel.org
Cc: dave.hansen@...el.com, namit@...are.com,
Meelis Roos <mroos@...ux.ee>,
"David S. Miller" <davem@...emloft.net>,
Borislav Petkov <bp@...en8.de>,
Andy Lutomirski <luto@...nel.org>,
Ingo Molnar <mingo@...hat.com>
Subject: Re: [PATCH v2] vmalloc: Fix issues with flush flag
> Switch VM_FLUSH_RESET_PERMS to use a regular TLB flush intead of
> vm_unmap_aliases() and fix calculation of the direct map for the
> CONFIG_ARCH_HAS_SET_DIRECT_MAP case.
>
> Meelis Roos reported issues with the new VM_FLUSH_RESET_PERMS flag on a
> sparc machine. On investigation some issues were noticed:
>
> 1. The calculation of the direct map address range to flush was wrong.
> This could cause problems on x86 if a RO direct map alias ever got loaded
> into the TLB. This shouldn't normally happen, but it could cause the
> permissions to remain RO on the direct map alias, and then the page
> would return from the page allocator to some other component as RO and
> cause a crash.
>
> 2. Calling vm_unmap_alias() on vfree could potentially be a lot of work to
> do on a free operation. Simply flushing the TLB instead of the whole
> vm_unmap_alias() operation makes the frees faster and pushes the heavy
> work to happen on allocation where it would be more expected.
> In addition to the extra work, vm_unmap_alias() takes some locks including
> a long hold of vmap_purge_lock, which will make all other
> VM_FLUSH_RESET_PERMS vfrees wait while the purge operation happens.
>
> 3. page_address() can have locking on some configurations, so skip calling
> this when possible to further speed this up.
>
> Fixes: 868b104d7379 ("mm/vmalloc: Add flag for freeing of special permsissions")
> Reported-by: Meelis Roos<mroos@...ux.ee>
> Cc: Meelis Roos<mroos@...ux.ee>
> Cc: Peter Zijlstra<peterz@...radead.org>
> Cc: "David S. Miller"<davem@...emloft.net>
> Cc: Dave Hansen<dave.hansen@...el.com>
> Cc: Borislav Petkov<bp@...en8.de>
> Cc: Andy Lutomirski<luto@...nel.org>
> Cc: Ingo Molnar<mingo@...hat.com>
> Cc: Nadav Amit<namit@...are.com>
> Signed-off-by: Rick Edgecombe<rick.p.edgecombe@...el.com>
> ---
>
> Changes since v1:
> - Update commit message with more detail
> - Fix flush end range on !CONFIG_ARCH_HAS_SET_DIRECT_MAP case
It does not work on my V445 where the initial problem happened.
[ 46.582633] systemd[1]: Detected architecture sparc64.
Welcome to Debian GNU/Linux 10 (buster)!
[ 46.759048] systemd[1]: Set hostname to <v445>.
[ 46.831383] systemd[1]: Failed to bump fs.file-max, ignoring: Invalid argument
[ 67.989695] rcu: INFO: rcu_sched detected stalls on CPUs/tasks:
[ 68.074706] rcu: 0-...!: (0 ticks this GP) idle=5c6/1/0x4000000000000000 softirq=33/33 fqs=0
[ 68.198443] rcu: 2-...!: (0 ticks this GP) idle=e7e/1/0x4000000000000000 softirq=67/67 fqs=0
[ 68.322198] (detected by 1, t=5252 jiffies, g=-939, q=108)
[ 68.402204] CPU[ 0]: TSTATE[0000000080001603] TPC[000000000043f298] TNPC[000000000043f29c] TASK[systemd-debug-g:89]
[ 68.556001] TPC[smp_synchronize_tick_client+0x18/0x1a0] O7[0xfff000010000691c] I7[xcall_sync_tick+0x1c/0x2c] RPC[alloc_set_pte+0xf4/0x300]
[ 68.750973] CPU[ 2]: TSTATE[0000000080001600] TPC[000000000043f298] TNPC[000000000043f29c] TASK[systemd-cryptse:88]
[ 68.904741] TPC[smp_synchronize_tick_client+0x18/0x1a0] O7[filemap_map_pages+0x3cc/0x3e0] I7[xcall_sync_tick+0x1c/0x2c] RPC[handle_mm_fault+0xa0/0x180]
[ 69.115991] rcu: rcu_sched kthread starved for 5252 jiffies! g-939 f0x0 RCU_GP_WAIT_FQS(5) ->state=0x402 ->cpu=3
[ 69.262239] rcu: RCU grace-period kthread stack dump:
[ 69.334741] rcu_sched I 0 10 2 0x06000000
[ 69.413495] Call Trace:
[ 69.448501] [000000000093325c] schedule+0x1c/0xc0
[ 69.517253] [0000000000936c74] schedule_timeout+0x154/0x260
[ 69.598514] [00000000004b65a4] rcu_gp_kthread+0x4e4/0xac0
[ 69.677261] [000000000047ecfc] kthread+0xfc/0x120
[ 69.746018] [00000000004060a4] ret_from_fork+0x1c/0x2c
[ 69.821014] [0000000000000000] 0x0
and hangs here, software watchdog kicks in soon.
--
Meelis Roos
Powered by blists - more mailing lists