[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <5fe26d02-b3d9-bcd1-196e-73a5932be83f@leemhuis.info>
Date: Wed, 5 Oct 2022 16:41:15 +0200
From: Thorsten Leemhuis <regressions@...mhuis.info>
To: "regressions@...ts.linux.dev" <regressions@...ts.linux.dev>
Cc: intel-gfx@...ts.freedesktop.org, iommu@...ts.linux.dev,
linux-kernel@...r.kernel.org
Subject: Re: [core-for-CI][PATCH] iommu: Remove iova cpu hotplugging flushing
#forregzbot
[Note: this mail is primarily send for documentation purposes and/or for
regzbot, my Linux kernel regression tracking bot. That's why I removed
most or all folks from the list of recipients, but left any that looked
like a mailing lists. These mails usually contain '#forregzbot' in the
subject, to make them easy to spot and filter out.]
[TLDR: I'm adding this regression report to the list of tracked
regressions; all text from me you find below is based on a few templates
paragraphs you might have encountered already already in similar form.]
Hi, this is your Linux kernel regression tracker.
On 22.09.22 12:10, Janusz Krzysztofik wrote:
> From: Chris Wilson <chris@...is-wilson.co.uk>
>
> Manual revert of commit f598a497bc7d ("iova: Add CPU hotplug handler to
> flush rcaches"). It is trying to instantiate a cpuhp notifier from inside
> a cpuhp callback. That code replaced intel_iommu implementation of
> flushing per-IOVA domain CPU rcaches which used a single instance of cpuhp
> held for the module lifetime.
Thanks for the report. To be sure below issue doesn't fall through the
cracks unnoticed, I'm adding it to regzbot, my Linux kernel regression
tracking bot:
#regzbot ^introduced f598a497bc7d
#regzbot title iommu/coretemp: possible circular locking dependency detected
#regzbot ignore-activity
This isn't a regression? This issue or a fix for it are already
discussed somewhere else? It was fixed already? You want to clarify when
the regression started to happen? Or point out I got the title or
something else totally wrong? Then just reply -- ideally with also
telling regzbot about it, as explained here:
https://linux-regtracking.leemhuis.info/tracked-regression/
Reminder for developers: When fixing the issue, add 'Link:' tags
pointing to the report (the mail this one replies to), as explained for
in the Linux kernel's documentation; above webpage explains why this is
important for tracked regressions.
Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat)
P.S.: As the Linux kernel's regression tracker I deal with a lot of
reports and sometimes miss something important when writing mails like
this. If that's the case here, don't hesitate to tell me in a public
reply, it's in everyone's interest to set the public record straight.
> <4>[ 6.928112] ======================================================
> <4>[ 6.928621] WARNING: possible circular locking dependency detected
> <4>[ 6.929225] 6.0.0-rc6-CI_DRM_12164-ga1f63e144e54+ #1 Not tainted
> <4>[ 6.929818] ------------------------------------------------------
> <4>[ 6.930415] cpuhp/0/15 is trying to acquire lock:
> <4>[ 6.931011] ffff888100e02a78 (&(&priv->bus_notifier)->rwsem){++++}-{3:3}, at: blocking_notifier_call_chain+0x20/0x50
> <4>[ 6.931533]
> but task is already holding lock:
> <4>[ 6.931534] ffffffff826490c0 (cpuhp_state-up){+.+.}-{0:0}, at: cpuhp_thread_fun+0x48/0x1f0
> <4>[ 6.933069]
> which lock already depends on the new lock.
>
> <4>[ 6.933070]
> the existing dependency chain (in reverse order) is:
> <4>[ 6.933071]
> -> #2 (cpuhp_state-up){+.+.}-{0:0}:
> <4>[ 6.933076] lock_acquire+0xd3/0x310
> <4>[ 6.933079] cpuhp_thread_fun+0xa6/0x1f0
> <4>[ 6.933082] smpboot_thread_fn+0x1b5/0x260
> <4>[ 6.933084] kthread+0xed/0x120
> <4>[ 6.933086] ret_from_fork+0x1f/0x30
> <4>[ 6.933089]
> -> #1 (cpu_hotplug_lock){++++}-{0:0}:
> <4>[ 6.933092] lock_acquire+0xd3/0x310
> <4>[ 6.933094] __cpuhp_state_add_instance+0x43/0x1c0
> <4>[ 6.933096] iova_domain_init_rcaches+0x199/0x1c0
> <4>[ 6.933099] iommu_setup_dma_ops+0x104/0x3d0
> <4>[ 6.933101] iommu_probe_device+0xa4/0x180
> <4>[ 6.933103] iommu_bus_notifier+0x2d/0x40
> <4>[ 6.933105] notifier_call_chain+0x31/0x90
> <4>[ 6.933108] blocking_notifier_call_chain+0x3a/0x50
> <4>[ 6.933110] device_add+0x3c1/0x900
> <4>[ 6.933112] pci_device_add+0x255/0x580
> <4>[ 6.933115] pci_scan_single_device+0xa6/0xd0
> <4>[ 6.933117] p2sb_bar+0x7f/0x220
> <4>[ 6.933120] i801_add_tco_spt.isra.18+0x2b/0xca [i2c_i801]
> <4>[ 6.933124] i801_add_tco+0xb1/0xfe [i2c_i801]
> <4>[ 6.933126] i801_probe.cold.25+0xa9/0x3a7 [i2c_i801]
> <4>[ 6.933129] pci_device_probe+0x95/0x110
> <4>[ 6.933132] really_probe+0xd6/0x350
> <4>[ 6.933134] __driver_probe_device+0x73/0x170
> <4>[ 6.933137] driver_probe_device+0x1a/0x90
> <4>[ 6.933140] __driver_attach+0xbc/0x190
> <4>[ 6.933141] bus_for_each_dev+0x72/0xc0
> <4>[ 6.933143] bus_add_driver+0x1bb/0x210
> <4>[ 6.933146] driver_register+0x66/0xc0
> <4>[ 6.933147] wmi_bmof_probe+0x3b/0xac [wmi_bmof]
> <4>[ 6.933150] do_one_initcall+0x53/0x2f0
> <4>[ 6.933152] do_init_module+0x45/0x1c0
> <4>[ 6.933154] load_module+0x1cd5/0x1ec0
> <4>[ 6.933156] __do_sys_finit_module+0xaf/0x120
> <4>[ 6.933158] do_syscall_64+0x37/0x90
> <4>[ 6.933160] entry_SYSCALL_64_after_hwframe+0x63/0xcd
> <4>[ 6.953757]
> -> #0 (&(&priv->bus_notifier)->rwsem){++++}-{3:3}:
> <4>[ 6.953779] validate_chain+0xb3f/0x2000
> <4>[ 6.953785] __lock_acquire+0x5a4/0xb70
> <4>[ 6.953786] lock_acquire+0xd3/0x310
> <4>[ 6.953787] down_read+0x39/0x140
> <4>[ 6.953790] blocking_notifier_call_chain+0x20/0x50
> <4>[ 6.953794] device_add+0x3c1/0x900
> <4>[ 6.953797] platform_device_add+0x108/0x240
> <4>[ 6.953799] coretemp_cpu_online+0xe1/0x15e [coretemp]
> <4>[ 6.953805] cpuhp_invoke_callback+0x181/0x8a0
> <4>[ 6.958244] cpuhp_thread_fun+0x188/0x1f0
> <4>[ 6.958267] smpboot_thread_fn+0x1b5/0x260
> <4>[ 6.958270] kthread+0xed/0x120
> <4>[ 6.958272] ret_from_fork+0x1f/0x30
> <4>[ 6.958274]
> other info that might help us debug this:
>
> <4>[ 6.958275] Chain exists of:
> &(&priv->bus_notifier)->rwsem --> cpu_hotplug_lock --> cpuhp_state-up
>
> <4>[ 6.961037] Possible unsafe locking scenario:
>
> <4>[ 6.961038] CPU0 CPU1
> <4>[ 6.961038] ---- ----
> <4>[ 6.961039] lock(cpuhp_state-up);
> <4>[ 6.961040] lock(cpu_hotplug_lock);
> <4>[ 6.961041] lock(cpuhp_state-up);
> <4>[ 6.961042] lock(&(&priv->bus_notifier)->rwsem);
> <4>[ 6.961044]
> *** DEADLOCK ***
>
> <4>[ 6.961044] 2 locks held by cpuhp/0/15:
> <4>[ 6.961046] #0: ffffffff82648f10 (cpu_hotplug_lock){++++}-{0:0}, at: cpuhp_thread_fun+0x48/0x1f0
> <4>[ 6.961053] #1: ffffffff826490c0 (cpuhp_state-up){+.+.}-{0:0}, at: cpuhp_thread_fun+0x48/0x1f0
> <4>[ 6.961058]
> stack backtrace:
> <4>[ 6.961059] CPU: 0 PID: 15 Comm: cpuhp/0 Not tainted 6.0.0-rc6-CI_DRM_12164-ga1f63e144e54+ #1
> <4>[ 6.961062] Hardware name: Intel Corporation NUC8i7HVK/NUC8i7HVB, BIOS HNKBLi70.86A.0047.2018.0718.1706 07/18/2018
> <4>[ 6.961063] Call Trace:
> <4>[ 6.961064] <TASK>
> <4>[ 6.961065] dump_stack_lvl+0x56/0x7f
> <4>[ 6.961069] check_noncircular+0x132/0x150
> <4>[ 6.961078] validate_chain+0xb3f/0x2000
> <4>[ 6.961083] __lock_acquire+0x5a4/0xb70
> <4>[ 6.961087] lock_acquire+0xd3/0x310
> <4>[ 6.961088] ? blocking_notifier_call_chain+0x20/0x50
> <4>[ 6.961093] down_read+0x39/0x140
> <4>[ 6.961097] ? blocking_notifier_call_chain+0x20/0x50
> <4>[ 6.961099] blocking_notifier_call_chain+0x20/0x50
> <4>[ 6.961102] device_add+0x3c1/0x900
> <4>[ 6.961106] ? dev_set_name+0x4e/0x70
> <4>[ 6.961109] platform_device_add+0x108/0x240
> <4>[ 6.961112] coretemp_cpu_online+0xe1/0x15e [coretemp]
> <4>[ 6.961117] ? create_core_data+0x550/0x550 [coretemp]
> <4>[ 6.961120] cpuhp_invoke_callback+0x181/0x8a0
> <4>[ 6.961124] cpuhp_thread_fun+0x188/0x1f0
> <4>[ 6.961129] ? smpboot_thread_fn+0x1e/0x260
> <4>[ 6.961131] smpboot_thread_fn+0x1b5/0x260
> <4>[ 6.961134] ? sort_range+0x20/0x20
> <4>[ 6.961135] kthread+0xed/0x120
> <4>[ 6.961137] ? kthread_complete_and_exit+0x20/0x20
> <4>[ 6.961139] ret_from_fork+0x1f/0x30
> <4>[ 6.961145] </TASK>
>
> Closes: https://gitlab.freedesktop.org/drm/intel/issues/6641
> Signed-off-by: Chris Wilson <chris@...is-wilson.co.uk>
> Signed-off-by: Janusz Krzysztofik <janusz.krzysztofik@...ux.intel.com>
> ---
> drivers/iommu/iova.c | 28 ----------------------------
> include/linux/cpuhotplug.h | 1 -
> include/linux/iova.h | 1 -
> 3 files changed, 30 deletions(-)
>
> diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c
> index 47d1983dfa2a4..f0136d0231f06 100644
> --- a/drivers/iommu/iova.c
> +++ b/drivers/iommu/iova.c
> @@ -31,16 +31,6 @@ unsigned long iova_rcache_range(void)
> return PAGE_SIZE << (IOVA_RANGE_CACHE_MAX_SIZE - 1);
> }
>
> -static int iova_cpuhp_dead(unsigned int cpu, struct hlist_node *node)
> -{
> - struct iova_domain *iovad;
> -
> - iovad = hlist_entry_safe(node, struct iova_domain, cpuhp_dead);
> -
> - free_cpu_cached_iovas(cpu, iovad);
> - return 0;
> -}
> -
> static void free_global_cached_iovas(struct iova_domain *iovad);
>
> static struct iova *to_iova(struct rb_node *node)
> @@ -255,21 +245,10 @@ int iova_cache_get(void)
> {
> mutex_lock(&iova_cache_mutex);
> if (!iova_cache_users) {
> - int ret;
> -
> - ret = cpuhp_setup_state_multi(CPUHP_IOMMU_IOVA_DEAD, "iommu/iova:dead", NULL,
> - iova_cpuhp_dead);
> - if (ret) {
> - mutex_unlock(&iova_cache_mutex);
> - pr_err("Couldn't register cpuhp handler\n");
> - return ret;
> - }
> -
> iova_cache = kmem_cache_create(
> "iommu_iova", sizeof(struct iova), 0,
> SLAB_HWCACHE_ALIGN, NULL);
> if (!iova_cache) {
> - cpuhp_remove_multi_state(CPUHP_IOMMU_IOVA_DEAD);
> mutex_unlock(&iova_cache_mutex);
> pr_err("Couldn't create iova cache\n");
> return -ENOMEM;
> @@ -292,7 +271,6 @@ void iova_cache_put(void)
> }
> iova_cache_users--;
> if (!iova_cache_users) {
> - cpuhp_remove_multi_state(CPUHP_IOMMU_IOVA_DEAD);
> kmem_cache_destroy(iova_cache);
> }
> mutex_unlock(&iova_cache_mutex);
> @@ -495,8 +473,6 @@ EXPORT_SYMBOL_GPL(free_iova_fast);
>
> static void iova_domain_free_rcaches(struct iova_domain *iovad)
> {
> - cpuhp_state_remove_instance_nocalls(CPUHP_IOMMU_IOVA_DEAD,
> - &iovad->cpuhp_dead);
> free_iova_rcaches(iovad);
> }
>
> @@ -755,10 +731,6 @@ int iova_domain_init_rcaches(struct iova_domain *iovad)
> }
> }
>
> - ret = cpuhp_state_add_instance_nocalls(CPUHP_IOMMU_IOVA_DEAD,
> - &iovad->cpuhp_dead);
> - if (ret)
> - goto out_err;
> return 0;
>
> out_err:
> diff --git a/include/linux/cpuhotplug.h b/include/linux/cpuhotplug.h
> index f61447913db97..8f541a6b63e41 100644
> --- a/include/linux/cpuhotplug.h
> +++ b/include/linux/cpuhotplug.h
> @@ -95,7 +95,6 @@ enum cpuhp_state {
> CPUHP_PAGE_ALLOC,
> CPUHP_NET_DEV_DEAD,
> CPUHP_PCI_XGENE_DEAD,
> - CPUHP_IOMMU_IOVA_DEAD,
> CPUHP_LUSTRE_CFS_DEAD,
> CPUHP_AP_ARM_CACHE_B15_RAC_DEAD,
> CPUHP_PADATA_DEAD,
> diff --git a/include/linux/iova.h b/include/linux/iova.h
> index c6ba6d95d79c2..fd77cd5bfa333 100644
> --- a/include/linux/iova.h
> +++ b/include/linux/iova.h
> @@ -37,7 +37,6 @@ struct iova_domain {
> struct iova anchor; /* rbtree lookup anchor */
>
> struct iova_rcache *rcaches;
> - struct hlist_node cpuhp_dead;
> };
>
> static inline unsigned long iova_size(struct iova *iova)
Powered by blists - more mailing lists