lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ded8b9bf-c9bb-8a41-541d-1bef354e4296@gmail.com>
Date: Wed, 15 Oct 2025 09:43:20 +0800
From: Hao Jia <jiahao.kernel@...il.com>
To: Aaron Lu <ziqianlu@...edance.com>
Cc: Valentin Schneider <vschneid@...hat.com>, Ben Segall
 <bsegall@...gle.com>, K Prateek Nayak <kprateek.nayak@....com>,
 Peter Zijlstra <peterz@...radead.org>,
 Chengming Zhou <chengming.zhou@...ux.dev>, Josh Don <joshdon@...gle.com>,
 Ingo Molnar <mingo@...hat.com>, Vincent Guittot
 <vincent.guittot@...aro.org>, Xi Wang <xii@...gle.com>,
 linux-kernel@...r.kernel.org, Juri Lelli <juri.lelli@...hat.com>,
 Dietmar Eggemann <dietmar.eggemann@....com>,
 Steven Rostedt <rostedt@...dmis.org>, Mel Gorman <mgorman@...e.de>,
 Chuyi Zhou <zhouchuyi@...edance.com>, Jan Kiszka <jan.kiszka@...mens.com>,
 Florian Bezdeka <florian.bezdeka@...mens.com>,
 Songtang Liu <liusongtang@...edance.com>, Chen Yu <yu.c.chen@...el.com>,
 Matteo Martelli <matteo.martelli@...ethink.co.uk>,
 Michal Koutný <mkoutny@...e.com>,
 Sebastian Andrzej Siewior <bigeasy@...utronix.de>
Subject: Re: [PATCH] sched/fair: Prevent cfs_rq from being unthrottled with
 zero runtime_remaining



On 2025/10/14 19:50, Aaron Lu wrote:
> On Tue, Oct 14, 2025 at 07:01:15PM +0800, Hao Jia wrote:
>>
>> Hello Aaron,
>>
>> Thank you for your reply.
>>
>> On 2025/10/14 17:11, Aaron Lu wrote:
>>> Hi Hao,
>>>
>>> On Tue, Oct 14, 2025 at 03:43:10PM +0800, Hao Jia wrote:
>>>>
>>>> Hello Aaron,
>>>>
>>>> On 2025/9/29 15:46, Aaron Lu wrote:
>>>>> When a cfs_rq is to be throttled, its limbo list should be empty and
>>>>> that's why there is a warn in tg_throttle_down() for non empty
>>>>> cfs_rq->throttled_limbo_list.
>>>>>
>>>>> When running a test with the following hierarchy:
>>>>>
>>>>>              root
>>>>>            /      \
>>>>>            A*     ...
>>>>>         /  |  \   ...
>>>>>            B
>>>>>           /  \
>>>>>          C*
>>>>>
>>>>> where both A and C have quota settings, that warn on non empty limbo list
>>>>> is triggered for a cfs_rq of C, let's call it cfs_rq_c(and ignore the cpu
>>>>> part of the cfs_rq for the sake of simpler representation).
>>>>>
>>>>
>>>> I encountered a similar warning a while ago and fixed it. I have a question
>>>> I'd like to ask. tg_unthrottle_up(cfs_rq_C) calls enqueue_task_fair(p) to
>>>> enqueue a task, which requires that the runtime_remaining of task p's entire
>>>> task_group hierarchy be greater than 0.
>>>>
>>>> In addition to the case you fixed above,
>>>> When bandwidth is running normally, Is it possible that there's a corner
>>>> case where cfs_A->runtime_remaining > 0, but cfs_B->runtime_remaining < 0
>>>> could trigger a similar warning?
>>>
>>> Do you mean B also has quota set and cfs_B's runtime_remaining < 0?
>>> In this case, B should be throttled and C is a descendent of B so should
>>> also be throttled, i.e. C can't be unthrottled when B is in throttled
>>> state. Do I understand you correctly?
>>>
>> Yes, both A and B have quota set.
>>
>> Is there a possible corner case?
>> Asynchronous unthrottling causes other running entities to completely
>> consume cfs_B->runtime_remaining (cfs_B->runtime_remaining < 0) but not
>> completely consume cfs_A->runtime_remaining (cfs_A->runtime_remaining > 0)
>> when we call unthrottle_cfs_rq(cfs_rq_A) .
> 
> Let me try to understand the situation here: in your described setup,
> all three task groups(A, B, C) have quota set?
> 
>>
>> When we unthrottle_cfs_rq(cfs_rq_A), cfs_A->runtime_remaining > 0, but if
>> cfs_B->runtime_remaining < 0 at this time,
> 
> Hmm... if cfs_B->runtime_remaining < 0, why it's not throttled?
> 
>> therefore, when enqueue_task_fair(p)->check_enqueue_throttle(cfs_rq_B)->throttle_cfs_rq(cfs_rq_B),
> 
> I assume p is a task of group B?
> So when A is unthrottled, since p is a throttled task of group B and B
> is still throttled, enqueue_task_fair(p) should not happen.
> 
>> an warnning may be triggered.
>>
>> My core question is:
>> When we call unthrottle_cfs_rq(cfs_rq_A), we only check
>> cfs_rq_A->runtime_remaining. However,
>> enqueue_task_fair(p)->enqueue_entity(C->B->A)->check_enqueue_throttle() does
> 
> According to this info, I assume p is a task of group C here. If
> unthrottle A would cause enqueuing p, that means: either group C and B
> do not have quota set or group C and B are in unthrottled state.
> 
>> require that the runtime_remaining of each task_group level of task p is
>> greater than 0.
> 
> If group C and B are in unthrottled state, their runtime_remaining
> should be > 0.
> 
>>
>> Can we guarantee this?
> 
> To guarantee this, a warn like below could be used. Can you try in your
> setup if you can hit it? Thanks.

Yes, I've already hit the cfs_rq->runtime_remaining < 0 condition in 
tg_unthrottle_up().

This morning, after applying your patch, I still get the same issue. 
However, As before, because cfs_rq->curr isn't NULL, 
check_enqueue_throttle() returns prematurely, preventing the triggering 
of throttle_cfs_rq().



Some information to share with you.

<...>-13123   [023] dN.1.    69.245648: tg_unthrottle_up: 
cfs_rq->runtime_remaining -5487 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c80ad300 cfs_rq->nr_queued 0
  sched-messaging-4299    [020] dNh2.   108.475966: tg_unthrottle_up: 
cfs_rq->runtime_remaining -1073 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57d2ef7400 cfs_rq->nr_queued 2
  sched-messaging-3514    [009] dNh2.   109.640727: tg_unthrottle_up: 
cfs_rq->runtime_remaining -11345 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad5a00 cfs_rq->nr_queued 4
            <...>-13872   [005] dNh2.   117.155177: tg_unthrottle_up: 
cfs_rq->runtime_remaining -15195 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad7e00 cfs_rq->nr_queued 5
            <...>-9861    [005] dNh2.   139.307579: tg_unthrottle_up: 
cfs_rq->runtime_remaining -41639 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f4f38400 cfs_rq->nr_queued 20
            <...>-9861    [005] dNh2.   139.307598: tg_unthrottle_up: 
cfs_rq->runtime_remaining -41639 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad7e00 cfs_rq->nr_queued 7
  sched-messaging-11165   [013] dNh2.   158.640827: tg_unthrottle_up: 
cfs_rq->runtime_remaining -32575 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e585693a400 cfs_rq->nr_queued 5
  sched-messaging-15337   [002] dNh2.   178.400559: tg_unthrottle_up: 
cfs_rq->runtime_remaining -29005 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ff4f6000 cfs_rq->nr_queued 7
  sched-messaging-8740    [010] dNh2.   178.842565: tg_unthrottle_up: 
cfs_rq->runtime_remaining -34448 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f4f4c800 cfs_rq->nr_queued 141
  sched-messaging-8740    [010] dNh2.   178.842583: tg_unthrottle_up: 
cfs_rq->runtime_remaining -34448 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad4e00 cfs_rq->nr_queued 4
            <...>-6101    [000] dNh1.   179.185475: tg_unthrottle_up: 
cfs_rq->runtime_remaining -14179 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ea202000 cfs_rq->nr_queued 116
  sched-messaging-6101    [000] dNh1.   179.185489: tg_unthrottle_up: 
cfs_rq->runtime_remaining -14179 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c7078400 cfs_rq->nr_queued 6
            <...>-7812    [002] dNh2.   184.228104: tg_unthrottle_up: 
cfs_rq->runtime_remaining -21154 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f597c200 cfs_rq->nr_queued 75
            <...>-7812    [002] dNh2.   184.228128: tg_unthrottle_up: 
cfs_rq->runtime_remaining -21154 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c7079600 cfs_rq->nr_queued 7
            <...>-4589    [006] dNh2.   185.872173: tg_unthrottle_up: 
cfs_rq->runtime_remaining -25309 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e584debc800 cfs_rq->nr_queued 173
            <...>-4589    [006] dNh2.   185.872192: tg_unthrottle_up: 
cfs_rq->runtime_remaining -25309 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad7800 cfs_rq->nr_queued 7
  sched-messaging-5155    [008] dNh2.   188.689689: tg_unthrottle_up: 
cfs_rq->runtime_remaining -14108 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ea20e800 cfs_rq->nr_queued 150
  sched-messaging-5155    [008] dNh2.   188.689704: tg_unthrottle_up: 
cfs_rq->runtime_remaining -14108 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad7600 cfs_rq->nr_queued 7
  sched-messaging-15265   [002] dNh2.   190.920442: tg_unthrottle_up: 
cfs_rq->runtime_remaining -25129 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ff4f6000 cfs_rq->nr_queued 6
            <...>-10191   [007] dNh1.   191.851170: tg_unthrottle_up: 
cfs_rq->runtime_remaining -27077 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad7200 cfs_rq->nr_queued 4
  sched-messaging-5963    [008] dNh2.   193.298329: tg_unthrottle_up: 
cfs_rq->runtime_remaining -41126 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57e9d31400 cfs_rq->nr_queued 156
  sched-messaging-5963    [008] dNh2.   193.298345: tg_unthrottle_up: 
cfs_rq->runtime_remaining -41126 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad7600 cfs_rq->nr_queued 6
            <...>-4877    [021] dNh2.   198.988637: tg_unthrottle_up: 
cfs_rq->runtime_remaining -19362 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ea2bf400 cfs_rq->nr_queued 86
            <...>-4877    [021] dNh2.   198.988657: tg_unthrottle_up: 
cfs_rq->runtime_remaining -19362 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c47e9e00 cfs_rq->nr_queued 6
            <...>-7344    [015] dNh2.   208.086556: tg_unthrottle_up: 
cfs_rq->runtime_remaining -31964 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e5856938a00 cfs_rq->nr_queued 4
            <...>-8384    [020] dNh2.   209.858709: tg_unthrottle_up: 
cfs_rq->runtime_remaining -28618 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f32ede00 cfs_rq->nr_queued 133
            <...>-8384    [020] dNh2.   209.858736: tg_unthrottle_up: 
cfs_rq->runtime_remaining -28618 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57d2ef7400 cfs_rq->nr_queued 7
  sched-messaging-7803    [000] dNh2.   210.380559: tg_unthrottle_up: 
cfs_rq->runtime_remaining -18845 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c7078400 cfs_rq->nr_queued 8
  sched-messaging-4388    [020] dNh2.   211.688504: tg_unthrottle_up: 
cfs_rq->runtime_remaining -22815 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57e7f0c600 cfs_rq->nr_queued 128
  sched-messaging-4388    [020] dNh2.   211.688533: tg_unthrottle_up: 
cfs_rq->runtime_remaining -22815 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57d2ef7400 cfs_rq->nr_queued 7
            <...>-15470   [014] dNh1.   223.440099: tg_unthrottle_up: 
cfs_rq->runtime_remaining -41950 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ff79ee00 cfs_rq->nr_queued 7
  sched-messaging-14343   [011] dNh2.   226.159614: tg_unthrottle_up: 
cfs_rq->runtime_remaining -24303 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f46d3a00 cfs_rq->nr_queued 2
            <...>-14543   [018] dNh1.   229.854321: tg_unthrottle_up: 
cfs_rq->runtime_remaining -23239 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e580acae000 cfs_rq->nr_queued 6
            <...>-9075    [010] dNh1.   231.397245: tg_unthrottle_up: 
cfs_rq->runtime_remaining -38148 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f4abb400 cfs_rq->nr_queued 108
            <...>-9075    [010] dNh1.   231.397267: tg_unthrottle_up: 
cfs_rq->runtime_remaining -38148 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad4e00 cfs_rq->nr_queued 5
            <...>-7271    [004] dNh2.   233.612978: tg_unthrottle_up: 
cfs_rq->runtime_remaining -24292 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57f418be00 cfs_rq->nr_queued 99
            <...>-7271    [004] dNh2.   233.612995: tg_unthrottle_up: 
cfs_rq->runtime_remaining -24292 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57c8ad4400 cfs_rq->nr_queued 6
            <...>-15191   [002] dNh1.   234.231866: tg_unthrottle_up: 
cfs_rq->runtime_remaining -28451 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ff4f6000 cfs_rq->nr_queued 6
            <...>-15181   [003] dNh2.   237.086491: tg_unthrottle_up: 
cfs_rq->runtime_remaining -20183 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ff4f7600 cfs_rq->nr_queued 4
  sched-messaging-6091    [013] dNh2.   243.412722: tg_unthrottle_up: 
cfs_rq->runtime_remaining -27394 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57ea209e00 cfs_rq->nr_queued 4
  sched-messaging-6091    [013] dNh2.   243.412742: tg_unthrottle_up: 
cfs_rq->runtime_remaining -27394 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e585693a400 cfs_rq->nr_queued 1
            <...>-6291    [010] dNh2.   245.166679: tg_unthrottle_up: 
cfs_rq->runtime_remaining -34474 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e57efee7600 cfs_rq->nr_queued 108
            <...>-4648    [017] dNh2.   246.414518: tg_unthrottle_up: 
cfs_rq->runtime_remaining -19404 cfs_rq->throttled 0 cfs_rq->curr 
ffff8e5856939600 cfs_rq->nr_queued 3



[Wed Oct 15 09:26:41 2025] ------------[ cut here ]------------
[Wed Oct 15 09:26:41 2025] WARNING: kernel/sched/fair.c:5897 at 
tg_unthrottle_up+0x2a6/0x340, CPU#23: sh/13123
[Wed Oct 15 09:26:41 2025] Modules linked in: xt_CHECKSUM xt_MASQUERADE 
xt_conntrack ipt_REJECT nf_reject_ipv4 xt_tcpudp nft_compat x_tables 
nft_chain_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 cfg80211 
nf_tables nfnetlink bridge stp llc cmac algif_hash algif_skcipher af_alg 
snd_hda_codec_intelhdmi bnep snd_hda_codec_alc269 
snd_hda_scodec_component snd_hda_codec_realtek_lib snd_hda_codec_generic 
binfmt_misc intel_uncore_frequency intel_uncore_frequency_common i915 
snd_hda_intel snd_sof_pci_intel_tgl snd_sof_pci_intel_cnl 
snd_sof_intel_hda_generic soundwire_intel soundwire_generic_allocation 
snd_sof_intel_hda_sdw_bpt snd_sof_intel_hda_common snd_soc_hdac_hda 
snd_sof_intel_hda_mlink snd_sof_intel_hda snd_hda_codec_hdmi 
snd_hda_ext_core soundwire_cadence snd_sof_pci snd_sof_xtensa_dsp 
snd_hda_codec snd_hda_core snd_sof snd_sof_utils x86_pkg_temp_thermal 
snd_intel_dspcfg intel_powerclamp snd_intel_sdw_acpi coretemp 
snd_soc_acpi_intel_match kvm_intel snd_soc_acpi_intel_sdca_quirks 
snd_soc_acpi snd_hwdep soundwire_bus
[Wed Oct 15 09:26:41 2025]  snd_soc_sdca snd_soc_core kvm snd_compress 
ac97_bus snd_pcm_dmaengine snd_pcm snd_seq_midi snd_seq_midi_event 
snd_rawmidi nls_iso8859_1 btusb snd_seq drm_buddy 
processor_thermal_device_pci btmtk ttm processor_thermal_device btrtl 
processor_thermal_wt_hint snd_seq_device btbcm irqbypass 
platform_temperature_control mei_hdcp ghash_clmulni_intel 
processor_thermal_soc_slider snd_timer btintel aesni_intel 
platform_profile processor_thermal_rfim drm_display_helper snd mei_me 
intel_rapl_msr rapl processor_thermal_rapl soundcore bluetooth 
intel_rapl_common intel_cstate cec think_lmi joydev mei input_leds crc8 
processor_thermal_wt_req firmware_attributes_class rc_core 8250_dw 
wmi_bmof processor_thermal_power_floor processor_thermal_mbox 
i2c_algo_bit int340x_thermal_zone int3400_thermal acpi_thermal_rel 
acpi_tad acpi_pad mac_hid sch_fq_codel msr parport_pc ppdev lp parport 
efi_pstore autofs4 btrfs blake2b_generic raid10 raid456 
async_raid6_recov async_memcpy async_pq async_xor async_tx xor raid6_pq 
raid1 raid0
[Wed Oct 15 09:26:41 2025]  spi_pxa2xx_platform hid_generic dw_dmac 
dw_dmac_core usbhid spi_pxa2xx_core nvme r8169 i2c_i801 hid nvme_core 
ahci i2c_mux realtek intel_lpss_pci i2c_smbus libahci intel_lpss idma64 
video wmi
[Wed Oct 15 09:26:41 2025] CPU: 23 UID: 0 PID: 13123 Comm: sh Kdump: 
loaded Not tainted 6.17.0+ #120 PREEMPT(voluntary)
[Wed Oct 15 09:26:41 2025] Hardware name: LENOVO 12JJA02KCD/3339, BIOS 
M4ZKT39A 11/22/2023
[Wed Oct 15 09:26:41 2025] RIP: 0010:tg_unthrottle_up+0x2a6/0x340
[Wed Oct 15 09:26:41 2025] Code: 48 8d 3d 00 00 00 00 e8 28 03 15 00 45 
8b 85 64 01 00 00 45 85 c0 0f 84 b9 fd ff ff 49 83 bd 68 01 00 00 00 0f 
8f ab fd ff ff <0f> 0b e9 a4 fd ff ff 48 8b 88 f0 0a 00 00 e9 7c ff ff 
ff 48 8b b1
[Wed Oct 15 09:26:41 2025] RSP: 0018:ffffd18529b8fae0 EFLAGS: 00010082
[Wed Oct 15 09:26:41 2025] RAX: 0000000000000006 RBX: ffff8e5f1fa72600 
RCX: 0000000000000000
[Wed Oct 15 09:26:41 2025] RDX: 0000000000000000 RSI: 0000000000000000 
RDI: 0000000000000000
[Wed Oct 15 09:26:41 2025] RBP: ffffd18529b8faf8 R08: 0000000000000001 
R09: 0000000000000000
[Wed Oct 15 09:26:41 2025] R10: 0000000000000000 R11: 0000000000000000 
R12: ffff8e57ca398c80
[Wed Oct 15 09:26:41 2025] R13: ffff8e57d1cc7a00 R14: 0000000000000000 
R15: ffff8e5f1fa72600
[Wed Oct 15 09:26:41 2025] FS:  00007fba99989740(0000) 
GS:ffff8e5f9826b000(0000) knlGS:0000000000000000
[Wed Oct 15 09:26:41 2025] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[Wed Oct 15 09:26:41 2025] CR2: 0000559e6f0ceda8 CR3: 000000014cee9002 
CR4: 0000000000f72ef0
[Wed Oct 15 09:26:41 2025] PKRU: 55555554
[Wed Oct 15 09:26:41 2025] Call Trace:
[Wed Oct 15 09:26:41 2025]  <TASK>
[Wed Oct 15 09:26:41 2025]  ? __pfx_tg_unthrottle_up+0x10/0x10
[Wed Oct 15 09:26:41 2025]  walk_tg_tree_from+0x66/0xd0
[Wed Oct 15 09:26:41 2025]  ? __pfx_tg_nop+0x10/0x10
[Wed Oct 15 09:26:41 2025]  unthrottle_cfs_rq+0xe4/0x280
[Wed Oct 15 09:26:41 2025]  tg_set_bandwidth+0x351/0x4d0
[Wed Oct 15 09:26:41 2025]  cpu_max_write+0xc3/0x130
[Wed Oct 15 09:26:41 2025]  cgroup_file_write+0x92/0x1a0
[Wed Oct 15 09:26:41 2025]  ? __check_object_size+0x27a/0x300
[Wed Oct 15 09:26:41 2025]  kernfs_fop_write_iter+0x15f/0x1f0
[Wed Oct 15 09:26:41 2025]  vfs_write+0x31b/0x430
[Wed Oct 15 09:26:41 2025]  ksys_write+0x6d/0xf0
[Wed Oct 15 09:26:41 2025]  __x64_sys_write+0x1d/0x30
[Wed Oct 15 09:26:41 2025]  x64_sys_call+0x1900/0x2760
[Wed Oct 15 09:26:41 2025]  do_syscall_64+0x83/0x8c0
[Wed Oct 15 09:26:41 2025]  ? __rseq_handle_notify_resume+0xaf/0x4e0
[Wed Oct 15 09:26:41 2025]  ? __do_sys_vfork+0x4f/0x70
[Wed Oct 15 09:26:41 2025]  ? __set_task_blocked+0x29/0x80
[Wed Oct 15 09:26:41 2025]  ? sigprocmask+0xb5/0xe0
[Wed Oct 15 09:26:41 2025]  ? __x64_sys_rt_sigprocmask+0x93/0xe0
[Wed Oct 15 09:26:41 2025]  ? x64_sys_call+0x1cc0/0x2760
[Wed Oct 15 09:26:41 2025]  ? do_syscall_64+0xbc/0x8c0
[Wed Oct 15 09:26:41 2025]  ? irqentry_exit+0x47/0x60
[Wed Oct 15 09:26:41 2025]  ? exc_page_fault+0x97/0x1b0
[Wed Oct 15 09:26:41 2025]  entry_SYSCALL_64_after_hwframe+0x76/0x7e
[Wed Oct 15 09:26:41 2025] RIP: 0033:0x7fba997148f7
[Wed Oct 15 09:26:41 2025] Code: 10 00 f7 d8 64 89 02 48 c7 c0 ff ff ff 
ff eb b7 0f 1f 00 f3 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 01 
00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 51 c3 48 83 ec 28 48 89 54 24 18 
48 89 74 24
[Wed Oct 15 09:26:41 2025] RSP: 002b:00007fffa912ef68 EFLAGS: 00000246 
ORIG_RAX: 0000000000000001
[Wed Oct 15 09:26:41 2025] RAX: ffffffffffffffda RBX: 000000000000000b 
RCX: 00007fba997148f7
[Wed Oct 15 09:26:41 2025] RDX: 000000000000000b RSI: 0000559e6f0ccda0 
RDI: 0000000000000001
[Wed Oct 15 09:26:41 2025] RBP: 0000559e6f0ccda0 R08: 0000559e6f0cccd7 
R09: 0000000000000000
[Wed Oct 15 09:26:41 2025] R10: 0000559e6f0cccd6 R11: 0000000000000246 
R12: 0000000000000001
[Wed Oct 15 09:26:41 2025] R13: 000000000000000b R14: 0000000000000000 
R15: 0000000000000000
[Wed Oct 15 09:26:41 2025]  </TASK>
[Wed Oct 15 09:26:41 2025] ---[ end trace 0000000000000000 ]---



> 
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index 3ef11783369d7..c347aa28c411a 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -5908,6 +5908,8 @@ static int tg_unthrottle_up(struct task_group *tg, void *data)
>   		cfs_rq->throttled_clock_self_time += delta;
>   	}
>   
> +	WARN_ON_ONCE(cfs_rq->runtime_enabled && cfs_rq->runtime_remaining <= 0);
> +
>   	/* Re-enqueue the tasks that have been throttled at this level. */
>   	list_for_each_entry_safe(p, tmp, &cfs_rq->throttled_limbo_list, throttle_node) {
>   		list_del_init(&p->throttle_node);

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ