[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <D4E120A4-D877-48CC-AE40-D55DBB6265D0@oracle.com>
Date: Mon, 2 Dec 2024 16:05:18 +0000
From: Chuck Lever III <chuck.lever@...cle.com>
To: Li Lingfeng <lilingfeng3@...wei.com>
CC: Dai Ngo <dai.ngo@...cle.com>, Jeff Layton <jlayton@...nel.org>,
Neil Brown
<neilb@...e.de>, Olga Kornievskaia <okorniev@...hat.com>,
Tom Talpey
<tom@...pey.com>,
Trond Myklebust <trond.myklebust@...merspace.com>,
Linux
NFS Mailing List <linux-nfs@...r.kernel.org>,
Linux Kernel Mailing List
<linux-kernel@...r.kernel.org>,
Yu Kuai <yukuai1@...weicloud.com>, Hou Tao
<houtao1@...wei.com>,
"zhangyi (F)" <yi.zhang@...wei.com>, yangerkun
<yangerkun@...wei.com>,
"chengzhihao1@...wei.com" <chengzhihao1@...wei.com>,
Li Lingfeng <lilingfeng@...weicloud.com>
Subject: Re: [bug report] deploying both NFS client and server on the same
machine triggle hungtask
> On Nov 28, 2024, at 2:22 AM, Li Lingfeng <lilingfeng3@...wei.com> wrote:
>
> Besides nfsd_file_shrinker, the nfsd_client_shrinker added by commit
> 7746b32f467b ("NFSD: add shrinker to reap courtesy clients on low memory
> condition") in 2022 and the nfsd_reply_cache_shrinker added by commit
> 3ba75830ce17 ("nfsd4: drc containerization") in 2019 may also trigger such
> an issue.
> Was this scenario not considered when designing the shrinkers for NFSD, or
> was it deemed unreasonable and not worth considering?
I'm speculating, but it is possible that the issue was
introduced by another patch in an area related to the
rwsem. Seems like there is a testing gap in this area.
Can you file a bugzilla report on bugzilla.kernel.org <http://bugzilla.kernel.org/>
under Filesystems/NFSD ?
> 在 2024/11/25 19:17, Li Lingfeng 写道:
>> Hi, we have found a hungtask issue recently.
>>
>> Commit 7746b32f467b ("NFSD: add shrinker to reap courtesy clients on low
>> memory condition") adds a shrinker to NFSD, which causes NFSD to try to
>> obtain shrinker_rwsem when starting and stopping services.
>>
>> Deploying both NFS client and server on the same machine may lead to the
>> following issue, since they will share the global shrinker_rwsem.
>>
>> nfsd nfs
>> drop_cache // hold shrinker_rwsem
>> write back, wait for rpc_task to exit
>> // stop nfsd threads
>> svc_set_num_threads
>> // clean up xprts
>> svc_xprt_destroy_all
>> rpc_check_timeout
>> rpc_check_connected
>> // wait for the connection to be disconnected
>> unregister_shrinker
>> // wait for shrinker_rwsem
>>
>> Normally, the client's rpc_task will exit after the server's nfsd thread
>> has processed the request.
>> When all the server's nfsd threads exit, the client’s rpc_task is expected
>> to detect the network connection being disconnected and exit.
>> However, although the server has executed svc_xprt_destroy_all before
>> waiting for shrinker_rwsem, the network connection is not actually
>> disconnected. Instead, the operation to close the socket is simply added
>> to the task_works queue.
>>
>> svc_xprt_destroy_all
>> ...
>> svc_sock_free
>> sockfd_put
>> fput_many
>> init_task_work // ____fput
>> task_work_add // add to task->task_works
>>
>> The actual disconnection of the network connection will only occur after
>> the current process finishes.
>> do_exit
>> exit_task_work
>> task_work_run
>> ...
>> ____fput // close sock
>>
>> Although it is not a common practice to deploy NFS client and server on
>> the same machine, I think this issue still needs to be addressed,
>> otherwise it will cause all processes trying to acquire the shrinker_rwsem
>> to hang.
>>
>> I don't have any ideas yet on how to solve this problem, does anyone have
>> any suggestions?
>>
>> Thanks.
>>
--
Chuck Lever
Powered by blists - more mailing lists