lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87lean7f2c.fsf@yhuang6-desk2.ccr.corp.intel.com>
Date:   Fri, 24 Nov 2023 16:04:11 +0800
From:   "Huang, Ying" <ying.huang@...el.com>
To:     "zhangpeng (AS)" <zhangpeng362@...wei.com>
Cc:     Yin Fengwei <fengwei.yin@...el.com>, <linux-mm@...ck.org>,
        <linux-kernel@...r.kernel.org>, <akpm@...ux-foundation.org>,
        <willy@...radead.org>, <aneesh.kumar@...ux.ibm.com>,
        <shy828301@...il.com>, <hughd@...gle.com>, <david@...hat.com>,
        <wangkefeng.wang@...wei.com>, <sunnanyong@...wei.com>
Subject: Re: [RFC PATCH] mm: filemap: avoid unnecessary major faults in
 filemap_fault()

"zhangpeng (AS)" <zhangpeng362@...wei.com> writes:

> On 2023/11/24 12:26, Huang, Ying wrote:
>
>> "Huang, Ying" <ying.huang@...el.com> writes:
>>
>>> "zhangpeng (AS)" <zhangpeng362@...wei.com> writes:
>>>
>>>> On 2023/11/23 13:26, Yin Fengwei wrote:
>>>>
>>>>> On 11/23/23 12:12, zhangpeng (AS) wrote:
>>>>>> On 2023/11/23 9:09, Yin Fengwei wrote:
>>>>>>
>>>>>>> Hi Peng,
>>>>>>>
>>>>>>> On 11/22/23 22:00, Peng Zhang wrote:
>>>>>>>> From: ZhangPeng <zhangpeng362@...wei.com>
>>>>>>>>
>>>>>>>> The major fault occurred when using mlockall(MCL_CURRENT | MCL_FUTURE)
>>>>>>>> in application, which leading to an unexpected performance issue[1].
>>>>>>>>
>>>>>>>> This caused by temporarily cleared pte during a read/modify/write update
>>>>>>>> of the pte, eg, do_numa_page()/change_pte_range().
>>>>>>>>
>>>>>>>> For the data segment of the user-mode program, the global variable area
>>>>>>>> is a private mapping. After the pagecache is loaded, the private anonymous
>>>>>>>> page is generated after the COW is triggered. Mlockall can lock COW pages
>>>>>>>> (anonymous pages), but the original file pages cannot be locked and may
>>>>>>>> be reclaimed. If the global variable (private anon page) is accessed when
>>>>>>>> vmf->pte is zeroed in numa fault, a file page fault will be triggered.
>>>>>>>>
>>>>>>>> At this time, the original private file page may have been reclaimed.
>>>>>>>> If the page cache is not available at this time, a major fault will be
>>>>>>>> triggered and the file will be read, causing additional overhead.
>>>>>>>>
>>>>>>>> Fix this by rechecking the pte by holding ptl in filemap_fault() before
>>>>>>>> triggering a major fault.
>>>>>>>>
>>>>>>>> [1] https://lore.kernel.org/linux-mm/9e62fd9a-bee0-52bf-50a7-498fa17434ee@huawei.com/
>>>>>>>>
>>>>>>>> Signed-off-by: ZhangPeng <zhangpeng362@...wei.com>
>>>>>>>> Signed-off-by: Kefeng Wang <wangkefeng.wang@...wei.com>
>>>>>>>> ---
>>>>>>>>     mm/filemap.c | 14 ++++++++++++++
>>>>>>>>     1 file changed, 14 insertions(+)
>>>>>>>>
>>>>>>>> diff --git a/mm/filemap.c b/mm/filemap.c
>>>>>>>> index 71f00539ac00..bb5e6a2790dc 100644
>>>>>>>> --- a/mm/filemap.c
>>>>>>>> +++ b/mm/filemap.c
>>>>>>>> @@ -3226,6 +3226,20 @@ vm_fault_t filemap_fault(struct vm_fault *vmf)
>>>>>>>>                 mapping_locked = true;
>>>>>>>>             }
>>>>>>>>         } else {
>>>>>>>> +        pte_t *ptep = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd,
>>>>>>>> +                          vmf->address, &vmf->ptl);
>>>>>>>> +        if (ptep) {
>>>>>>>> +            /*
>>>>>>>> +             * Recheck pte with ptl locked as the pte can be cleared
>>>>>>>> +             * temporarily during a read/modify/write update.
>>>>>>>> +             */
>>>>>>>> +            if (unlikely(!pte_none(ptep_get(ptep))))
>>>>>>>> +                ret = VM_FAULT_NOPAGE;
>>>>>>>> +            pte_unmap_unlock(ptep, vmf->ptl);
>>>>>>>> +            if (unlikely(ret))
>>>>>>>> +                return ret;
>>>>>>>> +        }
>>>>>>> I am curious. Did you try not to take PTL here and just check whether PTE is not NONE?
>>>>>> Thank you for your reply.
>>>>>>
>>>>>> If we don't take PTL, the current use case won't trigger this issue either.
>>>>> Is this verified by testing or just in theory?
>>>> If we add a delay between ptep_modify_prot_start() and ptep_modify_prot_commit(),
>>>> this issue will also trigger. Without delay, we haven't reproduced this problem
>>>> so far.
>>>>
>>>>>> In most cases, if we don't take PTL, this issue won't be triggered. However,
>>>>>> there is still a possibility of triggering this issue. The corner case is that
>>>>>> task 2 triggers a page fault when task 1 is between ptep_modify_prot_start()
>>>>>> and ptep_modify_prot_commit() in do_numa_page(). Furthermore,task 2 passes the
>>>>>> check whether the PTE is not NONE before task 1 updates PTE in
>>>>>> ptep_modify_prot_commit() without taking PTL.
>>>>> There is very limited operations between ptep_modify_prot_start() and
>>>>> ptep_modify_prot_commit(). While the code path from page fault to this check is
>>>>> long. My understanding is it's very likely the PTE is not NONE when do PTE check
>>>>> here without hold PTL (This is my theory. :)).
>>>> Yes, there is a high probability that this issue won't occur without taking PTL.
>>>>
>>>>> In the other side, acquiring/releasing PTL may bring performance impaction. It may
>>>>> not be big deal because the IO operations in this code path. But it's better to
>>>>> collect some performance data IMHO.
>>>> We tested the performance of file private mapping page fault (page_fault2.c of
>>>> will-it-scale [1]) and file shared mapping page fault (page_fault3.c of will-it-scale).
>>>> The difference in performance (in operations per second) before and after patch
>>>> applied is about 0.7% on a x86 physical machine.
>>> Whether is it improvement or reduction?
>> And I think that you need to test ramdisk cases too to verify whether
>> this will cause performance regression and how much.
>
> Yes, I will.
> In addition, are there any ramdisk test cases recommended? 😁

I think that you can start with the will-it-scale test case you used
before.  And you can try some workload with large number of major fault,
like file read with mmap.

--
Best Regards,
Huang, Ying

>> --
>> Best Regards,
>> Huang, Ying
>>
>>> --
>>> Best Regards,
>>> Huang, Ying
>>>
>>>> [1] https://github.com/antonblanchard/will-it-scale/tree/master
>>>>
>>>>> Regards
>>>>> Yin, Fengwei
>>>>>
>>>>>>> Regards
>>>>>>> Yin, Fengwei
>>>>>>>
>>>>>>>> +
>>>>>>>>             /* No page in the page cache at all */
>>>>>>>>             count_vm_event(PGMAJFAULT);
>>>>>>>>             count_memcg_event_mm(vmf->vma->vm_mm, PGMAJFAULT);

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ