lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <2ac2f797-6f72-ded6-bf0c-6b3c7b50a5f7@virtuozzo.com>
Date:   Tue, 20 Mar 2018 15:48:56 +0300
From:   Evgenii Shatokhin <eshatokhin@...tuozzo.com>
To:     Petr Mladek <pmladek@...e.com>
Cc:     Josh Poimboeuf <jpoimboe@...hat.com>,
        Jiri Kosina <jikos@...nel.org>,
        Miroslav Benes <mbenes@...e.cz>,
        Jason Baron <jbaron@...mai.com>,
        Joe Lawrence <joe.lawrence@...hat.com>,
        Jessica Yu <jeyu@...nel.org>, live-patching@...r.kernel.org,
        linux-kernel@...r.kernel.org
Subject: Re: [PATCH v10 05/10] livepatch: Support separate list for replaced
 patches.

On 20.03.2018 15:25, Petr Mladek wrote:
> On Mon 2018-03-19 16:43:24, Josh Poimboeuf wrote:
>> On Mon, Mar 19, 2018 at 04:02:07PM +0100, Petr Mladek wrote:
>>>> Can someone remind me why we're permanently disabling replaced patches?
>>>> I seem to remember being involved in that decision, but at least with
>>>> this latest version of the patches, it seems like it would be simpler to
>>>> just let 'replace' patches be rolled back to the previous state when
>>>> they're unpatched.  Then we don't need two lists of patches, the nops
>>>> can become more permanent, the replaced patches remain "enabled" but
>>>> inert, and the unpatching behavior is less surprising to the user, more
>>>> like a normal rollback.
>>>
>>> Yes, keeping the patches might make some things easier. But it might
>>> also bring some problems and it would make the feature less useful.
>>> The following arguments come to my mind:
>>>
>>> 1. The feature should help to keep the system in a consistent and
>>>     well defined state. It should not depend on what patches were
>>>     installed before.
>>
>> But the nops already accomplish that.  If they didn't, then this patch
>> set has a major problem.
> 
> The nops are enough to keep the functionality but they might harm
> the performance.
> 
> Livepatching is about preventing bugs without reboot. I could simply
> imagine that ftrace on a hot patch might cause performance
> problems on some workloads. And I would like to have a way out in
> this case.
> 
> Anyway, I am reworking the patchset so that it implements your
> approach first. The possibility to remove NOPs and replaced
> livepatches is done via a followup patch. This might help
> to discuss if the changes are worth it or not.
> 
> 
>>> 2. The feature should allow to unpatch some functions while keeping
>>>     the others patched.
>>>
>>>     The ftrace handler might cause some unwanted slowdown or other
>>>     problems. The performance might get restored only when we remove
>>>     the NOPs when they are not longer necessary.
>>
>> I'd say simplicity and maintainability of the code is more important
>> than an (imagined) performance issue.  The NOPs should be pretty fast
>> anyway.
>>
>> Not to mention that my proposal would make the behavior less surprising
>> and more user friendly (reverting a 'replace' patch restores it to its
>> previous state).
> 
> If the "disable" way works as expected, see below.
> 
> Also it is less surprising only if people understand the stack of
> patches. If they are familiar only with replace patches then it is
> normal for them that the patches get replaced. It is then like
> a package version update.
> 
> 
>>> 3. The handling of callbacks is already problematic. We run only
>>>     the ones from the last patch to make things easier.
>>>
>>>     We would need to come with something more complicated if we
>>>     want to support rollback to "random" patches on the stack.
>>>     And support for random patches is fundamental at least
>>>     from my point of view.
>>
>> Can you elaborate on what you mean by random patches and why it would
>> require something more complicated from the callbacks?
> 
> Let's say that we will use atomic replace for cumulative
> patches. Then every new patch knows what earlier patches did.
> It just did not know which of them was already installed. Therefore
> it needs to detect what callbacks were already called. The callbacks
> usually create or change something. So there should be something to check.
> Therefore the way forward should be rather straightforward.
> 
> The way back is more problematic. The callbacks in the new cumulative
> patch would need to store information about the previous state and
> be able to restore it when the patch gets disabled. It might more
> or less double the callbacks code and testing scenarios.
> 
> 
>>>> Along those lines, I'd also propose that we constrain our existing patch
>>>> stacking even further.  Right now we allow a new patch to be registered
>>>> on top of a disabled patch, though we don't allow the new patch to be
>>>> enabled until the previous patch gets enabled.  I'd propose we no longer
>>>> allow that condition.  We should instead enforce that all existing
>>>> patches are *enabled* before allowing a new patch to be registered on
>>>> top.  That way the patch stacking is even more sane, and there are less
>>>> "unusual" conditions to worry about.  We have enough of those already.
>>>> Each additional bit of flexibility has a maintenance cost, and this one
>>>> isn't worth it IMO.
>>>
>>> Again, this might make some things easier but it might also bring
>>> problems.
>>>
>>> For example, we would need to solve the situation when the last
>>> patch is disabled and cannot be removed because the transition
>>> was forced. This might be more common after removing the immediate
>>> feature.
>>
>> I would stop worrying about forced patches so much :-)
> 
> I have already seen blocked transition several times. It is true that
> it was with kGraft. But we just do not have enough real life experience
> with the upstream livepatch code.
> 
> 
>> Forced patches already come with a disclaimer, and we can't bend over
>> backwards for them.  In such a rare case, the admin can just re-enable
>> the forced patch before loading the 'replace' patch.
> 
> 
>>> Also it might be less user friendly.
>>
>> I don't know, does anybody really care about this case (patching on top
>> of a disabled patch)?  It just adds to the crazy matrix of possible
>> scenarios we have to keep in our heads, which means more bugs, for very
>> little (hypothetical) gain.
> 
> It depends if the we remove the replaced patches or not. If they are
> removed then replacing disabled patches is rather trivial from both
> coding and understanding side.
> 
> I am going to add this as a separate patch as well. Let's discuss
> it with the code.
> 
> 
>>> White the atomic replace could make things easier for both developers
>>> and users.
>>
>> I agree that atomic replace is a useful feature and I'm not arguing
>> against it, so maybe I missed your point?
> 
> Your suggestion allows easier code but it reduces the advantages of
> the atomic replace feature. We would achieve almost the same results
> with a normal livepatch where the functions behave like in
> the original code.
> 
> Also removing replaced patches can be seen as a clean up after
> each patch. It might be more code but the target system might be
> easier to debug. Also we do not need to mind about various
> disable scenarios.

+1

In our case (ReadyKernel), we mostly use cumulative patches with 
occasional "normal" patches as hot fixes and such.

After the new cumulative patch has been loaded and enabled, all the 
previously loaded patches are unloaded. Yes, it is seen as a cleanup.

So, most of the time, a single binary patch is loaded and active. It 
helps debugging if issues show up and it also makes it easier to explain 
to the customers which fixes are in effect on their machines.

Just my 2 cents.

> 
> Best Regards,
> Petr
> .
> 

Regards,
Evgenii

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ