[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4DD9C8B9.5070004@intel.com>
Date: Mon, 23 May 2011 10:38:49 +0800
From: Huang Ying <ying.huang@...el.com>
To: Ingo Molnar <mingo@...e.hu>
CC: huang ying <huang.ying.caritas@...il.com>,
Len Brown <lenb@...nel.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
Andi Kleen <andi@...stfloor.org>,
"Luck, Tony" <tony.luck@...el.com>,
"linux-acpi@...r.kernel.org" <linux-acpi@...r.kernel.org>,
Andi Kleen <ak@...ux.intel.com>,
"Wu, Fengguang" <fengguang.wu@...el.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Borislav Petkov <bp@...en8.de>
Subject: Re: [PATCH 5/9] HWPoison: add memory_failure_queue()
On 05/22/2011 09:25 PM, Ingo Molnar wrote:
>>> The generalization that *would* make sense is not at the irq_work level
>>> really, instead we could generalize a 'struct event' for kernel internal
>>> producers and consumers of events that have no explicit PMU connection.
>>>
>>> This new 'struct event' would be slimmer and would only contain the fields
>>> and features that generic event consumers and producers need. Tracing
>>> events could be updated to use these kinds of slimmer events.
>>>
>>> It would still plug nicely into existing event ABIs, would work with event
>>> filters, etc. so the tooling side would remain focused and unified.
>>>
>>> Something like that. It is rather clear by now that splitting out irq_work
>>> was a mistake. But mistakes can be fixed and some really nice code could
>>> come out of it! Would you be interested in looking into this?
>>
>> Yes. This can transfer hardware error data from kernel to user space. Then,
>> how to do hardware error recovering in this big picture? IMHO, we will need
>> to call something like memory_failure_queue() in IRQ context for memory
>> error.
>
> That's where 'active filters' come into the picture - see my other mail (that
> was in the context of unidentified NMI errors/events) where i outlined how they
> would work in this case and elsewhere. Via active filters we could share most
> of the code, gain access to the events and still have kernel driven policy
> action.
Is that something as follow?
- NMI handler run for the hardware error, where hardware error
information is collected and put into perf ring buffer as 'event'.
- Some 'active filters' are run for each 'event' in NMI context.
- Some operations can not be done in NMI handler, so they are delayed to
an IRQ handler (can be done with something like irq_work).
- Some other 'active filters' are run for each 'event' in IRQ context.
(For memory error, we can call memory_failure_queue() here).
Where some 'active filters' are kernel built-in, some 'active filters'
can be customized via kernel command line or by user space.
If my understanding as above is correct, I think this is a general and
complex solution. It is a little hard for user to understand which
'active filters' are in effect. He may need some runtime assistant to
understand the code (maybe /sys/events/active_filters, which list all
filters in effect now), because that is hard only by reading the source
code. Anyway, this is a design style choice.
There are still some issues, I don't know how to solve in above framework.
- If there are two processes request the same type of hardware error
events. One hardware error event will be copied to two ring buffers
(each for one process), but the 'active filters' should be run only
once for each hardware error event.
- How to deal with ring-buffer overflow? For example, there is full of
corrected memory error in ring-buffer, and now a recoverable memory
error occurs but it can not be put into perf ring buffer because of
ring-buffer overflow, how to deal with the recoverable memory error?
Best Regards,
Huang Ying
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists