[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <583A7D67.50003@mellanox.com>
Date: Sun, 27 Nov 2016 08:29:59 +0200
From: Roi Dayan <roid@...lanox.com>
To: Daniel Borkmann <daniel@...earbox.net>,
Cong Wang <xiyou.wangcong@...il.com>
CC: <roid@...lanox.com>,
Linux Kernel Network Developers <netdev@...r.kernel.org>,
Jiri Pirko <jiri@...lanox.com>,
John Fastabend <john.fastabend@...il.com>
Subject: Re: [Patch net-next] net_sched: move the empty tp check from
->destroy() to ->delete()
On 27/11/2016 06:47, Roi Dayan wrote:
>
>
> On 27/11/2016 02:33, Daniel Borkmann wrote:
>> On 11/26/2016 12:09 PM, Daniel Borkmann wrote:
>>> On 11/26/2016 07:46 AM, Cong Wang wrote:
>>>> On Thu, Nov 24, 2016 at 7:20 AM, Daniel Borkmann
>>>> <daniel@...earbox.net> wrote:
>> [...]
>>>>> Ok, strange, qdisc_destroy() calls into ops->destroy(), where ingress
>>>>> drops its entire chain via tcf_destroy_chain(), so that will be NULL
>>>>> eventually. The tps are freed by call_rcu() as well as qdisc itself
>>>>> later on via qdisc_rcu_free(), where it frees per-cpu bstats as well.
>>>>> Outstanding readers should either bail out due to if (!cl) or can
>>>>> still
>>>>> process the chain until read section ends, but during that time,
>>>>> cl->q
>>>>> resp. bstats should be good. Do you happen to know what's at address
>>>>> ffff880a68b04028? I was wondering wrt call_rcu() vs call_rcu_bh(),
>>>>> but
>>>>> at least on ingress (netif_receive_skb_internal()) we hold
>>>>> rcu_read_lock()
>>>>> here. The KASAN report is reliably happening at this location, right?
>>>>
>>>> I am confused as well, I don't see how it could be related to my
>>>> patch yet.
>>>> I will take a deep look in the weekend.
>
>
>
> Hi Cong,
>
> When reported the new trace I didn't mean it's related to your patch,
> I just wanted to point it out it exposed something. I should have been
> clear about it.
>
>
>>>
>>> Ok, I'm currently on the run. Got too late yesterday night, but I'll
>>> write what I found in the evening today, not related to ingress though.
>>
>> Just pushed out my analysis to netdev under "[PATCH net] net, sched:
>> respect
>> rcu grace period on cls destruction". My conclusion is that both
>> issues are
>> actually separate, and that one is small enough where we could route
>> it via
>> net actually. Perhaps this at the same time shrinks your "[PATCH
>> net-next]
>> net_sched: move the empty tp check from ->destroy() to ->delete()" to a
>> reasonable size that it's suitable to net as well. Your
>> ->delete()/->destroy()
>> one is definitely needed, too. The tp->root one is independant of
>> ->delete()/
>> ->destroy() as they are different races and tp->root could also
>> happen when
>> you just destroy the whole tp directly. I think that seems like a
>> good path
>> forward to me.
>>
>> Thanks,
>> Daniel
>
>
>
> Hi Daniel,
>
> As for the tainted kernel. I was in old (week or two) net-next tree
> and only cherry-picked from latest net-next related patches to
> Mellanox HCA, cls_api, cls_flower, devlink. so those are the tainted
> modules.
> I have the issue reproducing in that tree so wanted it to check it
> with Cong's patch instead of latest net-next.
> I'll try running reproducing the issue with your new patch and later
> try latest net-next as well.
>
> Thanks,
> Roi
>
Hi,
I tested "[PATCH net] net, sched: respect rcu grace period on cls
destruction" and could not reproduce my original issue.
I rebased "[Patch net-next] net_sched: move the empty tp check from
->destroy() to ->delete()" over to test it in the same tree and got into
a new trace in fl_delete.
[35659.012123] BUG: KASAN: wild-memory-access on address 1ffffffff803ca31
[35659.020042] Write of size 1 by task ovs-vswitchd/20135
[35659.025878] CPU: 19 PID: 20135 Comm: ovs-vswitchd Tainted:
G O 4.9.0-rc3+ #18
[35659.035948] Hardware name: HP ProLiant DL380p Gen8, BIOS P70 07/01/2015
[35659.043730] Call Trace:
[35659.046619] [<ffffffff95b6dc42>] dump_stack+0x63/0x81
[35659.052456] [<ffffffff955fbbf8>] kasan_report_error+0x408/0x4e0
[35659.059402] [<ffffffff955fc2e8>] kasan_report+0x58/0x60
[35659.065428] [<ffffffff952d5e8d>] ? call_rcu_sched+0x1d/0x20
[35659.072119] [<ffffffffc01e0701>] ? fl_destroy_filter+0x21/0x30
[cls_flower]
[35659.080217] [<ffffffffc01e1ccf>] ? fl_delete+0x1df/0x2e0 [cls_flower]
[35659.087580] [<ffffffff955fa4ca>] __asan_store1+0x4a/0x50
[35659.093697] [<ffffffffc01e1ccf>] fl_delete+0x1df/0x2e0 [cls_flower]
[35659.100870] [<ffffffff9653ecba>] tc_ctl_tfilter+0x10da/0x1b90
0x1d02 is in fl_delete (net/sched/cls_flower.c:805).
800 struct cls_fl_filter *f = (struct cls_fl_filter *) arg;
801
802 rhashtable_remove_fast(&head->ht, &f->ht_node,
803 head->ht_params);
804 __fl_delete(tp, f);
805 *last = list_empty(&head->filters);
806 return 0;
807 }
Thanks,
Roi
Powered by blists - more mailing lists