lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <55AF4FD7.2010009@transip.nl>
Date:	Wed, 22 Jul 2015 10:09:59 +0200
From:	Frank Schreuder <fschreuder@...nsip.nl>
To:	Florian Westphal <fw@...len.de>
CC:	Nikolay Aleksandrov <nikolay@...ulusnetworks.com>,
	Johan Schuijt <johan@...nsip.nl>,
	Eric Dumazet <eric.dumazet@...il.com>,
	"nikolay@...hat.com" <nikolay@...hat.com>,
	"davem@...emloft.net" <davem@...emloft.net>,
	"chutzpah@...too.org" <chutzpah@...too.org>,
	Robin Geuze <robing@...nsip.nl>,
	netdev <netdev@...r.kernel.org>
Subject: Re: reproducable panic eviction work queue



Op 7/21/2015 om 8:34 PM schreef Florian Westphal:
> Frank Schreuder <fschreuder@...nsip.nl> wrote:
>
> [ inet frag evictor crash ]
>
> We believe we found the bug.  This patch should fix it.
>
> We cannot share list for buckets and evictor, the flag member is
> subject to race conditions so flags & INET_FRAG_EVICTED test is not
> reliable.
>
> It would be great if you could confirm that this fixes the problem
> for you, we'll then make formal patch submission.
>
> Please apply this on kernel without previous test patches, wheter you
> use affected -stable or net-next kernel shouldn't matter since those are
> similar enough.
>
> Many thanks!
>
> diff --git a/include/net/inet_frag.h b/include/net/inet_frag.h
> --- a/include/net/inet_frag.h
> +++ b/include/net/inet_frag.h
> @@ -45,6 +45,7 @@ enum {
>    * @flags: fragment queue flags
>    * @max_size: maximum received fragment size
>    * @net: namespace that this frag belongs to
> + * @list_evictor: list of queues to forcefully evict (e.g. due to low memory)
>    */
>   struct inet_frag_queue {
>   	spinlock_t		lock;
> @@ -59,6 +60,7 @@ struct inet_frag_queue {
>   	__u8			flags;
>   	u16			max_size;
>   	struct netns_frags	*net;
> +	struct hlist_node	list_evictor;
>   };
>   
>   #define INETFRAGS_HASHSZ	1024
> diff --git a/net/ipv4/inet_fragment.c b/net/ipv4/inet_fragment.c
> index 5e346a0..1722348 100644
> --- a/net/ipv4/inet_fragment.c
> +++ b/net/ipv4/inet_fragment.c
> @@ -151,14 +151,13 @@ evict_again:
>   		}
>   
>   		fq->flags |= INET_FRAG_EVICTED;
> -		hlist_del(&fq->list);
> -		hlist_add_head(&fq->list, &expired);
> +		hlist_add_head(&fq->list_evictor, &expired);
>   		++evicted;
>   	}
>   
>   	spin_unlock(&hb->chain_lock);
>   
> -	hlist_for_each_entry_safe(fq, n, &expired, list)
> +	hlist_for_each_entry_safe(fq, n, &expired, list_evictor)
>   		f->frag_expire((unsigned long) fq);
>   
>   	return evicted;
> @@ -284,8 +283,7 @@ static inline void fq_unlink(struct inet_frag_queue *fq, struct inet_frags *f)
>   	struct inet_frag_bucket *hb;
>   
>   	hb = get_frag_bucket_locked(fq, f);
> -	if (!(fq->flags & INET_FRAG_EVICTED))
> -		hlist_del(&fq->list);
> +	hlist_del(&fq->list);
>   	spin_unlock(&hb->chain_lock);
>   }
>   
Hi Florian,

Thanks for the patch!

After implementing the patch in our setup we are no longer able to 
reproduct the kernel panic.
Unfortunately the server load increases after 5/10 minutes and the logs 
are getting spammed with stacktraces.
I included a snippet below.

Do you have any insights on why this happens, and how we can resolve this?

Thanks,
Frank


Jul 22 09:44:17 dommy0 kernel: [  360.121516] Modules linked in: 
parport_pc ppdev lp parport bnep rfcomm bluetooth rfkill uinput nfsd 
auth_rpcgss oid_registry nfs_acl nfs lockd grace fscache sunrpc loop 
coretemp kvm ttm drm_kms_helper iTCO_wdt drm psmouse ipmi_si 
iTCO_vendor_support tpm_tis tpm ipmi_msghandler i2c_algo_bit i2c_core 
i7core_edac dcdbas serio_raw pcspkr wmi lpc_ich edac_core mfd_core evdev 
button acpi_power_meter processor thermal_sys ext4 crc16 mbcache jbd2 
sd_mod sg sr_mod cdrom hid_generic usbhid ata_generic hid crc32c_intel 
ata_piix mptsas scsi_transport_sas mptscsih libata mptbase ehci_pci 
scsi_mod uhci_hcd ehci_hcd usbcore usb_common ixgbe dca ptp bnx2 
pps_core mdio
Jul 22 09:44:17 dommy0 kernel: [  360.121560] CPU: 3 PID: 42 Comm: 
kworker/3:1 Tainted: G        W    L 3.18.18-transip-1.6 #1
Jul 22 09:44:17 dommy0 kernel: [  360.121562] Hardware name: Dell Inc. 
PowerEdge R410/01V648, BIOS 1.12.0 07/30/2013
Jul 22 09:44:17 dommy0 kernel: [  360.121567] Workqueue: events 
inet_frag_worker
Jul 22 09:44:17 dommy0 kernel: [  360.121568] task: ffff880224574490 ti: 
ffff8802240a0000 task.ti: ffff8802240a0000
Jul 22 09:44:17 dommy0 kernel: [  360.121570] RIP: 
0010:[<ffffffff810c0872>]  [<ffffffff810c0872>] del_timer_sync+0x42/0x60
Jul 22 09:44:17 dommy0 kernel: [  360.121575] RSP: 
0018:ffff8802240a3d48  EFLAGS: 00000246
Jul 22 09:44:17 dommy0 kernel: [  360.121576] RAX: 0000000000000200 RBX: 
0000000000000000 RCX: 0000000000000000
Jul 22 09:44:17 dommy0 kernel: [  360.121578] RDX: ffff88022215ce40 RSI: 
0000000000300000 RDI: ffff88022215cdf0
Jul 22 09:44:17 dommy0 kernel: [  360.121579] RBP: 0000000000000003 R08: 
ffff880222343c00 R09: 0000000000000101
Jul 22 09:44:17 dommy0 kernel: [  360.121581] R10: 0000000000000000 R11: 
0000000000000027 R12: ffff880222343c00
Jul 22 09:44:17 dommy0 kernel: [  360.121582] R13: 0000000000000101 R14: 
0000000000000000 R15: 0000000000000027
Jul 22 09:44:17 dommy0 kernel: [  360.121584] FS: 0000000000000000(0000) 
GS:ffff88022f260000(0000) knlGS:0000000000000000
Jul 22 09:44:17 dommy0 kernel: [  360.121585] CS:  0010 DS: 0000 ES: 
0000 CR0: 000000008005003b
Jul 22 09:44:17 dommy0 kernel: [  360.121587] CR2: 00007fb1e9884095 CR3: 
000000021c084000 CR4: 00000000000007e0
Jul 22 09:44:17 dommy0 kernel: [  360.121588] Stack:
Jul 22 09:44:17 dommy0 kernel: [  360.121589]  ffff88022215cdf0 
ffffffff8149289e ffffffff81a8aa30 ffffffff81a8aa38
Jul 22 09:44:17 dommy0 kernel: [  360.121592]  0000000000000286 
ffff88022215ce88 ffffffff8149287f 0000000000000394
Jul 22 09:44:17 dommy0 kernel: [  360.121594]  ffffffff81a87100 
0000000000000001 000000000000007c 0000000000000000
Jul 22 09:44:17 dommy0 kernel: [  360.121596] Call Trace:
Jul 22 09:44:17 dommy0 kernel: [  360.121600] [<ffffffff8149289e>] ? 
inet_evict_bucket+0x11e/0x140
Jul 22 09:44:17 dommy0 kernel: [  360.121602] [<ffffffff8149287f>] ? 
inet_evict_bucket+0xff/0x140
Jul 22 09:44:17 dommy0 kernel: [  360.121605] [<ffffffff814929b0>] ? 
inet_frag_worker+0x60/0x210
Jul 22 09:44:17 dommy0 kernel: [  360.121609] [<ffffffff8107e3a2>] ? 
process_one_work+0x142/0x3b0
Jul 22 09:44:17 dommy0 kernel: [  360.121612] [<ffffffff815078ed>] ? 
schedule+0x1d/0x70
Jul 22 09:44:17 dommy0 kernel: [  360.121614] [<ffffffff8107eb94>] ? 
worker_thread+0x114/0x440
Jul 22 09:44:17 dommy0 kernel: [  360.121617] [<ffffffff815073ad>] ? 
__schedule+0x2cd/0x7b0
Jul 22 09:44:17 dommy0 kernel: [  360.121619] [<ffffffff8107ea80>] ? 
create_worker+0x1a0/0x1a0
Jul 22 09:44:17 dommy0 kernel: [  360.121622] [<ffffffff81083dfc>] ? 
kthread+0xbc/0xe0
Jul 22 09:44:17 dommy0 kernel: [  360.121624] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:44:17 dommy0 kernel: [  360.121627] [<ffffffff8150b218>] ? 
ret_from_fork+0x58/0x90
Jul 22 09:44:17 dommy0 kernel: [  360.121629] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:44:17 dommy0 kernel: [  360.121631] Code: 75 29 be 3c 04 00 00 
48 c7 c7 0c 73 71 81 e8 26 72 fa ff 48 89 df e8 6e ff ff ff 85 c0 79 18 
66 2e 0f 1f 84 00 00 00 00 00 f3 90 <48> 89 df e8 56 ff ff ff 85 c0 78 
f2 5b 90 c3 66 66 66 66 66 66

Jul 22 09:44:27 dommy0 kernel: [  370.097476] Task dump for CPU 3:
Jul 22 09:44:27 dommy0 kernel: [  370.097478] kworker/3:1     R running 
task        0    42      2 0x00000008
Jul 22 09:44:27 dommy0 kernel: [  370.097482] Workqueue: events 
inet_frag_worker
Jul 22 09:44:27 dommy0 kernel: [  370.097483]  0000000000000004 
ffffffff81849240 ffffffff810b9464 00000000000003dc
Jul 22 09:44:27 dommy0 kernel: [  370.097485]  ffff88022f26d4c0 
ffffffff81849180 ffffffff81849240 ffffffff818b4e40
Jul 22 09:44:27 dommy0 kernel: [  370.097488]  ffffffff810bc797 
0000000000000000 ffffffff810c6dc9 0000000000000092
Jul 22 09:44:27 dommy0 kernel: [  370.097490] Call Trace:
Jul 22 09:44:27 dommy0 kernel: [  370.097491]  <IRQ> 
[<ffffffff810b9464>] ? rcu_dump_cpu_stacks+0x84/0xc0
Jul 22 09:44:27 dommy0 kernel: [  370.097499] [<ffffffff810bc797>] ? 
rcu_check_callbacks+0x407/0x650
Jul 22 09:44:27 dommy0 kernel: [  370.097501] [<ffffffff810c6dc9>] ? 
timekeeping_update.constprop.8+0x89/0x1b0
Jul 22 09:44:27 dommy0 kernel: [  370.097504] [<ffffffff810c7ec5>] ? 
update_wall_time+0x225/0x5c0
Jul 22 09:44:27 dommy0 kernel: [  370.097507] [<ffffffff810cfcb0>] ? 
tick_sched_do_timer+0x30/0x30
Jul 22 09:44:27 dommy0 kernel: [  370.097510] [<ffffffff810c14df>] ? 
update_process_times+0x3f/0x80
Jul 22 09:44:27 dommy0 kernel: [  370.097513] [<ffffffff810cfb27>] ? 
tick_sched_handle.isra.12+0x27/0x70
Jul 22 09:44:27 dommy0 kernel: [  370.097515] [<ffffffff810cfcf5>] ? 
tick_sched_timer+0x45/0x80
Jul 22 09:44:27 dommy0 kernel: [  370.097518] [<ffffffff810c1d76>] ? 
__run_hrtimer+0x66/0x1b0
Jul 22 09:44:27 dommy0 kernel: [  370.097522] [<ffffffff8101c5c5>] ? 
read_tsc+0x5/0x10
Jul 22 09:44:27 dommy0 kernel: [  370.097524] [<ffffffff810c2519>] ? 
hrtimer_interrupt+0xf9/0x230
Jul 22 09:44:27 dommy0 kernel: [  370.097528] [<ffffffff81046d86>] ? 
smp_apic_timer_interrupt+0x36/0x50
Jul 22 09:44:27 dommy0 kernel: [  370.097531] [<ffffffff8150c0bd>] ? 
apic_timer_interrupt+0x6d/0x80
Jul 22 09:44:27 dommy0 kernel: [  370.097532]  <EOI> 
[<ffffffff8150ad89>] ? _raw_spin_lock+0x9/0x30
Jul 22 09:44:27 dommy0 kernel: [  370.097537] [<ffffffff814927bb>] ? 
inet_evict_bucket+0x3b/0x140
Jul 22 09:44:27 dommy0 kernel: [  370.097539] [<ffffffff8149287f>] ? 
inet_evict_bucket+0xff/0x140
Jul 22 09:44:27 dommy0 kernel: [  370.097542] [<ffffffff814929b0>] ? 
inet_frag_worker+0x60/0x210
Jul 22 09:44:27 dommy0 kernel: [  370.097545] [<ffffffff8107e3a2>] ? 
process_one_work+0x142/0x3b0
Jul 22 09:44:27 dommy0 kernel: [  370.097547] [<ffffffff815078ed>] ? 
schedule+0x1d/0x70
Jul 22 09:44:27 dommy0 kernel: [  370.097550] [<ffffffff8107eb94>] ? 
worker_thread+0x114/0x440
Jul 22 09:44:27 dommy0 kernel: [  370.097552] [<ffffffff815073ad>] ? 
__schedule+0x2cd/0x7b0
Jul 22 09:44:27 dommy0 kernel: [  370.097554] [<ffffffff8107ea80>] ? 
create_worker+0x1a0/0x1a0
Jul 22 09:44:27 dommy0 kernel: [  370.097557] [<ffffffff81083dfc>] ? 
kthread+0xbc/0xe0
Jul 22 09:44:27 dommy0 kernel: [  370.097559] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:44:27 dommy0 kernel: [  370.097562] [<ffffffff8150b218>] ? 
ret_from_fork+0x58/0x90
Jul 22 09:44:27 dommy0 kernel: [  370.097564] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0

Jul 22 09:44:53 dommy0 kernel: [  396.106303] Modules linked in: 
parport_pc ppdev lp parport bnep rfcomm bluetooth rfkill uinput nfsd 
auth_rpcgss oid_registry nfs_acl nfs lockd grace fscache sunrpc loop 
coretemp kvm ttm drm_kms_helper iTCO_wdt drm psmouse ipmi_si 
iTCO_vendor_support tpm_tis tpm ipmi_msghandler i2c_algo_bit i2c_core 
i7core_edac dcdbas serio_raw pcspkr wmi lpc_ich edac_core mfd_core evdev 
button acpi_power_meter processor thermal_sys ext4 crc16 mbcache jbd2 
sd_mod sg sr_mod cdrom hid_generic usbhid ata_generic hid crc32c_intel 
ata_piix mptsas scsi_transport_sas mptscsih libata mptbase ehci_pci 
scsi_mod uhci_hcd ehci_hcd usbcore usb_common ixgbe dca ptp bnx2 
pps_core mdio
Jul 22 09:44:53 dommy0 kernel: [  396.106347] CPU: 3 PID: 42 Comm: 
kworker/3:1 Tainted: G        W    L 3.18.18-transip-1.6 #1
Jul 22 09:44:53 dommy0 kernel: [  396.106348] Hardware name: Dell Inc. 
PowerEdge R410/01V648, BIOS 1.12.0 07/30/2013
Jul 22 09:44:53 dommy0 kernel: [  396.106353] Workqueue: events 
inet_frag_worker
Jul 22 09:44:53 dommy0 kernel: [  396.106355] task: ffff880224574490 ti: 
ffff8802240a0000 task.ti: ffff8802240a0000
Jul 22 09:44:53 dommy0 kernel: [  396.106356] RIP: 
0010:[<ffffffff8149288d>]  [<ffffffff8149288d>] 
inet_evict_bucket+0x10d/0x140
Jul 22 09:44:53 dommy0 kernel: [  396.106359] RSP: 
0018:ffff8802240a3d58  EFLAGS: 00000206
Jul 22 09:44:53 dommy0 kernel: [  396.106361] RAX: 0000000000000000 RBX: 
0000000000000286 RCX: 0000000000000000
Jul 22 09:44:53 dommy0 kernel: [  396.106362] RDX: ffff88022215ce40 RSI: 
0000000000300000 RDI: ffff88022215cdf0
Jul 22 09:44:53 dommy0 kernel: [  396.106364] RBP: 0000000000000003 R08: 
ffff880222343c00 R09: 0000000000000101
Jul 22 09:44:53 dommy0 kernel: [  396.106365] R10: 0000000000000000 R11: 
0000000000000027 R12: 0000000000000000
Jul 22 09:44:53 dommy0 kernel: [  396.106366] R13: 0000000000000000 R14: 
ffff880222343c00 R15: 0000000000000101
Jul 22 09:44:53 dommy0 kernel: [  396.106368] FS: 0000000000000000(0000) 
GS:ffff88022f260000(0000) knlGS:0000000000000000
Jul 22 09:44:53 dommy0 kernel: [  396.106370] CS:  0010 DS: 0000 ES: 
0000 CR0: 000000008005003b
Jul 22 09:44:53 dommy0 kernel: [  396.106371] CR2: 00007fb1e9884095 CR3: 
000000021c084000 CR4: 00000000000007e0
Jul 22 09:44:53 dommy0 kernel: [  396.106372] Stack:
Jul 22 09:44:53 dommy0 kernel: [  396.106373]  ffffffff81a8aa30 
ffffffff81a8aa38 0000000000000286 ffff88022215ce88
Jul 22 09:44:53 dommy0 kernel: [  396.106376]  ffffffff8149287f 
0000000000000394 ffffffff81a87100 0000000000000001
Jul 22 09:44:53 dommy0 kernel: [  396.106378]  000000000000007c 
0000000000000000 00000000000000c0 ffffffff814929b0
Jul 22 09:44:53 dommy0 kernel: [  396.106380] Call Trace:
Jul 22 09:44:53 dommy0 kernel: [  396.106383] [<ffffffff8149287f>] ? 
inet_evict_bucket+0xff/0x140
Jul 22 09:44:53 dommy0 kernel: [  396.106386] [<ffffffff814929b0>] ? 
inet_frag_worker+0x60/0x210
Jul 22 09:44:53 dommy0 kernel: [  396.106390] [<ffffffff8107e3a2>] ? 
process_one_work+0x142/0x3b0
Jul 22 09:44:53 dommy0 kernel: [  396.106393] [<ffffffff815078ed>] ? 
schedule+0x1d/0x70
Jul 22 09:44:53 dommy0 kernel: [  396.106396] [<ffffffff8107eb94>] ? 
worker_thread+0x114/0x440
Jul 22 09:44:53 dommy0 kernel: [  396.106398] [<ffffffff815073ad>] ? 
__schedule+0x2cd/0x7b0
Jul 22 09:44:53 dommy0 kernel: [  396.106401] [<ffffffff8107ea80>] ? 
create_worker+0x1a0/0x1a0
Jul 22 09:44:53 dommy0 kernel: [  396.106403] [<ffffffff81083dfc>] ? 
kthread+0xbc/0xe0
Jul 22 09:44:53 dommy0 kernel: [  396.106406] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:44:53 dommy0 kernel: [  396.106409] [<ffffffff8150b218>] ? 
ret_from_fork+0x58/0x90
Jul 22 09:44:53 dommy0 kernel: [  396.106411] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:44:53 dommy0 kernel: [  396.106412] Code: a0 00 00 00 41 ff 94 
24 70 40 00 00 48 85 db 75 e5 48 83 c4 28 89 e8 5b 5d 41 5c 41 5d 41 5e 
41 5f c3 0f 1f 40 00 f0 41 ff 47 68 <48> 8b 44 24 08 66 83 00 01 48 89 
df e8 92 df c2 ff f0 41 ff 4f

Jul 22 09:45:21 dommy0 kernel: [  424.094444] Modules linked in: 
parport_pc ppdev lp parport bnep rfcomm bluetooth rfkill uinput nfsd 
auth_rpcgss oid_registry nfs_acl nfs lockd grace fscache sunrpc loop 
coretemp kvm ttm drm_kms_helper iTCO_wdt drm psmouse ipmi_si 
iTCO_vendor_support tpm_tis tpm ipmi_msghandler i2c_algo_bit i2c_core 
i7core_edac dcdbas serio_raw pcspkr wmi lpc_ich edac_core mfd_core evdev 
button acpi_power_meter processor thermal_sys ext4 crc16 mbcache jbd2 
sd_mod sg sr_mod cdrom hid_generic usbhid ata_generic hid crc32c_intel 
ata_piix mptsas scsi_transport_sas mptscsih libata mptbase ehci_pci 
scsi_mod uhci_hcd ehci_hcd usbcore usb_common ixgbe dca ptp bnx2 
pps_core mdio
Jul 22 09:45:21 dommy0 kernel: [  424.094487] CPU: 3 PID: 42 Comm: 
kworker/3:1 Tainted: G        W    L 3.18.18-transip-1.6 #1
Jul 22 09:45:21 dommy0 kernel: [  424.094488] Hardware name: Dell Inc. 
PowerEdge R410/01V648, BIOS 1.12.0 07/30/2013
Jul 22 09:45:21 dommy0 kernel: [  424.094492] Workqueue: events 
inet_frag_worker
Jul 22 09:45:21 dommy0 kernel: [  424.094494] task: ffff880224574490 ti: 
ffff8802240a0000 task.ti: ffff8802240a0000
Jul 22 09:45:21 dommy0 kernel: [  424.094495] RIP: 
0010:[<ffffffff810c08ac>]  [<ffffffff810c08ac>] del_timer+0x1c/0x70
Jul 22 09:45:21 dommy0 kernel: [  424.094500] RSP: 
0018:ffff8802240a3d28  EFLAGS: 00000246
Jul 22 09:45:21 dommy0 kernel: [  424.094502] RAX: ffffffff81895380 RBX: 
0000000000000000 RCX: 0000000000000000
Jul 22 09:45:21 dommy0 kernel: [  424.094503] RDX: ffff88022215ce40 RSI: 
0000000000300000 RDI: ffff88022215cdf0
Jul 22 09:45:21 dommy0 kernel: [  424.094505] RBP: 0000000000000000 R08: 
ffff880222343c00 R09: 0000000000000101
Jul 22 09:45:21 dommy0 kernel: [  424.094506] R10: 0000000000000000 R11: 
0000000000000027 R12: 0000000000000000
Jul 22 09:45:21 dommy0 kernel: [  424.094507] R13: ffff8802245a8000 R14: 
ffff880222343c00 R15: 0000000000000101
Jul 22 09:45:21 dommy0 kernel: [  424.094509] FS: 0000000000000000(0000) 
GS:ffff88022f260000(0000) knlGS:0000000000000000
Jul 22 09:45:21 dommy0 kernel: [  424.094511] CS:  0010 DS: 0000 ES: 
0000 CR0: 000000008005003b
Jul 22 09:45:21 dommy0 kernel: [  424.094512] CR2: 00007fb1e9884095 CR3: 
000000021c084000 CR4: 00000000000007e0
Jul 22 09:45:21 dommy0 kernel: [  424.094513] Stack:
Jul 22 09:45:21 dommy0 kernel: [  424.094514]  0000000000000296 
ffff88022215cdf0 ffff88022215cdf0 0000000000000003
Jul 22 09:45:21 dommy0 kernel: [  424.094517]  ffffffff81a87100 
ffffffff814927f7 ffffffff81a8aa30 ffffffff81a8aa38
Jul 22 09:45:21 dommy0 kernel: [  424.094519]  0000000000000286 
ffff88022215ce88 ffffffff8149287f 0000000000000394
Jul 22 09:45:21 dommy0 kernel: [  424.094521] Call Trace:
Jul 22 09:45:21 dommy0 kernel: [  424.094524] [<ffffffff814927f7>] ? 
inet_evict_bucket+0x77/0x140
Jul 22 09:45:21 dommy0 kernel: [  424.094527] [<ffffffff8149287f>] ? 
inet_evict_bucket+0xff/0x140
Jul 22 09:45:21 dommy0 kernel: [  424.094529] [<ffffffff814929b0>] ? 
inet_frag_worker+0x60/0x210
Jul 22 09:45:21 dommy0 kernel: [  424.094533] [<ffffffff8107e3a2>] ? 
process_one_work+0x142/0x3b0
Jul 22 09:45:21 dommy0 kernel: [  424.094536] [<ffffffff815078ed>] ? 
schedule+0x1d/0x70
Jul 22 09:45:21 dommy0 kernel: [  424.094539] [<ffffffff8107eb94>] ? 
worker_thread+0x114/0x440
Jul 22 09:45:21 dommy0 kernel: [  424.094541] [<ffffffff815073ad>] ? 
__schedule+0x2cd/0x7b0
Jul 22 09:45:21 dommy0 kernel: [  424.094544] [<ffffffff8107ea80>] ? 
create_worker+0x1a0/0x1a0
Jul 22 09:45:21 dommy0 kernel: [  424.094546] [<ffffffff81083dfc>] ? 
kthread+0xbc/0xe0
Jul 22 09:45:21 dommy0 kernel: [  424.094549] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:45:21 dommy0 kernel: [  424.094552] [<ffffffff8150b218>] ? 
ret_from_fork+0x58/0x90
Jul 22 09:45:21 dommy0 kernel: [  424.094554] [<ffffffff81083d40>] ? 
kthread_create_on_node+0x1c0/0x1c0
Jul 22 09:45:21 dommy0 kernel: [  424.094555] Code: 66 66 66 66 66 66 2e 
0f 1f 84 00 00 00 00 00 48 83 ec 28 48 89 5c 24 10 48 89 6c 24 18 31 ed 
4c 89 64 24 20 48 83 3f 00 48 89 fb <48> c7 47 38 00 00 00 00 74 30 48 
8d 7f 18 48 8d 74 24 08 e8 0c

-- 

TransIP BV

Schipholweg 11E
2316XB Leiden
E: fschreuder@...nsip.nl
I: https://www.transip.nl

--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ