[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <170fa0d20805200833s30c0a856p5f08250879a182b9@mail.gmail.com>
Date: Tue, 20 May 2008 11:33:13 -0400
From: "Mike Snitzer" <snitzer@...il.com>
To: "Neil Brown" <neilb@...e.de>
Cc: linux-raid@...r.kernel.org, linux-kernel@...r.kernel.org,
paul.clements@...eleye.com
Subject: Re: [RFC][PATCH] md: avoid fullsync if a faulty member missed a dirty transition
On Tue, May 20, 2008 at 11:30 AM, Mike Snitzer <snitzer@...il.com> wrote:
> On Mon, May 19, 2008 at 1:27 AM, Neil Brown <neilb@...e.de> wrote:
> > On Monday May 19, snitzer@...il.com wrote:
> > >
> > > Hi Neil,
> > >
> > > Sorry about not getting back with you sooner. Thanks for putting
> > > significant time to chasing this problem.
> > >
> > > I tested your most recent patch and unfortunately still hit the case
> > > where the nbd member becomes degraded yet the array continues to clear
> > > bits (events_cleared of the non-degraded member is higher than the
> > > degraded member). Is this behavior somehow expected/correct?
> >
> > It shouldn't be..... ahhh.
> > There is a delay between noting that the bit can be cleared, and
> > actually writing the zero to disk. This is obviously intentional
> > in case the bit gets set again quickly.
> > I'm sampling the event count at the latter point instead of the
> > former, and there is time for it to change.
> >
> > Maybe this patch on top of what I recently sent out?
>
> Hi Neil,
>
> We're much closer. The events_cleared is symmetric on both the failed
> and active member of the raid1. But there have been some instances
> where the md thread hits a deadlock during my testing. What follows
> is the backtrace and live crash info:
>
> md0_raid1 D 000002c4b6483a7f 0 11249 2 (L-TLB)
> ffff81005747dce0 0000000000000046 0000000000000000 ffff8100454c53c0
> 000000000000000a ffff810048fbd0c0 000000000000000a ffff810048fbd0c0
> ffff81007f853840 000000000000148e ffff810048fbd2b0 0000000362c10780
> Call Trace:
> [<ffffffff88ba8503>] :md_mod:bitmap_daemon_work+0x249/0x4d3
> [<ffffffff802457a5>] autoremove_wake_function+0x0/0x2e
> [<ffffffff88ba53b3>] :md_mod:md_check_recovery+0x20/0x4a5
> [<ffffffff8044cb5c>] thread_return+0x0/0xf1
> [<ffffffff88bbe0eb>] :raid1:raid1d+0x25/0xd09
> [<ffffffff8023bcd7>] lock_timer_base+0x26/0x4b
> [<ffffffff8023bd4d>] try_to_del_timer_sync+0x51/0x5a
> [<ffffffff8023bd62>] del_timer_sync+0xc/0x16
> [<ffffffff8044d38a>] schedule_timeout+0x92/0xad
> [<ffffffff88ba6c6c>] :md_mod:md_thread+0xeb/0x101
> [<ffffffff802457a5>] autoremove_wake_function+0x0/0x2e
> [<ffffffff88ba6b81>] :md_mod:md_thread+0x0/0x101
> [<ffffffff8024564d>] kthread+0x47/0x76
> [<ffffffff8020aa38>] child_rip+0xa/0x12
> [<ffffffff80245606>] kthread+0x0/0x76
> [<ffffffff8020aa2e>] child_rip+0x0/0x12
>
> crash> bt 11249
> PID: 11249 TASK: ffff810048fbd0c0 CPU: 3 COMMAND: "md0_raid1"
> #0 [ffff81005747dbf0] schedule at ffffffff8044cb5c
> #1 [ffff81005747dce8] bitmap_daemon_work at ffffffff88ba8503
> #2 [ffff81005747dd68] md_check_recovery at ffffffff88ba53b3
> #3 [ffff81005747ddb8] raid1d at ffffffff88bbe0eb
> #4 [ffff81005747ded8] md_thread at ffffffff88ba6c6c
> #5 [ffff81005747df28] kthread at ffffffff8024564d
> #6 [ffff81005747df48] kernel_thread at ffffffff8020aa38
>
> 0xffffffff88ba84ee <bitmap_daemon_work+0x234>: callq
> 0xffffffff802458ec <prepare_to_wait>
> 0xffffffff88ba84f3 <bitmap_daemon_work+0x239>: mov 0x18(%rbx),%rax
> 0xffffffff88ba84f7 <bitmap_daemon_work+0x23d>: mov 0x28(%rax),%eax
> 0xffffffff88ba84fa <bitmap_daemon_work+0x240>: test $0x2,%al
> 0xffffffff88ba84fc <bitmap_daemon_work+0x242>: je
> 0xffffffff88ba8505 <bitmap_daemon_work+0x24b>
> 0xffffffff88ba84fe <bitmap_daemon_work+0x244>: callq
> 0xffffffff8044c200 <__sched_text_start>
> 0xffffffff88ba8503 <bitmap_daemon_work+0x249>: jmp
> 0xffffffff88ba84d6 <bitmap_daemon_work+0x21c>
> 0xffffffff88ba8505 <bitmap_daemon_work+0x24b>: mov 0x18(%rbx),%rdi
> 0xffffffff88ba8509 <bitmap_daemon_work+0x24f>: mov %rbp,%rsi
> 0xffffffff88ba850c <bitmap_daemon_work+0x252>: add $0x200,%rdi
> 0xffffffff88ba8513 <bitmap_daemon_work+0x259>: callq
> 0xffffffff802457f6 <finish_wait>
>
> So running with your latest patches seems to introduce a race in
> bitmap_daemon_work's if (unlikely((*bmc & COUNTER_MAX) ==
> COUNTER_MAX)) { } block.
Err, that block is in bitmap_startwrite()...
Mike
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists