[NTLUG:Discuss] RAID recovery help!

Richard Geoffrion ntlug at rain4us.net
Tue Nov 4 22:19:03 CST 2003


WHAT!?!? WHY!?!?   Why is this not working! What am I missing?

[info]
root at gwifs:/var/log# cat /proc/mdstat
Personalities : [raid0] [raid1]
read_ahead 1024 sectors
md0 : active raid1 sdb1[1] sda1[0]
      32000 blocks [2/2] [UU]

md1 : active raid1 sdb5[1] sda5[0](F)
      4000064 blocks [2/1] [_U]

md2 : active raid1 sdb6[1] sda6[0]
      1999936 blocks [2/2] [UU]

md3 : active raid1 sdb7[1] sda7[0]
      4000064 blocks [2/2] [UU]

md4 : active raid1 sdb8[1] sda8[0]
      4000064 blocks [2/2] [UU]

md5 : active raid1 sdb9[1] sda9[0]
      57400128 blocks [2/2] [UU]

unused devices: <none>
root at gwifs:/var/log# raidhotadd /dev/md1 /dev/sdb5
/dev/md1: can not hot-add disk: disk busy!

[/info]

The server had an unscheduled reboot this past weekend...but the message.log
shows no /dev/md1 raid errors.

[/var/log/messages snipet]
Nov  2 11:30:57 gwifs kernel: md: considering sdb5 ...
Nov  2 11:30:58 gwifs kernel: md:  adding sdb5 ...
Nov  2 11:30:58 gwifs kernel: md:  adding sda5 ...
Nov  2 11:30:58 gwifs kernel: md: created md1
Nov  2 11:30:58 gwifs kernel: md: bind<sda5,1>
Nov  2 11:30:58 gwifs kernel: md: bind<sdb5,2>
Nov  2 11:30:58 gwifs kernel: md: running: <sdb5><sda5>
Nov  2 11:30:58 gwifs kernel: md: sdb5's event counter: 0000006a
Nov  2 11:30:58 gwifs kernel: md: sda5's event counter: 0000006a
Nov  2 11:30:58 gwifs kernel: md: RAID level 1 does not need chunksize!
Continuing anyway.
Nov  2 11:30:58 gwifs kernel: md1: max total readahead window set to 124k
Nov  2 11:30:58 gwifs kernel: md1: 1 data-disks, max readahead per
data-disk: 124k
Nov  2 11:30:58 gwifs kernel: raid1: device sdb5 operational as mirror 1
Nov  2 11:30:58 gwifs kernel: raid1: device sda5 operational as mirror 0
Nov  2 11:30:58 gwifs kernel: raid1: raid set md1 active with 2 out of 2
mirrors
Nov  2 11:30:58 gwifs kernel: md: updating md1 RAID superblock on device
Nov  2 11:30:58 gwifs kernel: md: sdb5 [events: 0000006b]<6>(write) sdb5's
sb offset: 4000064
Nov  2 11:30:58 gwifs kernel: md: delaying resync of md1 until md5 has
finished resync (they share one or$
Nov  2 11:30:58 gwifs kernel: md: sda5 [events: 0000006b]<6>(write) sda5's
sb offset: 4000064

[/snippet]

well....further looking DID turn up this...

[another snippet]
Nov  2 12:00:01 gwifs kernel: md: syncing RAID array md1
Nov  2 12:00:01 gwifs kernel: md: minimum _guaranteed_ reconstruction speed:
100 KB/sec/disc.
Nov  2 12:00:01 gwifs kernel: md: using maximum available idle IO bandwith
(but not more than 100000 KB/s$
Nov  2 12:00:01 gwifs kernel: md: using 124k window, over a total of 4000064
blocks.
Nov  2 12:00:01 gwifs kernel: md: delaying resync of md2 until md1 has
finished resync (they share one or$
Nov  2 12:00:01 gwifs kernel: md: delaying resync of md3 until md1 has
finished resync (they share one or$
Nov  2 12:00:58 gwifs kernel: raid1: mirror resync was not fully finished,
restarting next time.
Nov  2 12:00:58 gwifs kernel: md: recovery thread got woken up ...
Nov  2 12:00:58 gwifs kernel: md: updating md1 RAID superblock on device
Nov  2 12:00:58 gwifs kernel: md: sdb5 [events: 0000006c]<6>(write) sdb5's
sb offset: 4000064
Nov  2 12:00:58 gwifs kernel: md: (skipping faulty sda5 )
Nov  2 12:00:58 gwifs kernel: md: recovery thread finished ...
Nov  2 12:00:58 gwifs kernel: md: md_do_sync() got signal ... exiting
Nov  2 12:00:58 gwifs kernel: raid1: mirror resync was not fully finished,
restarting next time.
Nov  2 12:00:58 gwifs last message repeated 25 times
[/another snippet]

so...how can /dev/md1 (sda5) be faulting...but all the other mdXs be
fine...they are all on the same physical disk!


--
Richard




More information about the Discuss mailing list