Hello, I have a RAID5 array on my CentOS 5.6 x86_64 workstation which "suddenly" failed to work (actually after the system could not resume from a suspend). I had recently issues after moving the workstation to another office, where one of the disks got accidently unplugged. But the RAID was working and it had reconstructed (as far as I can tell) the data. After I replugged the disk, it was working normally over the last two days (enough to get back to a proper state I guess/hoped) This RAID is used as an LVM volume group for all my important data, among them the root of the operating system(s). It based on four partitions on four separate disks (the third partition of each disk, 3 active, one spare) When booting, I get an error message similar to: raid5 failed: No md superblock detected on /dev/md0. and the LVM volume group does not come up. I then booted using the CentOS 5.6 LiveCD and tried to run a few mdadm command (see just below). It seems that there are some data still lying around, but I'm not very experienced with RAID and I thought that I would ask for advice before trying commands which may impact the data such as assemble, etc. Thanks in advance for any help! Mathieu [root at livecd ~]# mdadm --misc -E /dev/md0 mdadm: No md superblock detected on /dev/md0. [root at livecd ~]# mdadm --misc -Q /dev/md0 /dev/md0: is an md device which is not active /dev/md0: No md super block found, not an md component. [root at livecd ~]# mdadm --misc -D /dev/md0 mdadm: md device /dev/md0 does not appear to be active. [root at livecd ~]# mdadm -E /dev/sd*3 /dev/sda3: Magic : a92b4efc Version : 0.90.00 UUID : 7533411a:f066a145:1e89d48e:1a8374a3 Creation Time : Tue Dec 1 12:01:05 2009 Raid Level : raid5 Used Dev Size : 204796416 (195.31 GiB 209.71 GB) Array Size : 409592832 (390.62 GiB 419.42 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 0 Update Time : Wed Aug 17 14:47:36 2011 State : active Active Devices : 2 Working Devices : 2 Failed Devices : 1 Spare Devices : 0 Checksum : ed6d5dcd - correct Events : 38857 Layout : left-symmetric Chunk Size : 256K Number Major Minor RaidDevice State this 0 8 3 0 active sync /dev/sda3 0 0 8 3 0 active sync /dev/sda3 1 1 0 0 1 faulty removed 2 2 8 51 2 active sync /dev/sdd3 /dev/sdb3: Magic : a92b4efc Version : 0.90.00 UUID : 7533411a:f066a145:1e89d48e:1a8374a3 Creation Time : Tue Dec 1 12:01:05 2009 Raid Level : raid5 Used Dev Size : 204796416 (195.31 GiB 209.71 GB) Array Size : 409592832 (390.62 GiB 419.42 GB) Raid Devices : 3 Total Devices : 4 Preferred Minor : 0 Update Time : Wed Aug 10 11:52:15 2011 State : clean Active Devices : 3 Working Devices : 4 Failed Devices : 0 Spare Devices : 1 Checksum : ed63a948 - correct Events : 9022 Layout : left-symmetric Chunk Size : 256K Number Major Minor RaidDevice State this 1 8 19 1 active sync /dev/sdb3 0 0 8 3 0 active sync /dev/sda3 1 1 8 19 1 active sync /dev/sdb3 2 2 8 51 2 active sync /dev/sdd3 3 3 8 35 3 spare /dev/sdc3 /dev/sdc3: Magic : a92b4efc Version : 0.90.00 UUID : 7533411a:f066a145:1e89d48e:1a8374a3 Creation Time : Tue Dec 1 12:01:05 2009 Raid Level : raid5 Used Dev Size : 204796416 (195.31 GiB 209.71 GB) Array Size : 409592832 (390.62 GiB 419.42 GB) Raid Devices : 3 Total Devices : 3 Preferred Minor : 0 Update Time : Tue Aug 16 17:48:59 2011 State : clean Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Checksum : ed6bf2b0 - correct Events : 10670 Layout : left-symmetric Chunk Size : 256K Number Major Minor RaidDevice State this 1 8 35 1 active sync /dev/sdc3 0 0 8 3 0 active sync /dev/sda3 1 1 8 35 1 active sync /dev/sdc3 2 2 8 51 2 active sync /dev/sdd3 /dev/sdd3: Magic : a92b4efc Version : 0.90.00 UUID : 7533411a:f066a145:1e89d48e:1a8374a3 Creation Time : Tue Dec 1 12:01:05 2009 Raid Level : raid5 Used Dev Size : 204796416 (195.31 GiB 209.71 GB) Array Size : 409592832 (390.62 GiB 419.42 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 0 Update Time : Wed Aug 17 14:47:36 2011 State : active Active Devices : 2 Working Devices : 2 Failed Devices : 1 Spare Devices : 0 Checksum : ed6d5e01 - correct Events : 38857 Layout : left-symmetric Chunk Size : 256K Number Major Minor RaidDevice State this 2 8 51 2 active sync /dev/sdd3 0 0 8 3 0 active sync /dev/sda3 1 1 0 0 1 faulty removed 2 2 8 51 2 active sync /dev/sdd3 [root at livecd ~]#
> [root at livecd ~]# mdadm --misc -E /dev/md0 > mdadm: No md superblock detected on /dev/md0. > [root at livecd ~]# mdadm --misc -Q /dev/md0 > /dev/md0: is an md device which is not active > /dev/md0: No md super block found, not an md component. > [root at livecd ~]# mdadm --misc -D /dev/md0 > mdadm: md device /dev/md0 does not appear to be active.I could fix the issue. Since the information on internet is a bit messy and scary here is a summary of the problem and of the solution, for future reference: ## PROBLEM Due to some failure related to suspend the RAID5 array became inconsistent. The symptom was that the superblock (which allow auto-configuration of the RAID array) was not recognized. But the underlying member partitions were still recognized as per the following command:> [root at livecd ~]# mdadm -E /dev/sd*3which provided details about their states ## SOLUTION (detailed output of the commands at the end of the post) # Assemble the array with force option mdadm -v --assemble --force /dev/md0 /dev/sd{a,b,c,d}3 # NOTE: assembling without force option was not enough [root at livecd ~]# mdadm -v --assemble /dev/md0 /dev/sd{a,b,c,d}3 ... mdadm: /dev/md0 assembled from 2 drives - not enough to start the array while not clean - consider --force. # Check state mdadm -D /dev/md0 ... Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 0 0 1 removed 2 8 51 2 active sync /dev/sdd3 # Add missing partitions to the array mdadm /dev/md0 -a /dev/sdb3 mdadm /dev/md0 -a /dev/sdc3 # Check that it is now OK mdadm -D /dev/md0 ... Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 19 1 spare rebuilding /dev/sdb3 2 8 51 2 active sync /dev/sdd3 3 8 35 - spare /dev/sdc3 # and watch it rebuild: ... Rebuild Status : 3% complete Many thanks to the CentOS LiveCD team, they saved the day. Cheers, Mathieu ## DETAILED OUTPUT [root at livecd ~]# mdadm -v --assemble --force /dev/md0 /dev/sd{a,b,c,d}3 mdadm: looking for devices for /dev/md0 mdadm: /dev/sda3 is identified as a member of /dev/md0, slot 0. mdadm: /dev/sdb3 is identified as a member of /dev/md0, slot 1. mdadm: /dev/sdc3 is identified as a member of /dev/md0, slot 1. mdadm: /dev/sdd3 is identified as a member of /dev/md0, slot 2. mdadm: added /dev/sdc3 to /dev/md0 as 1 mdadm: added /dev/sdd3 to /dev/md0 as 2 mdadm: added /dev/sda3 to /dev/md0 as 0 mdadm: /dev/md0 has been started with 2 drives (out of 3). [root at livecd ~]# mdadm -D /dev/md0 /dev/md0: Version : 0.90 Creation Time : Tue Dec 1 12:01:05 2009 Raid Level : raid5 Array Size : 409592832 (390.62 GiB 419.42 GB) Used Dev Size : 204796416 (195.31 GiB 209.71 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Wed Aug 17 14:47:36 2011 State : clean, degraded Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 256K UUID : 7533411a:f066a145:1e89d48e:1a8374a3 Events : 0.38857 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 0 0 1 removed 2 8 51 2 active sync /dev/sdd3 [root at livecd ~]# mdadm /dev/md0 -a /dev/sdb3 mdadm: re-added /dev/sdb3 [root at livecd ~]# mdadm /dev/md0 -a /dev/sdc3 mdadm: added /dev/sdc3 [root at livecd ~]# mdadm -D /dev/md0 /dev/md0: Version : 0.90 Creation Time : Tue Dec 1 12:01:05 2009 Raid Level : raid5 Array Size : 409592832 (390.62 GiB 419.42 GB) Used Dev Size : 204796416 (195.31 GiB 209.71 GB) Raid Devices : 3 Total Devices : 4 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Wed Aug 17 14:47:36 2011 State : clean, degraded, recovering Active Devices : 2 Working Devices : 4 Failed Devices : 0 Spare Devices : 2 Layout : left-symmetric Chunk Size : 256K Rebuild Status : 0% complete UUID : 7533411a:f066a145:1e89d48e:1a8374a3 Events : 0.38857 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 19 1 spare rebuilding /dev/sdb3 2 8 51 2 active sync /dev/sdd3 3 8 35 - spare /dev/sdc3