Bug#422554: After reboot, drives marked faulty/removed get reassebled as the mirrored array instead of the healthy ones

Binzberger Viktor bviktor at filozofia.bme.hu
Sun May 6 21:23:44 UTC 2007


Package: mdadm
Version: 2.5.6-9

I don't know whether this is a bug, but here's what happened.
I had two mirrors operational /dev/hd[ac]1 and /dev/hd[ac]3 on md0 and md1, respectively.
On Apr 23, /dev/hdc started to complain about bad sectors. I first marked it as faulty, 
and then removed it from the array with mdadm. I didn't have physical access to the machine, 
so I didn't remove /dev/hdc. On Apr 27, the owner of the machine rebooted the machine with ssh -> "reboot"

It is just now that I've discovered that ever since the reboot, mdadm reassembled the bad hdc[13]
partitions as md?, instead of using the healthy hda[13]!

Here is what I see now:

cat /proc/mdstat:

Personalities : [raid1]
md1 : active raid1 hdc3[1]
      29294400 blocks [2/1] [_U]

md0 : active raid1 hdc1[0]
      4883648 blocks [2/1] [U_]

unused devices: <none>

mdadm -D /dev/md0

/dev/md0:
        Version : 00.90.03
  Creation Time : Tue Jan  1 00:49:59 2002
     Raid Level : raid1
     Array Size : 4883648 (4.66 GiB 5.00 GB)
    Device Size : 4883648 (4.66 GiB 5.00 GB)
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Sun May  6 23:19:41 2007
          State : clean, degraded
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0

           UUID : 712b8608:01ac77cf:8ac60fa3:f9a08b86
         Events : 0.160454

    Number   Major   Minor   RaidDevice State
       0      22        1        0      active sync   /dev/hdc1
       1       0        0        1      removed


mdadm --examine /dev/hda1
/dev/hda1:
          Magic : a92b4efc
        Version : 00.90.01
           UUID : 712b8608:01ac77cf:8ac60fa3:f9a08b86
  Creation Time : Tue Jan  1 00:49:59 2002
     Raid Level : raid1
    Device Size : 4883648 (4.66 GiB 5.00 GB)
     Array Size : 4883648 (4.66 GiB 5.00 GB)
   Raid Devices : 1
  Total Devices : 1
Preferred Minor : 0

    Update Time : Fri Apr 27 20:58:51 2007
          State : clean
 Active Devices : 1
Working Devices : 1
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 2319a8de - correct
         Events : 0.64780


      Number   Major   Minor   RaidDevice State
this     0       3        1        0      active sync   /dev/hda1

   0     0       3        1        0      active sync   /dev/hda1


mdadm --examine /dev/hdc1

/dev/hdc1:
          Magic : a92b4efc
        Version : 00.90.01
           UUID : 712b8608:01ac77cf:8ac60fa3:f9a08b86
  Creation Time : Tue Jan  1 00:49:59 2002
     Raid Level : raid1
    Device Size : 4883648 (4.66 GiB 5.00 GB)
     Array Size : 4883648 (4.66 GiB 5.00 GB)
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 0

    Update Time : Sun May  6 22:33:38 2007
          State : clean
 Active Devices : 1
Working Devices : 1
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 232883dc - correct
         Events : 0.159894


      Number   Major   Minor   RaidDevice State
this     0      22        1        0      active sync   /dev/hdc1

   0     0      22        1        0      active sync   /dev/hdc1
   1     1       0        0        1      faulty removed


Is this thing OK? I don't have the opportunity to experiment with this machine, and I had to fix it in the meanwhile, so 
I cannot give you further debug information on the issue.

Cheers, 
Viktor







More information about the pkg-mdadm-devel mailing list