Package: mdadm Version: 1.8.1-1 Severity: grave Justification: causes data loss
Hi, I created a device md0 w/ 3 block devices; hdf1, hdg1, and hdh1. It was set to use raid5. Today, hdg1 disappeared from the machine. Unfortunately, mdadm 1.8.1 does not appear to be able to assemble md devices in degraded mode: [EMAIL PROTECTED]:/dev$ sudo mdadm -A /dev/md0 --scan mdadm: failed to RUN_ARRAY /dev/md0: Invalid argument in kernel logs: md: md driver 0.90.1 MAX_MD_DEVS=256, MD_SB_DISKS=27 md: md0 stopped. md: bind<hdh1> md: bind<hdf1> md: md0: raid array is not clean -- starting background reconstruction raid5: measuring checksumming speed 8regs : 1352.000 MB/sec 8regs_prefetch: 1132.000 MB/sec 32regs : 924.000 MB/sec 32regs_prefetch: 916.000 MB/sec pII_mmx : 2388.000 MB/sec p5_mmx : 3204.000 MB/sec raid5: using function: p5_mmx (3204.000 MB/sec) md: raid5 personality registered as nr 4 raid5: device hdf1 operational as raid disk 0 raid5: device hdh1 operational as raid disk 2 raid5: cannot start dirty degraded array for md0 RAID5 conf printout: --- rd:3 wd:2 fd:1 disk 0, o:1, dev:hdf1 disk 2, o:1, dev:hdh1 raid5: failed to run raid set md0 md: pers->run() failed ... [EMAIL PROTECTED]:/dev$ sudo mdadm -E /dev/hdh1 /dev/hdh1: Magic : a92b4efc Version : 00.90.00 UUID : 47c630a9:142ec28b:a28fb6b2:8fc0ce54 Creation Time : Mon Feb 24 05:17:28 2003 Raid Level : raid5 Device Size : 156288256 (149.05 GiB 160.04 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 0 Update Time : Tue Jan 25 20:46:04 2005 State : dirty Active Devices : 2 Working Devices : 2 Failed Devices : 1 Spare Devices : 0 Checksum : c12cdab5 - correct Events : 0.1697073 Layout : left-symmetric Chunk Size : 32K Number Major Minor RaidDevice State this 2 34 65 2 active sync /dev/ide/host2/bus1/target1/lun0/part1 0 0 33 65 0 active sync /dev/ide/host2/bus0/target1/lun0/part1 1 1 0 0 1 faulty removed 2 2 34 65 2 active sync /dev/ide/host2/bus1/target1/lun0/part1 Google informed me that I should try the following: [EMAIL PROTECTED]:/dev$ sudo mdadm -S /dev/md0 [EMAIL PROTECTED]:/dev$ sudo mdadm -Af /dev/md0 /dev/hdf1 /dev/hdh1 mdadm: failed to RUN_ARRAY /dev/md0: Invalid argument [EMAIL PROTECTED]:/dev$ sudo mdadm -S /dev/md0 [EMAIL PROTECTED]:/dev$ sudo mdadm -Af --scan /dev/md0 /dev/hdf1 /dev/hdh1 mdadm: failed to RUN_ARRAY /dev/md0: Invalid argument mdadm: /dev/hdf1 not identified in config file. mdadm: /dev/hdh1 not identified in config file. I then downgraded mdadm 1.7.0-1, and: [EMAIL PROTECTED]:~$ sudo mdadm -S /dev/md0 [EMAIL PROTECTED]:~$ sudo mdadm -Af /dev/md0 /dev/hdf1 /dev/hdh1 mdadm: /dev/md0 has been started with 2 drives (out of 3). According to Neil Brown, mdadm 1.8.1 is a development version (http://www.issociate.de/board/post/141215/Multipath_problem.html); he recommends 1.8.0. If this is indeed the case, we should definitely not be releasing sarge w/ this version. The severity of this bug is grave because the whole point of running w/ raid5 is so that if a drive fails, data can be recovered from the remaining 2 drives. Not being able to assemble the md device to recover such data is a bit of a show stopper. -- System Information: Debian Release: 3.1 APT prefers unstable APT policy: (500, 'unstable'), (500, 'testing') Architecture: i386 (i686) Kernel: Linux 2.6.10-1-k7 Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8) -- To UNSUBSCRIBE, email to [EMAIL PROTECTED] with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]