Package: mdadm Version: 2.6.7.1 Severity: important i use ubuntu 10.04 . it's possible that i forgot once to "--set-faulty" before removing a device. first i created raid1 with 2 devices. after encountering the bug i grow the raid from 2 devices to 4 hoping it will make room for other disks to add. here is the config:
dennis1 ~ $ cat /proc/mdstatiscuous mode Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md1 : active raid1 sdd1[2] sdc1[0] sda1[1] 30716160 blocks [3/3] [UUU] bitmap: 14/235 pages [56KB], 64KB chunk md3 : active raid1 sda3[4](S) sdd3[5] sdc3[0] 452550976 blocks [4/1] [U___] bitmap: 45/216 pages [180KB], 1024KB chunk unused devices: <none> dennis1 ~ $ sudo mdadm --examine /dev/sdc3 [sudo] password for denis: /dev/sdc3: Magic : a92b4efc Version : 00.90.00 UUID : 4c2769e4:568869bc:2a972c67:788b998b (local to host dennis1) Creation Time : Wed Dec 22 05:55:45 2010 Raid Level : raid1 Used Dev Size : 452550976 (431.59 GiB 463.41 GB) Array Size : 452550976 (431.59 GiB 463.41 GB) Raid Devices : 4 Total Devices : 3 Preferred Minor : 3 Update Time : Sun Dec 26 06:40:59 2010 State : clean Internal Bitmap : present Active Devices : 1 Working Devices : 3 Failed Devices : 3 Spare Devices : 2 Checksum : a424be5c - correct Events : 170512 Number Major Minor RaidDevice State this 0 8 35 0 active sync /dev/sdc3 0 0 8 35 0 active sync /dev/sdc3 1 1 0 0 1 faulty removed 2 2 0 0 2 faulty removed 3 3 0 0 3 faulty removed 4 4 8 3 4 spare /dev/sda3 5 5 8 51 5 spare /dev/sdd3 dennis1 ~ $ sudo mdadm --examine /dev/sda3 /dev/sda3: Magic : a92b4efc Version : 00.90.00 UUID : 4c2769e4:568869bc:2a972c67:788b998b (local to host dennis1) Creation Time : Wed Dec 22 05:55:45 2010 Raid Level : raid1 Used Dev Size : 452550976 (431.59 GiB 463.41 GB) Array Size : 452550976 (431.59 GiB 463.41 GB) Raid Devices : 4 Total Devices : 3 Preferred Minor : 3 Update Time : Sun Dec 26 06:42:05 2010 State : clean Internal Bitmap : present Active Devices : 1 Working Devices : 3 Failed Devices : 3 Spare Devices : 2 Checksum : a424becc - correct Events : 170550 Number Major Minor RaidDevice State this 4 8 3 4 spare /dev/sda3 0 0 8 35 0 active sync /dev/sdc3 1 1 0 0 1 faulty removed 2 2 0 0 2 faulty removed 3 3 0 0 3 faulty removed 4 4 8 3 4 spare /dev/sda3 5 5 8 51 5 spare /dev/sdd3 dennis1 ~ $ sudo mdadm --detail /dev/md3 /dev/md3: Version : 00.90 Creation Time : Wed Dec 22 05:55:45 2010 Raid Level : raid1 Array Size : 452550976 (431.59 GiB 463.41 GB) Used Dev Size : 452550976 (431.59 GiB 463.41 GB) Raid Devices : 4 Total Devices : 3 Preferred Minor : 3 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Sun Dec 26 06:44:20 2010 State : active, degraded Active Devices : 1 Working Devices : 3 Failed Devices : 0 Spare Devices : 2 UUID : 4c2769e4:568869bc:2a972c67:788b998b (local to host dennis1) Events : 0.170627 Number Major Minor RaidDevice State 0 8 35 0 active sync /dev/sdc3 1 0 0 1 removed 5 8 51 2 spare rebuilding /dev/sdd3 3 0 0 3 removed 4 8 3 - spare /dev/sda3 out of curiosity i will : dennis1 ~ $ sudo mdadm --grow --force -n 1 /dev/md3 mdadm: Cannot set raid-devices for /dev/md3: Device or resource busy dennis1 ~ $ sudo mdadm --grow -n 2 /dev/md3 dennis1 ~ $ sudo mdadm --grow --force -n 1 /dev/md3 mdadm: Cannot set raid-devices for /dev/md3: Device or resource busy dennis1 ~ $ cat /proc/mdstat Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md1 : active raid1 sdd1[2] sdc1[0] sda1[1] 30716160 blocks [3/3] [UUU] bitmap: 16/235 pages [64KB], 64KB chunk md3 : active raid1 sda3[2](S) sdd3[3] sdc3[0] 452550976 blocks [2/1] [U_] bitmap: 45/216 pages [180KB], 1024KB chunk unused devices: <none> dennis1 ~ $ sudo mdadm --examine /dev/sdc3 /dev/sdc3: Magic : a92b4efc Version : 00.90.00 UUID : 4c2769e4:568869bc:2a972c67:788b998b (local to host dennis1) Creation Time : Wed Dec 22 05:55:45 2010 Raid Level : raid1 Used Dev Size : 452550976 (431.59 GiB 463.41 GB) Array Size : 452550976 (431.59 GiB 463.41 GB) Raid Devices : 2 Total Devices : 3 Preferred Minor : 3 Update Time : Sun Dec 26 06:49:12 2010 State : clean Internal Bitmap : present Active Devices : 1 Working Devices : 3 Failed Devices : 1 Spare Devices : 2 Checksum : a424c261 - correct Events : 170800 Number Major Minor RaidDevice State this 0 8 35 0 active sync /dev/sdc3 0 0 8 35 0 active sync /dev/sdc3 1 1 0 0 1 faulty removed 2 2 8 3 2 spare /dev/sda3 3 3 8 51 3 spare /dev/sdd3 dennis1 ~ $ sudo mdadm --grow -n 5 /dev/md3 dennis1 ~ $ sudo mdadm --examine /dev/sdc3 /dev/sdc3: Magic : a92b4efc Version : 00.90.00 UUID : 4c2769e4:568869bc:2a972c67:788b998b (local to host dennis1) Creation Time : Wed Dec 22 05:55:45 2010 Raid Level : raid1 Used Dev Size : 452550976 (431.59 GiB 463.41 GB) Array Size : 452550976 (431.59 GiB 463.41 GB) Raid Devices : 5 Total Devices : 3 Preferred Minor : 3 Update Time : Sun Dec 26 06:49:53 2010 State : clean Internal Bitmap : present Active Devices : 1 Working Devices : 3 Failed Devices : 4 Spare Devices : 2 Checksum : a424c2fd - correct Events : 170826 Number Major Minor RaidDevice State this 0 8 35 0 active sync /dev/sdc3 0 0 8 35 0 active sync /dev/sdc3 1 1 0 0 1 faulty removed 2 2 0 0 2 faulty removed 3 3 0 0 3 faulty removed 4 4 0 0 4 faulty removed 5 5 8 3 5 spare /dev/sda3 6 6 8 51 6 spare /dev/sdd3 dennis1 ~ $ cat /proc/mdstat Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md1 : active raid1 sdd1[2] sdc1[0] sda1[1] 30716160 blocks [3/3] [UUU] bitmap: 14/235 pages [56KB], 64KB chunk md3 : active raid1 sda3[5](S) sdd3[6] sdc3[0] 452550976 blocks [5/1] [U____] bitmap: 45/216 pages [180KB], 1024KB chunk unused devices: <none> dennis1 ~ $ sudo mdadm /dev/md3 -add /dev/sda3 /dev/sdd3 [sudo] password for denis: mdadm: option -d not valid in manage mode dennis1 ~ $ sudo mdadm /dev/md3 -add /dev/sda3 mdadm: option -d not valid in manage mode dennis1 ~ $ sudo mdadm /dev/md3 -a /dev/sda3 mdadm: Cannot open /dev/sda3: Device or resource busy dennis1 ~ $ sudo mdadm /dev/md3 -a /dev/sdd3 mdadm: Cannot open /dev/sdd3: Device or resource busy the simptoms are identical: if i remove now one of the spares, zero the superblock and add it back it will start rebuilding on it untill around 50% and then it will stop leaving me with the above data. this happens only for the md3 device. for the md1 never had a problem. thanks Denis Toader -- To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org