Package: mdadm Version: 2.5.6-9 Severity: normal Hello, I am trying to setup raid1 on my pc. I have 2 identical drives. 270+gb. each with 3 partitions. 30gb hdb1 -> md0 250gb hdb2 ->md2 4gb swap hdb5 ->md4
Initially my raid had only one drive. I have added the second one with mdadm --add /dev/md2 /dev/hda1 then 2 then 4 It started doing recovery for drives. IT finished for md0,md4 but for md2 it is in infinite loop. IT goes to 15 % and starts again hplinux:/home/lucas# cat /proc/mdstat Personalities : [raid1] [raid6] [raid5] [raid4] [raid0] md4 : active raid1 hda5[0] hdb5[1] 1951744 blocks [2/2] [UU] md2 : active raid1 hdb2[1] hda2[2] 276438400 blocks [2/1] [_U] [>....................] recovery = 2.8% (7898752/276438400) finish=2448.2min speed=1826K/sec md0 : active raid1 hda1[0] hdb1[1] 34178176 blocks [2/2] [UU] unused devices: <none> Oct 11 19:26:50 hplinux kernel: md: md2: sync done. Oct 11 19:26:51 hplinux kernel: md: syncing RAID array md2 Oct 11 19:26:51 hplinux kernel: md: minimum _guaranteed_ reconstruction speed: 1000 KB/sec/disc. Oct 11 19:26:51 hplinux kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for reconstruction. Oct 11 19:26:51 hplinux kernel: md: using 128k window, over a total of 276438400 blocks. Oct 11 19:32:21 hplinux kernel: md: md2: sync done. Oct 11 19:32:21 hplinux mdadm: RebuildFinished event detected on md device /dev/md2 Oct 11 19:32:21 hplinux kernel: md: syncing RAID array md2 Oct 11 19:32:21 hplinux kernel: md: minimum _guaranteed_ reconstruction speed: 1000 KB/sec/disc. Oct 11 19:32:21 hplinux kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for reconstruction. Oct 11 19:32:21 hplinux kernel: md: using 128k window, over a total of 276438400 blocks. Oct 11 19:32:21 hplinux mdadm: RebuildStarted event detected on md device /dev/md2 Oct 11 19:36:16 hplinux kernel: md: md2: sync done. Oct 11 19:36:17 hplinux kernel: md: syncing RAID array md2 why is this appening? I have checked both partitions for bad blocks with e2fsck. Is there something I am missing? The loops just goes on and on. ITs been 72h so I don't know how much longer with the drives survive this 20mb/s read and write. Thanks, lucas -- Package-specific info: --- mount output /dev/md0 on / type ext3 (rw,errors=remount-ro) tmpfs on /lib/init/rw type tmpfs (rw,nosuid,mode=0755) proc on /proc type proc (rw,noexec,nosuid,nodev) sysfs on /sys type sysfs (rw,noexec,nosuid,nodev) procbususb on /proc/bus/usb type usbfs (rw) udev on /dev type tmpfs (rw,mode=0755) tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev) devpts on /dev/pts type devpts (rw,noexec,nosuid,gid=5,mode=620) /dev/md2 on /files type ext3 (rw) nfsd on /proc/fs/nfsd type nfsd (rw) rpc_pipefs on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw) --- mdadm.conf # mdadm.conf # # Please refer to mdadm.conf(5) for information about this file. # # by default, scan all partitions (/proc/partitions) for MD superblocks. # alternatively, specify devices to scan, using wildcards if desired. DEVICE partitions # auto-create devices with Debian standard permissions CREATE owner=root group=disk mode=0660 auto=yes # automatically tag new arrays as belonging to the local system HOMEHOST <system> # instruct the monitoring daemon where to send mail alerts MAILADDR root # definitions of existing MD arrays # This file was auto-generated on Fri, 05 Oct 2007 21:54:54 -0500 # by mkconf $Id: mkconf 261 2006-11-09 13:32:35Z madduck $ ARRAY /dev/md0 level=raid1 num-devices=2 UUID=8f8b90a7:642e9c4d:b1274a75:6a339511 ARRAY /dev/md2 level=raid1 num-devices=2 UUID=6934dbf9:dfe9d7d1:b1274a75:6a339511 ARRAY /dev/md4 level=raid1 num-devices=2 UUID=967e922e:2688de6a:b1274a75:6a339511 --- /proc/mdstat: Personalities : [raid1] [raid6] [raid5] [raid4] [raid0] md4 : active raid1 hda5[0] hdb5[1] 1951744 blocks [2/2] [UU] md2 : active raid1 hdb2[1] hda2[2] 276438400 blocks [2/1] [_U] [>....................] recovery = 2.4% (6757632/276438400) finish=2488.6min speed=1805K/sec md0 : active raid1 hda1[0] hdb1[1] 34178176 blocks [2/2] [UU] unused devices: <none> --- /proc/partitions: major minor #blocks name 3 0 312571224 hda 3 1 34178256 hda1 3 2 276438487 hda2 3 3 1 hda3 3 5 1951866 hda5 3 64 312571224 hdb 3 65 34178256 hdb1 3 66 276438487 hdb2 3 67 1 hdb3 3 69 1951866 hdb5 9 0 34178176 md0 9 2 276438400 md2 9 4 1951744 md4 --- initrd.img-2.6.18-4-486: 21377 blocks etc/mdadm etc/mdadm/mdadm.conf scripts/local-top/mdadm sbin/mdadm lib/modules/2.6.18-4-486/kernel/drivers/md/multipath.ko lib/modules/2.6.18-4-486/kernel/drivers/md/raid0.ko lib/modules/2.6.18-4-486/kernel/drivers/md/raid1.ko lib/modules/2.6.18-4-486/kernel/drivers/md/raid10.ko lib/modules/2.6.18-4-486/kernel/drivers/md/md-mod.ko lib/modules/2.6.18-4-486/kernel/drivers/md/linear.ko lib/modules/2.6.18-4-486/kernel/drivers/md/raid456.ko lib/modules/2.6.18-4-486/kernel/drivers/md/xor.ko --- /proc/modules: dm_snapshot 15644 0 - Live 0xe0c59000 dm_mirror 18000 0 - Live 0xe0c48000 dm_mod 48952 2 dm_snapshot,dm_mirror, Live 0xe0c63000 raid0 7808 0 - Live 0xe082e000 raid456 113168 0 - Live 0xe08c0000 xor 14344 1 raid456, Live 0xe0857000 raid1 19968 3 - Live 0xe0851000 md_mod 67860 5 raid0,raid456,raid1, Live 0xe0890000 --- volume detail: --- /proc/cmdline root=/dev/md0 ro --- grub: kernel /boot/vmlinuz-2.6.18-4-486 root=/dev/md0 ro kernel /boot/vmlinuz-2.6.18-4-486 root=/dev/hda1 ro kernel /boot/vmlinuz-2.6.18-4-486 root=/dev/hda1 ro single -- System Information: Debian Release: 4.0 APT prefers stable APT policy: (500, 'stable') Architecture: i386 (i686) Shell: /bin/sh linked to /bin/bash Kernel: Linux 2.6.18-4-486 Locale: LANG=en_US.UTF-8, LC_CTYPE=en_US.UTF-8 (charmap=UTF-8) Versions of packages mdadm depends on: ii debconf [debconf-2.0] 1.5.11 Debian configuration management sy ii libc6 2.3.6.ds1-13 GNU C Library: Shared libraries ii lsb-base 3.1-23.1 Linux Standard Base 3.1 init scrip ii makedev 2.3.1-83 creates device files in /dev Versions of packages mdadm recommends: ii module-init-tools 3.3-pre4-2 tools for managing Linux kernel mo ii postfix [mail-transport-agent 2.3.8-2+b1 A high-performance mail transport -- debconf information: * mdadm/autostart: true mdadm/mail_to: root mdadm/initrdstart_msg_errmd: * mdadm/initrdstart: all mdadm/initrdstart_msg_errconf: mdadm/initrdstart_notinconf: false mdadm/initrdstart_msg_errexist: mdadm/initrdstart_msg_intro: mdadm/autocheck: true mdadm/initrdstart_msg_errblock: mdadm/start_daemon: true -- To UNSUBSCRIBE, email to [EMAIL PROTECTED] with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]