Following the above, I did this: 1) stop nfs server daemon, unmount /home, stop /dev/md4 2) make a new md4 with default chunk size (64KiB) but otherwise the same as above 3) made a new xfs filesystem on md4 using md3 as log device, as before 4) mount this as /home, export via nfs 5) restore 1.2TB of backed up data 6) copy several large files over the nfs mount And there are no problems.
So the only significant difference: default chunk size OK, non-default chunk size fails. I guess somewhere in the raid10 code an assumption is made about chunk size. Some documentation of what I did: mdadm -C /dev/md4 -v -n 6 -l 10 -p f2 -f /dev/sd[abcdef]3 mdadm: chunk size defaults to 64K mdadm: /dev/sda3 appears to be part of a raid array: level=raid10 devices=6 ctime=Tue Sep 28 14:34:19 2010 mdadm: /dev/sdb3 appears to be part of a raid array: level=raid10 devices=6 ctime=Tue Sep 28 14:34:19 2010 mdadm: /dev/sdc3 appears to be part of a raid array: level=raid10 devices=6 ctime=Tue Sep 28 14:34:19 2010 mdadm: /dev/sdd3 appears to be part of a raid array: level=raid10 devices=6 ctime=Tue Sep 28 14:34:19 2010 mdadm: /dev/sde3 appears to be part of a raid array: level=raid10 devices=6 ctime=Tue Sep 28 14:34:19 2010 mdadm: /dev/sdf3 appears to be part of a raid array: level=raid10 devices=6 ctime=Tue Sep 28 14:34:19 2010 mdadm: size set to 961980160K Continue creating array? y mdadm: array /dev/md4 started. r...@wibert:~# mkfs.xfs -f -d su=65536,sw=6 -l logdev=/dev/md3 -L home /dev/md4 meta-data=/dev/md4 isize=256 agcount=32, agsize=22546416 blks = sectsz=512 attr=2 data = bsize=4096 blocks=721485120, imaxpct=5 = sunit=16 swidth=96 blks naming =version 2 bsize=4096 ascii-ci=0 log =/dev/md3 bsize=4096 blocks=32112, version=2 = sectsz=512 sunit=0 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 r...@wibert:~# mdadm --detail /dev/md4 /dev/md4: Version : 00.90 Creation Time : Thu Sep 30 14:27:19 2010 Raid Level : raid10 Array Size : 2885940480 (2752.25 GiB 2955.20 GB) Used Dev Size : 961980160 (917.42 GiB 985.07 GB) Raid Devices : 6 Total Devices : 7 Preferred Minor : 4 Persistence : Superblock is persistent Update Time : Fri Oct 1 09:21:19 2010 State : clean Active Devices : 6 Working Devices : 7 Failed Devices : 0 Spare Devices : 1 Layout : near=1, far=2 Chunk Size : 64K UUID : 4e420fd5:e9f8a4ff:fcc517c0:fe041f9d (local to host wibert) Events : 0.38 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 19 1 active sync /dev/sdb3 2 8 35 2 active sync /dev/sdc3 3 8 51 3 active sync /dev/sdd3 4 8 67 4 active sync /dev/sde3 5 8 83 5 active sync /dev/sdf3 6 8 147 - spare /dev/sdj3 When did non-default chunk size last work on raid10? http://blog.jamponi.net/2008/07/raid56-and-10-benchmarks-on-26255_10.html -- raid10 fails - "lost page write due to I/O error on md4" and " raid10_make_request bug: can't convert block across chunks or bigger than 128k 1623343324 20" - write fails, remote nfs mount of filesystem becomes unusable https://bugs.launchpad.net/bugs/651846 You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. -- ubuntu-bugs mailing list ubuntu-bugs@lists.ubuntu.com https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs