Here is what happened... I recently upgraded a system running 6 x 2TB HDDs with an EFI motherboard and 6 x 3TB HDDs. The final step in the process was growing the RAID-5 array using metadata v0.90 (/dev/md1) consisting of 6 component devices of just under 2TB each to use devices of just under 3TB each. At the time I forgot about the limitation that metadata 0.90 does not support component devices over 2TB. However, the grow completed successfully and I was using the system just fine for about 2 eeks. LVM2 is using /dev/md1 as a physical volume for volume group radagast and pvdisplay showed that /dev/md1 has a size of 13.64 TiB. I had been writing data to it regularly and I believe that I had well exceeded the original size of the old array (about 9.4TB). All was fine until a few days ago when I rebooted the system. The system booted back up to a point, when it could not mount some of the file systems that were on logical volumes on /dev/md1. So, it seems that the mdadm --grow operation was successful, but upon boot the mdadm --assemble operation completed, but not with the same size array as after the grow operation. Here is some relavant information: $ sudo pvdisplay /dev/md1 --- Physical volume --- PV Name /dev/md1 VG Name radagast PV Size 13.64 TiB / not usable 2.81 MiB Allocatable yes PE Size 4.00 MiB Total PE 3576738 Free PE 561570 Allocated PE 3015168 PV UUID 0ay0Ai-jcws-yPAR-DP83-Fha5-LZDO-341dQt Detail of /dev/md1 after the attempt to reboot. Unfortunately I don't have any detail of the array prior to the grow or reboot. However, the pvdisplay above does show the 13.64 TiB size of the array after the grow operation. $ sudo mdadm --detail /dev/md1 /dev/md1: Version : 00.90 Creation Time : Wed May 20 17:19:50 2009 Raid Level : raid5 Array Size : 3912903680 (3731.64 GiB 4006.81 GB) Used Dev Size : 782580736 (746.33 GiB 801.36 GB) Raid Devices : 6 Total Devices : 6 Preferred Minor : 1 Persistence : Superblock is persistent Update Time : Fri Jun 10 00:35:43 2011 State : clean Active Devices : 6 Working Devices : 6 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 256K UUID : 6650f3f8:19abfca8:e368bf24:bd0fce41 Events : 0.6539960 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 19 1 active sync /dev/sdb3 2 8 67 2 active sync /dev/sde3 3 8 51 3 active sync /dev/sdd3 4 8 35 4 active sync /dev/sdc3 5 8 83 5 active sync /dev/sdf3 Code: # mdadm --detail /dev/md1 /dev/md1: Version : 00.90 Creation Time : Wed May 20 17:19:50 2009 Raid Level : raid5 Array Size : 3912903680 (3731.64 GiB 4006.81 GB) Used Dev Size : 782580736 (746.33 GiB 801.36 GB) Raid Devices : 6 Total Devices : 6 Preferred Minor : 1 Persistence : Superblock is persistent Update Time : Tue Jun 7 02:19:18 2011 State : clean Active Devices : 6 Working Devices : 6 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 256K UUID : 6650f3f8:19abfca8:e368bf24:bd0fce41 Events : 0.6539960 Number Major Minor RaidDevice State 0 8 3 0 active sync /dev/sda3 1 8 19 1 active sync /dev/sdb3 2 8 67 2 active sync /dev/sde3 3 8 51 3 active sync /dev/sdd3 4 8 35 4 active sync /dev/sdc3 5 8 83 5 active sync /dev/sdf3 And here is the partitioning layout of each drive. /dev/sd[abcdef] have identical partitioning layouts. $ sudo parted /dev/sda print Model: ATA Hitachi HDS72303 (scsi) Disk /dev/sda: 3001GB Sector size (logical/physical): 512B/512B Partition Table: gpt Number Start End Size File system Name Flags 1 17.4kB 1066kB 1049kB bios_grub 2 1066kB 207MB 206MB ext3 raid 3 207MB 3001GB 3000GB raid