[Beowulf] Large raid rebuild times

Joe Landman landman at scalableinformatics.com
Fri Nov 20 06:43:48 PST 2009


james bardin wrote:
> Hello,
> 
> Has anyone here seen any numbers, or tested themselves, the rebuild

Hi James,

   Yes, we test this (by purposely failing a drive on the storage we 
ship to our customers).

> times for large raid arrays (raid 6 specifically)?

   Yes, RAID6 with up to 24 drives per RAID, up to 2TB drives.  Due to 
the bit error rate failure models for the uncorrectable errors on disks, 
RAID5 is *strongly* contra-indicated for storage of more than a few 
small TB, and more than a few drives (less then 5 and less than 1TB). 
The risk of a second failure during rebuild is simply unacceptably high, 
which would/does permanently take out your data in RAID5.

> I can't seem to find anything concrete to go by, and I haven't had to
> rebuild anything larger than a few TB. What happens when you loose a
> 2TB drive in a 20TB array for instance? Do any hardware raid solutions
> help. I don't think ZFS is an option right now, so I'm looking at

   We have customers with 32TB raw per RAID, and when a drive fails, it 
rebuilds.  Rebuild time is a function of how fast the card is set up to 
do rebuilds, you can tune the better cards in terms of "background" 
rebuild performance.  For low rebuild speeds, we have seen 24 hours+, 
for high rebuild speeds, we have seen 12-15 hours for the 32TB.

   ZFS is probably not what you want to do ... building a critical 
dependency upon a product that has a somewhat uncertain future ...

Joe


-- 
Joseph Landman, Ph.D
Founder and CEO
Scalable Informatics, Inc.
email: landman at scalableinformatics.com
web  : http://scalableinformatics.com
        http://scalableinformatics.com/jackrabbit
phone: +1 734 786 8423 x121
fax  : +1 866 888 3112
cell : +1 734 612 4615



More information about the Beowulf mailing list