Page 1 of 1
Quantifying Rebuild Times
Posted: Thu Jul 31, 2014 5:01 am
by slink
Any one know of how to test rebuild times on RAID5 900GB FC drives? I know this is hugely dependent on the data written on the drive and the number of drives participating in the rebuild but is there any way to estimate or calculate this kind of thing?
Basically my proposal to use RAID5 has been questioned by cautious stakeholders and they want to use RAID6 to head off the possibility of multiple near simultaneous drive failures. The explanations of fast raid are being persuasive but they want cold hard figures from HP. I've contacted our rep but wondered if anybody on here could suggest anything.
Also, anyone know of any performance comparisons between FC RAID5 3+1 & RAID6 4+2?
thanks
Re: Quantifying Rebuild Times
Posted: Fri Aug 01, 2014 1:12 am
by Richard Siemers
Here is an old one, that compares R10 and the various R5s...
http://3parblog.typepad.com/.a/6a00e553 ... 970c-popupAnd here is a comparative illustration, without specific numbers:
- Screen Shot 2014-08-01 at 1.10.53 AM.png (182.82 KiB) Viewed 17229 times
Re: Quantifying Rebuild Times
Posted: Tue Aug 05, 2014 7:40 pm
by Josh26
It's an incredibly frustrating thing because people will invariably talk about "traditional" raid5 and get their jimmies rustled over running 96 disks or whatever reliant on not seeing two failures at once. Whereas the RAID5 in a 3par world is closer aligned to "5+0".
In a R5 3+1 configuration, one disk going offline only leaves those remaining three vulnerable, and relocating appropriate chunks has been a matter of minutes in our experience.
Re: Quantifying Rebuild Times
Posted: Thu Aug 07, 2014 12:12 pm
by Richard Siemers
Drive "failures" is my favorite topic, as its my most common issue.
This is a 300gb 15K FC drive that was aprox. 83% allocated.
Aug 03 2014 04:24:53 CDT,pd 351 failure: hardware failed for I/O- Internal reason:- Sense key 0004 : Hardware error.- Asc/ascq 0032/0000 : No defect spare location available.- All used chunklets on this disk will be relocated.,Disk fail alert
Aug 03 2014 05:39:34 CDT,Magazine 8:7:3 Physical Disk 351 Failed (Vacated Invalid Media Failed Hardware),Component state change
So 1h 15m for a "vacate" operation. = Aprox 52 Mb/sec
However, this is a vacate operation, not a rebuild from parity... hence, I don't think there is any exposure to double disk failure during THIS specific case. I "presume" what happened is a block was written, then failed verification... since there were no more "defect spare" locations available, I assume that immediate write was immediately serviced on a new disk, in a system "spare" chunk let, then the remainder of the drive was vacated. My theory being even though this drive "failed", its data never is not subject to "double drive jeopardy".
Does anyone have a log sample of a true drive failure where data has to be rebuilt from parity?
Re: Quantifying Rebuild Times
Posted: Sun Aug 10, 2014 5:33 am
by bajorgensen
For what it's worth I am not in the storage dept,
but we do have thousands of drives in 3PAR systems.
For all the cases I am aware of, there has only been predictive drive failures.
First time I discovered a 3PAR in degraded state I was concerned as we did not get an alert,
but the 3PAR will vacate the drive slowly in the background (low priority?)
and then fail the drive/alert HP.
The process seems slow, but I guess a rebuild would have much higher priority.
Re: Quantifying Rebuild Times
Posted: Mon Aug 11, 2014 8:25 pm
by afidel
bajorgensen wrote:For what it's worth I am not in the storage dept,
but we do have thousands of drives in 3PAR systems.
For all the cases I am aware of, there has only been predictive drive failures.
First time I discovered a 3PAR in degraded state I was concerned as we did not get an alert,
but the 3PAR will vacate the drive slowly in the background (low priority?)
and then fail the drive/alert HP.
The process seems slow, but I guess a rebuild would have much higher priority.
I'm pretty sure it's yes on low priority, 3Par seems to do all background tasks on a 20% or less threshold (ie if I compare what AO or multiple tuneld jobs does as far as data movement on the backend to what I can push with a few svmotions it's in the 15-20% range). I also only have predictive failures to go by as far as time to rebuild (that shows to me that their monitoring is quite good as we've had hard failures on all of our previous arrays).