RAID Performance Calculators
-
@DustinB3403 said:
Well SATA supports up to 6GB/S
With my calculations I can push 4.4GB/m or 700MB/S (write)
So I don't think so.
Not a single drive. But an array definitely can.
-
@Dashrender my calculations are a 12 disk RAID 5 array.
A bigger system might.
-
@DustinB3403 said:
Yeah I did the math on the SSD drives above, and the rates IOPS is 4.4 GB/m.
Drive performance is not measured in GB/m. It is measured in IOPS.
-
Looking at throughput numbers for drives is almost always useless. If you are building a streaming video server or a backup target that takes a single backup stream at a time, okay, there is a time where throughput can matter. But it is rare.
There is a reason why IOPS is the only number generally used when talking storage performance - because it is the only one of significance. It is only because of this that things like SANs have any hope of working as they have terribly slow throughput bottlenecks between them and the servers that they support. But most businesses can run from iSCSI over 1GigE wires. Why? Because it is the IOPS that matter, rarely the throughput.
If you look at throughput numbers, you will come up with some crazily dangerous comparisons that will lead you in some terrible decision making directions.
-
@scottalanmiller said:
There is a reason why IOPS is the only number generally used when talking storage performance - because it is the only one of significance.
I've not even bothered to count how many times he has been told that in this thread yet he keeps not listening.
-
@JaredBusch said:
@scottalanmiller said:
There is a reason why IOPS is the only number generally used when talking storage performance - because it is the only one of significance.
I've not even bothered to count how many times he has been told that in this thread yet he keeps not listening.
Speaking to him, he was confused and thought that GB/m was an IOPS measurement and did not realize that IOPS was the units.
-
So how to figure out the IOPS of an array? Start with getting the IOPS numbers from the drives. If we are dealing with 50,000 Read IOPS and 100,000 Write IOPS you just use the formula from the below link and you will get the rough IOPS number:
Now you have to deal with total capacity of the RAID controller. You can only have so many IOPS before the RAID controller cannot push it.
-
This article helps explain the issues with the RAID controller limits...
http://mangolassi.it/topic/2072/testing-the-limits-of-the-dell-h710-raid-controller-with-ssd
-
OK So I read though the articles provided, Thank you @scottalanmiller .
In doing the math, I have 1 remaining question. Should I use QD1 or QD32 read/write performance markers?
I did the math with QD1 IOPS 10000/40000 respectively. With an 80/20 Ratio a baseline. (I'm sure this needs to be verified with DPACK though).
But if I do the math with QD32 I'm guessing I'll have a dramtically different resulting number. Since QD32 is 197,000 /88,000 respectively.
-
How many IOPs do you need? Assuming you're not in an IOP defect now, DPACk should tell you that to. But if you are in a defect today, it will be much harder to know.
If you have the time and resources, you could see about throwing an SSD in a system, loading it up with your workload and see what DPACK tells you then...
-
With QD1 on a RAID 5 12 disk array I'm looking at 48,000 IOPS.
If I use QD32 on a RAID 5 12 disk array I'm looking at 525,600 IOPS.
Can someone clarify this?
-
@DustinB3403 said:
In doing the math, I have 1 remaining question. Should I use QD1 or QD32 read/write performance markers?
That's tough. That you can only determine from measuring your actual usage and, in reality, you need a big blend of numbers. I'd start by getting a number from both to provide a range of reasonable possibilities.
-
@DustinB3403 said:
With QD1 on a RAID 5 12 disk array I'm looking at 48,000 IOPS.
If I use QD32 on a RAID 5 12 disk array I'm looking at 525,600 IOPS.
Can someone clarify this?
You'll likely be somewhere very much in the middle. These are kind of the best case and worst case numbers on a very large curve.
-
@Dashrender We likely need very little IOPS. We don't run any intensive applications, most literally Network shares and Domain functions.
-
One of the secrets of IOPS, and of many many things in IT, is that there is not a real answer in terms of an actual number, the numbers are massive estimates. What you should actually get is a big curve or 3D wave that represents IOPS under different conditions.
-
Should I really be concerned about this?
The goal is to run and host VM's and network share data off of two XenServer Host. Or should I simply state that "on the low end, we'd see IOPS at 48K to 525K, which is still mountains faster than the SR disk we have now in stand alone servers"
-
@DustinB3403 said:
@Dashrender We likely need very little IOPS. We don't run any intensive applications, most literally Network shares and Domain functions.
What is your current IOP availability? Are you having drive related issues today? If not, then as long as you match the current number (nearly impossible to not blow it away with SSD drives) you should be golden.
If I have 8 SAS 15K drives in RAID 10 today to replace them with 8 SSD in RAID 5, I personally wouldn't even look at IOP numbers as they are 10-1000x more than they were before, and probably 100-1000x more.
-
@Dashrender sadly we're in a RAID 5 SR Array at the moment.
I'm just looking for more "ammunition' for this proposal.
-
Then your gain will be even greater!
-
@DustinB3403 said:
Should I really be concerned about this?
You need to be concerned that you have "enough IOPS." But realistically you are looking at overrunning your controller with that many SSDs. So.... is it reasonable to worry that your smallish company's needs will be in excess of the IOPS possible from a high end enterprise RAID controller?
No, not reasonable. You would know if you were doing something super crazy that would require special case storage. In the real world, small companies of say 500 and fewer users have been able to function just fine off of large RAID 6 spinning rust arrays for years. Just moving from 10K to 15K drives can be a jump beyond what is needed. And then moving from R6 to R10 gives a bit leap. A large R10 array of 10K drives is enough for nearly any companies.
Having a 12 disk R5 SSD array is so many orders of magnitude faster than even very large, very fast spinning disk array that it is effectively impossible that you have a need for storage of that performance magnitude. If you did, you would be dysfunctional today, right?