Start a Conversation

This post is more than 5 years old

Solved!

Go to Solution

721

June 2nd, 2010 22:00

AVM Storage Pool Selection, Performance?

Hello All,

Been researching in the forums and have read the AVM PDFs and have yet to find a definitive answer on some of these quesitons, so any assistance would be great before I engage EMC support directly. I hope this discussion will help all.

We are about to deploy our NS-480 for the VMWare over NFS use case. Our primary concern is performance. From what I understand, AVM is the way to go, however, our storage pool choice is causing us some consternation.

For instance, lets assume that I have about 80 15K 300GB SAS drives in my backend Clariion. I have options of many pools, but for now lets narrow it down to just clar_r5_performance and clar_r5_economy.

If I create a clar_r5_performance using 4 disk groups of 4+1 RAID5, then it is my understanding that, when I add a filesystem, AVM will look for available space in two of those disk groups, add a slice on each (balanced across SPs if posible), and then stripe them together. As a result, that filesystem gets the performance of 10 spindles. While this should be sufficient, it does have concern for some of our higher performance workloads, or if we need to scale a filesystem beyond those spindles. What is the best way to handle this? Can I move a filesystem to a new storage group? Can I restripe the filesystem over more disk groups?

Is there any way AVM would create a filesystem and stripe across all the disk groups in the storage pool? Say 4 or 8 disk groups in the clar_r5_performance pool. I understand that we incur a greater risk as if any RAID group backing that filesystem suffers a double-drive failure, we lose the data, but we have contingencies for that.

Also, does clar_r5_economy provide a significant IOPS benefit over clar_r5_performance? It's my understanding that each individual filesystem would get 18 spindles instead of 10  so I would get more IOPS there for my more demanding applications. This is of course assuming two filesystems didn't reside on the same spindles and shared  them of course, we plan to use is_greedy to help avoid this. I could also use a RAID10 storage pool at this point too if necessary, but this is discouraged as these applications are likely to require high capacity but low write performance, mostly high read performance is required and the expense of RAID10 is not really necessary when RAID5 is sufficient if I can get the spindle count adequate to support the load.

For these applications, read performance matters more than extreme data protection (as we are replicating and several other methods to protect). If this is the case then, would clar_r5_economy provide better performance than clar_r5_performance or is there some special aspect of AVM that I am missing?

Also, am I worrying too much? Is there some magic beyond write caching, etc. that makes AVM performance scale linearly as spindles are added, or do I just make special exceptions for those virtual machines that require more than 10 spindles of performance (e.g. a RAID10 pool or manual filesystems created on large metavolume or RAID10 groups)?

I just want to make sure I'm not missing something here and fully understand the strengths and limitations of AVM, and avoid problems down the road if possible. I really don't anticipate any individual VM needing more than 18 spindles of IOPS performance, but I would like to get it as optimized as possible in case one does

Thanks, I know that was long and rambling so if you can answer even a subset, I would appreciate it.

366 Posts

June 3rd, 2010 14:00

Hi,

Your understanding is almost correct, except for the amount of raid groups AVM will try to use for each filesystem it creates.

By default, the AVM ( on both clar_r5_performance and clar_r5_economy ) will first try to use 4 dvols from different RG's balanced on the SP's. That gives you 20 spindles on 4+1, and 36 spindles on 8+1.

It was found that striping over more than 4 dvols most of the times makes no difference due to the I/O queues on the Data Mover.

AVM should be used on most of the environments.

Also, 4+1 is preferred when you have several disks. If you use 8+1, you would end up with a combination of 4+1 and 8+1, except if you leave empty spaces on the DAE's.

In your VMWare NFS environment, this probably will not be an issue, but I think this exercise is valid before goes to the implementation phase.

Gustavo Barreto.

3 Posts

June 3rd, 2010 14:00

Thanks Gustavo, after re-reading the paper I noticed there that it would use four RAID groups instead of two, I was mislead by the diagram and didn't fully read the selection algorithm information.

OK, that sounds good, I think we have a clear picture of our layout. I feel pretty comfortable with 4+1 if it spreads across four RAID groups, and we'll be sure to SP and bus-balance the DAE's. I would prefer 4+1 as well for the reasons you stated (clean DAE shelves with 3 groups each, and easier expansion, adding in drive groups of 5 instead of drive groups of 8).

One further question: If I were to build a RAID10 group for AVM, can I place the mirrored drives on separate DAEs which reside on separate buses such so that, if I were to lose an entire DAE (say via accidental fiber disconnection), the data would still be online and available? I don't anticipate this as a current use case, but there is a specific application where this may be required, so I was just curious.

Thanks.

366 Posts

June 4th, 2010 05:00

Hi,

Yes, you can do it.

This is not an usual configuration, but the Celerra does not care for the disks placement on the backend. It only will check for the raid type, protection and amount of disks on the raid group.

Your concern makes sense, but to be honest I have never saw a situation like you described.

Gustavo Barreto.

9 Legend

 • 

20.4K Posts

June 5th, 2010 06:00

No Events found!

Top