EMC 20% Unified Storage Guarantee: Final Reprise

Hi! You might remember me from such blog posts as: EMC 20% Unified Storage Guarantee !EXPOSED! and the informational EMC Unified Storage Capacity Calculator – The Tutorial! – Well, here I’d like to bring to you the final word on this matter! (Well, my final word.. I’m sure well after I’m no longer discussing this… You will be, which is cool, I love you guys and your collaboration!)

Disclaimer: I am in no way saying I am the voice of EMC, nor am I assuming that Mike Richardson is infact the voice of NetApp, but I know we’re both loud, so our voices are heard regardless :)

So on to the meat of the ‘argument’ so to speak (That’d be some kind of vegan meat substitute being that I’m vegan!)

EMC Unified Storage Guarantee

Unified Storage Guarantee - EMC Unified Storage is 20% more efficient. Guaranteed.

I find it’d be useful if I quote the text of the EMC Guarantee, and then as appropriate drill down into each selected section in our comparable review on this subject.

It’s easy to be efficient with EMC.

EMC® unified storage brings efficiency to a whole new level. We’ve even created a capacity calculator so you can configure efficiency results for yourself. You’ll discover that EMC requires 20% less raw capacity to achieve your unified storage needs. This translates to superior storage efficiency when compared to other unified storage arrays—even those utilizing their own documented best practices.

If we’re not more efficient, we’ll match the shortfall

If for some unlikely reason the capacity calculator does not demonstrate that EMC is 20% more efficient, we’ll match the shortfall with additional storage. That’s how confident we are.

The guarantee to end all guarantees

Storage efficiency is one of EMC’s fundamental strengths. Even though our competitors try to match it by altering their systems, turning off options, changing defaults or tweaking configurations—no amount of adjustments can counter the EMC unified storage advantage.

Here’s the nitty-gritty, for you nitty-gritty types
  • The 20% guarantee is for EMC unified storage (file and block—at least 20% of each)
  • It’s based on out-of-the-box best practices
  • There’s no need to compromise availability to achieve efficiency
  • There are no caveats on types of data you must use
  • There’s no need to auto-delete snapshots to get results

This guarantee is based on standard out-of-the-box configurations. Let us show you how to configure your unified storage to get even more efficiency. Try our capacity calculator today.

Okay, now that we have THAT part out of the way.. What does this mean? Why am I stating the obvious (so to speak)  Let’s drill this down to the discussions at hand.

The 20% guarantee is for EMC unified storage (file and block—at least 20% of each)

This is relatively straight-forward.  It simply says “Build a Unified Configuration – which is Unified” SAN is SAN, NAS is NAS, but when you combine them together you get a Unified Configuration! – Not much to read in to that.  Just that you’re likely to see the benefit of 20% or greater in a Unified scenario, than you are in a comparable SAN or NAS only scenario.

It’s based on out-of-the-box best practices

I cannot stress this enough.   Out-Of-Box Best practices.   What does that mean?    Universally, I can build a configuration which will say to this “20% efficiency guarantee” Muhahah! Look what I did! I made this configuration which CLEARLY is less than 20%! Even going into the negative percentile! I AM CHAMPION GIVE ME DISK NOW!".   Absolutely.  I’ve seen it, and heard it touted (Hey, even humor me as I discuss a specific use-case which me and Mike Richardson have recently discussed.)    But building a one-off configuration which makes your numbers appear ‘more right’ v using your company subscribed best practices (and out of box configurations) is what is being proposed here.   If it weren’t for best practices we’d have R0 configurations spread across every workload, with every feature and function under the sun disabled to say ‘look what I can doo!”

So, I feel it is important to put this matter to bed (because so many people have been losing their time and sleep over this debate and consideration)  I will take this liberty to quote from a recent blog post by Mike Richardson – Playing to Lose, Hoping to Win: EMC’s Latest Guarantee (Part 2)    In this article written by Mike he did some –great- analysis.  We’re talking champion.  He went through and used the calculator, built out use-cases and raid groups, really gave it a good and solid run through (which I appreciate!)   He was extremely honest, forthright and open and communicative about his experience, configuration and building this out with the customer in mind.   To tell you the truth, Mike truly inspired me to follow-up with this final reprise.

Reading through Mike’s article I would like to quote (in context) the following from it:

NetApp Usable Capacity in 20+2 breakdown

The configuration I recommend is to the left.  With 450GB FC drives, the maximum drive count you can have in a 32bit aggr is 44.  This divides evenly into 2 raidgroups of 20+2.  I am usually comfortable recommending between 16 and 22 RG size, although NetApp supports FC raidgroup sizes up to 28 disks.  Starting with the same amount of total disks (168 – 3 un-needed spares), the remaining disks are split into 8 RAID DP raidgroups. After subtracting an additional 138GB for the root volumes, the total usable capacity for either NAS or SAN is just under 52TB.

I love that Mike was able to share this image from the Internal NetApp calculator tool (It’s really useful to build out RG configurations) and it gives a great breakdown of disk usage.

For the sake of argument for those who cannot make it out from the picture, what Mike has presented here is a 22 disk RAID-DP RG (20+2 disks – Made up of 168 FC450 disks with 7 spares) I’d also like to note that snapshot reserve has been changed from the default of 20% to 0% in the case of this example.

Being I do not have access to the calculator tool which Mike used, I used my own spreadsheet run calculator which more or less confirms what Mike’s tool is saying to be absolutely true!   But this got me thinking!    (Oh no! Don’t start thinking on me now!)    And I was curious.   Hey, sure this deviates from best practices a bit, right? But BP’s change at times, right?

So being that I rarely like to have opinions of my own, and instead like to base it on historical evidence founded factually and referenced in others… I sent the following txt message to various people I know (Some Former Netappians’s, some close friends who manage large scale enterprise NetApp accounts, etc (etc is for the protection of those I asked ;))

The TXT Message was: “Would you ever create a 20+2 FC RG with netapp?”

That seems pretty straight forward.   Right? Here is a verbatim summation of the responses I received.

  • Sorry, I forgot about this email.  To be brief, NO.
  • “It depends, I know (customer removed) did 28, 16 is the biggest I would do”
  • I would never think to do that… unless it came as a suggestion from NetApp for some perfemance reasons… (I blame txting for typo’s ;))
  • Nope we never use more then 16
  • Well rebuild times would be huge.

So, sure this is a small sampling (of the responses I received) but I notice a resonating pattern there.   The resounding response is a NO.   But wait, what does that have to do with a hole in the wall?   Like Mike said, NetApp can do RG sizes of up to 28 disks.   Also absolutely 100% accurate, and in a small number of use-cases I have found situations in which people have exceeded 16 disk RG’s.   So, I decided to do a little research and see what the community has said on this matter of RG sizes. (This happened out of trying to find a Raid6 RG Rebuild Guide – I failed)

I found a few articles I’d like to reference here:

  • Raid Group size 8, 16, 28?

    • According to the resiliency guide Page 11:

      NetApp recommends using the default RAID group sizes when using RAID-DP.

    • Eugene makes some good points here –

      • All disks in an aggregate are supposed to participate in IO operations.  There is a performance penalty during reconstruction as well as risks; "smaller" RG sizes are meant to minimize both.

      • There is a maximum number of data disks that can contribute space to an aggregate for a 16TB aggregate composed entirely of a give disk size, so I’ve seen RG sizes deviate from the recommended based on that factor (You don’t want/need a RG of 2 data+2parity just to add 2 more data disks to an aggr….). Minimizing losses to parity is not a great solution to any capacity issue.

      • my $0.02.

    • An enterprise account I’m familiar has been using NetApp storage since F300 days and they have tested all types of configurations and have found performance starts to flatline after 16 disks.  I think the most convincing proof that 16 is the sweet spot is the results on spec.org.  NetApp tests using 16 disk RAID groups.

  • Raid group size recommendation

      • Okay, maybe not the best reference considering I was fairly active in the response on the subject in July and August of 2008 in this particular thread.  Though read through it if you like, I guess the best take away I can get from it (which I happened to have said…)
        • I was looking at this from two aspects: Performance, and long-term capacity.
        • My sources for this were a calculator and capacity documents.
        • Hopefully this helped bring some insight into the operation  and my decisions around it.
          • (Just goes to show… I don’t have opinions… only citeable evidence Well, and real world customer experiences as well;))
    • Raid group size with FAS3140 and DS4243
      • I found this in the DS4243 Disk Shelf Technical FAQ document
      • WHAT ARE THE BEST PRACTICES FOR CONFIGURING RAID GROUPS IN FULLY LOADED CONFIGURATIONS?
      • For one shelf: two RAID groups with maximum size 12. (It is possible in this case that customers will configure one big RAID group of size 23–21 data and 2 parity; however, NetApp recommends two RAID groups).
    • Managing performance degradation over time
    • Aggregate size and "overhead" and % free rules of thumb.
    • Why should we not reserve Snap space for SAN volumes?
      • All around good information, conversation and discussion around filling up Aggr’s – No need to drill down to a specific point.

So, what does all of this mean other than the fact that I appear to have too much time on my hands? :)

Well, to sum up what I’m seeing and considering we are in the section titled ‘out of box best practices’

  1. Best Practices and recommendations (as well as expert guidance and general use) seem to dictate a 14+2, 16 disk RG
    1. Can that number be higher.  Yes, but that would serve to be counter to out-of-box best practices, not to mention it seems your performance will not benefit as seen in the comments mentioned above (and the fact that spec.org tests are run in that model)
  2. By default the system will have a reserve, and not set to 0% – so if I were to strip out all of the reserve which is there for a reason – my usable capacity will go up in spades, but I’m not discussing a modified configuration; I’m comparing against a default, out-of-box best practices configuration, which by default calls for a 5% aggr snap reserve, 20% vol snap reserve for NAS and a SAN Fractional Reserve of 100%
    1. Default Snapshot reserve, and TR-3483 helps provide backing information and discussion around this subject. (Friendly modifications from Aaron Delp’s NetApp Setup Cheat Sheet)
  3. In order to maintain these ‘out of box best practices’ and enable for a true model of thin provisioning (albeit, not what I am challenging here, especially being that Mike completely whacked the reserve space for snapshots – Nonetheless… in our guarantee side of the house we have the ‘caveat’ of “There’s no need to auto-delete snapshots to get results” – Which is simply saying, Even if you were to have your default system out of box, in order to achieve, strive and take things to the next level you would need to enable “Volume Auto-Grow” on NetApp, or it’s sister function “Snap Auto Delete” the first of which is nice as it’s not disruptive to your backups, but you can’t grow when you’ve hit your peak! So your snapshots would then be at risk.   Don’t put your snapshots at risk!
  4. Blog posts are not evidence for updating of Best Practices, nor does it change your defaults out of box.   What am I talking about here?  (Hi Dimitris!)   Dimitri wrote this –great- blog post NetApp usable space – beyond the FUD whereby he goes into the depth and discussion of what we’ve been talking about these past weeks, he makes a lot of good points, and even goes so far as to validate a lot of what I’ve said, which I greatly appreciate.    But taking things a little too far, he ‘recommends’ snap reserve 0, fractional reserve 0, snap autodelete on, etc.    As a former NetApp engineer I would strongly recommend a lot of ‘changes’ to the defaults and the best practices as the use-case fit, however I did not set a holistic “Let’s win this capacity battle at the sake of compromising my customers data”   And by blindly doing exactly what he suggested here, you are indeed putting your data integrity and recovery at risk.   

I’ve noticed that.. I actually covered all of the other bullet points in this article without needing to actually drill into them separately.  :) So, allow me to do some summing up on this coverage.

If we compare an EMC RAID6 Configuration to a NetApp RAID-DP Configuration, with file and block (at least 20% of each) using out of box default best practices, you will be able to achieve no compromise availability, no compromise efficiency regardless of data type, with no need to auto-delete your snapshots to gain results.   So that’s a guarantee you can write home about, 20% guaranteed in ‘caveats’ you can fit into a single paragraph (and not a 96 page document ;))

Now, I’m sure, no.. Let me give a 100% guarantee… that someone is going to call ‘foul’ on this whole thing, and this will be the hot-bed post of the week, I completely get it.   But what you the reader really are wondering “Yea, 20% Guarantee.. Guarantee of what? How am I supposed to learn about Unified?”

Welcome to the EMC Unified Storage – Next Generation Efficiency message!

Welcome to the EMC Unisphere – Next Generation Storage Management Simplicity

I mean, obviously once you’re over the whole debate of ‘storage, capacity, performance’ you want to actually be able to pay to play (or, $0 PO to play, right? ;))

But I say.. Why wait?  We’re all intelligent and savvy individuals.  What if I said you could in the comfort of your own home (or lab) start playing with this technology today with little effort on your behalf.     I say, don’t wait.   Go download now and start playing.

For those of you who are familiar with the infamous Celerra VSA as published in Chad’s blog numerous times New Celerra VSA (5.6.48.701) and Updated “SRM4 in a box” guide things have recently gone to a whole new level with the introduction of Nicholas Weaver’s UBER VSA!  Besser UBER : Celerra VSA UBER v2 – Which takes the ‘work’ out of set up.  In fact, all set up requires is an ESX Server, VMware Workstation, VMware Fusion (or in my particular case, I do testing on VMware Viewer to prove you can do it) and BAM! You’re ready to go and you have a Unified array at your disposal!

Celerra VSA UBER Version 2 – Workstation
Celerra VSA UBER Version 2 – OVA (ESX)

Though I wouldn’t stop there, if you’re already talking Unified and playing with File data at all, run don’t walk to download (and play with) the latest FMA Virtual Appliance! Get yer EMC FMA Virtual Appliance here!

Benefits of Automated File Tiering/Active Archiving

But don’t let sillie little Powerpoint slides tell you anything about it, listen to talking heads on youtube instead :)

I won’t include all of the videos here, but I adore the way the presenter in this video says ‘series’ :) – But, deep dive and walk through in FMA in Minutes!

    Okay! Fine! I’ve downloaded the Unified VSA, I’ve checked out FMA and seen how it might help.. but how does this help my storage efficiency message? What are you trying to tell me?  If I leave you with anything at this point, let’s break it down into a few key points.

    • Following best practices will garner you a 20% greater efficiency before you even start to get efficient with technologies like Thin Provisioning, FAST, Fast Cache, FMA, etc
    • With the power of a little bandwidth, you’re able to download fully functional Virtual Appliances to allow you to play with and learn the Unified Storage line today.
    • The power of managing your File Tiering architecture and Archiving policy is at your finger tips with the FMA Virtual Appliance.
    • I apparently have too much time on my hands.  (I actually don’t… but it can certainly look that way :))
    • Talk to your TC, Rep, Partner (whoever) about Unified.   Feel free to reference this blog post if you want, if there is nothing else to learn from this, I want you – the end user to be educated :)
    • I appreciate all of your comments, feedback, positive and negative commentary on the subjectI encourage you to question everything, me, the competition, the FUD and even the facts.   I research first, ask questions, ask questions later and THEN shoot.    The proof is in the pudding.  Or in my case, a unique form of Vegan pudding.

    Good luck out there, I await the maelstrom, the fun, the joy.   Go download some VSA’s, watch some videos, and calculate, calculate, calculate!   Take care! – Christopher :)

    EMC Unified Storage Capacity Calculator – The Tutorial!

    The latest update to this is included here in the Final Reprise! EMC 20% Unified Storage Guarantee: Final Reprise

    After all of the brouhaha and discussion from a recent post EMC 20% Unified Storage Guarantee !EXPOSED! I thought it valuable to dive a little deeper into our own calculator.

    EMC Storage Guarantee

    I’m sure like me, some of you may have tried to use the calculator and found it to be really cool, but you also may have experienced a few bouts of frustration.   It’s okay, I completely get it.  I get it so much that I’m writing this article to help reveal some of the challenges and how to overcome them.

    For starters, one of the coolest bits about the EMC Unified Storage Capacity Calculator… is the fact it has a –help- option right there on screen.  I totally get it if you didn’t notice it, or feel you’re above ‘help’, I’m with you wholly! But I decided ‘Why not.. what does the “?” unveil? Wait for it… it unveils secrets to your success! And a breakdown of the ‘sauce’ so to speak!

    EMC Unified Storage Capacity Calculator

    When you first launch the Capacity Calculator for the first time, you should see a screen which looks like this – It defaults to a NX4 with nothing configured or set up.

    Configuration and Templates

    NAS (Templates) SAN (Custom) Capacity Breakdown

    Regardless of which System Model you choose, NX4, NS-120, NS-480 or NS-960, the “?” help text for the NAS/SAN/Breakdown will be the same across the board.  Exceptions being that the help file will specifically declare whether you’re looking at a specific/respective model.   

    SAN Custom Configuration Not Enough Space - Error! 

     

    As you start to fill the system with disks you may at some point come across an error such as ‘not enough space’, this will usually come up when you’re playing around with SAN configurations, or NAS (Custom) configs.    There is no need to be worried or alarmed when this happens.   All this is saying is that based upon the configuration you have ‘defaulted’ in the column/tab you’re working in, there is not enough “space” in that particular tray to add the disks.  

    Adding Hot Spares to ConfigurationMoving Between Trays

    There are two ways to resolve this.    Either change to a disk format you can work with (Such as HS (Hot Spare)) as seen above, or using the arrows in the System Model diagram, you can move to another tray entirely!

    System Models

    NX4 System Model NS-120 System Model NS-480 System Model NS-960 System Model

    What I find to be particularly useful and cool is the fact that when you select a particular system, hover over the “?” in the System Model section is that it will give you a breakdown of details about the system. (No more needing to go search the internet or call your TC asking “How many drives will my system take!?”  Not only that, but it also provides you with details of how you’d go about building this configuration – both in this simulator so to speak, and respectively when you go live with this as a real configuration.   Sweet if you ask me!

    Total Usable Capacity

    NX4 Total Usage Capacity NS-120 Total Usage Capacity NS-480 Total Usage Capacity NS-960 Total Usage Capacity

    One particularly useful and cool bit about this is it not only tells you specifically what kinds of disks are required of the system, one particular complaint I’ve heard from some folks was about not knowing how many Spares were recommended in their configuration.  Well, check out the ‘caution’ symbol!

    Unrealistic Configuration on an NS-960 Required Hot Spares for an NS-960 Configuration

    I built the following un-realistic configuration so we could drill down in to the system to see what it will report for ‘spares’ required.  Based upon this example, it looks like I need spares of every type, EFC, FC and ATA!    (I populated a tray of each type of disk to make this as unrealistic as possible :))  Pretty cool if you ask me!

    But for the most part, this accounts for all of the ‘errors’, common or otherwise which I am noticing are encountered using this calculator.

    Feel free to give it a good run through, but I’m so glad to see that a majority (read:all) of our concerns of how it operates and functions are actually solved right here in the help file! And in the case of SAN Custom Configuration (read:lack of templates) the little workaround for ‘lack of space’ above seems to address that in whole!

    I hope you find the EMC Unified Storage Capacity Calculator to be as cool as I do, and that you get the best out of it!

    Thanks and for those of you who haven’t played with the Celerra Virtual Appliance yet – Go download the Uber version here! Besser UBER : Celerra VSA UBER v2 (That’ll give you the ability to play around with the Celerra today without having to buy the hardware… nothing spells getting familiar than actually playing with a fully functioning system!)

    Thanks, and good luck!

    The latest update to this is included here in the Final Reprise! EMC 20% Unified Storage Guarantee: Final Reprise

    EMC 20% Unified Storage Guarantee !EXPOSED!

    The latest update to this is included here in the Final Reprise! EMC 20% Unified Storage Guarantee: Final Reprise

    For those of you who know me (and those who don’t, hi! Pleased to meet you!) I spent a lot of time at NetApp battling the storage efficiency game, always trying to justify where all of the storage space went in a capacity bound situation.   However since joining EMC, all I would ever hear from the competition is how ‘space inefficient’ we were and frankly, I’m glad to see the release of the EMC Capacity Calculator to let you decide for yourself where your efficiency goes.   Recently we announced this whole "Unified Storage Guarantee" and to be honest with you, I couldn’t believe what I was hearing. So I decided to take the marketing hype, set it on fire and start drilling down into the details, because that’s the way I roll. :)

    EMC Storage Guarantee

    I decided to generate two workload sets side by side to compare what you get when you use the Calculators

    I have a set of requirements – ~131TB of File/Services data, and 4TB of Highly performing random IO SAN storage

    There is an ‘advisory’ on the EMC guarantee that you have at least 20% SAN and 20% NAS in order to guarantee a 20% space efficiency over others – So I modified my configuration to include at least 20% of both SAN and NAS (But let me tell you, when I had it as just NAS.. It was just as pretty :))

    Using NetApp’s Storage Efficiency Calculator I assigned the following data:

    Storage Efficiency Calculator

    That seems pretty normal, nothing too out of the ordinary – I left all of the defaults otherwise as we all know that ‘cost per TB’ is relative depending upon any number of circumstances!

    So, I click ‘Calculate’ and it generates this (beautiful) web page, check it out! – There is other data at the bottom which is ‘cut off’ due to my resolution, but I guarantee it was nothing more than marketing jibber jabber and contained no technical details.

    Storage Efficiency Calculator

    So, taking a look at that – this is pretty sweet, it gives me a cool little tubular breakdown, tells me that to meet my requirements of 135TB I’ll require 197TB in my NetApp Configuration – that’s really cool, it’s very forthright and forth coming.

    What’s even cooler is there are checkboxes I can uncheck in order to ‘equalize’ things so to speak. And considering that the EMC Guarantee is based upon Useable up front without enabling any features! Let me take this moment to establish some equality for a second.

    Storage Efficiency Calculator

    All I’ve done is uncheck Thin Provisioning (EMC can do that too, but doesn’t require you to do that as part of the Guarantee, because we all know… some times… customers WON’T thin provision certain workloads, so I get it!)   Also turning off deduplication, just so I get a good feel for how many spindles I’ll be eating up from a performance perspective – And turning off dev/test clone (which didn’t really make much difference since I had little DB in this configuration)

    Now, through no effort of my own, the chart updated a little bit to report that NetApp now requires 387TB to manage the same workload a second ago required 197TB. That’s a little odd, but hey, what do I know.. This is just a calculator taking data and presenting it to me!

    Now… with the very same details thrown into the EMC Capacity Calculator, lets take a look at how it looks.

    image

    According to this, I start with a Raw Capacity of ~207TB and through all of the ways as defined on screen, I end up with 135TB Total usable, with at least 20% SAN and about twice that in NAS – Looks fairly interesting, right?

    But lets take things one step further. Let’s scrap Snapshots on both sides of the fence. Throw caution in to the wind.. No snapshots.. What does that do to my capacity requirements for the same ~135TB Usable I was looking for in the original configurations.

    clip_image005[4]I updated this slide to accurately reflect more realistic R5 sets for the EFD disks.  In addition I introduced an ADDITIONAL spare disk, which should 'hurt' my values and make me appear less efficient.

    On the NetApp side I reclaim 27TB of Useable space (to make it 360TB Raw)- while on the EMC side I reclaim 15TB of useable space [150TB Useable now] while Still 207TB Raw.

    But we both know the value of having snapshots in these file-type data scenarios, so we’ll leave the snapshots enabled – and now it’s time to do some math – Help me as I go through this, and pardon any errors.

    Configuration NetApp RAW NetApp Useable Raw v Useable % EMC RAW EMC Useable Raw v Useable % Difference
    FILE+DB                
    Default Checkboxes   197 TB 135 TB 68% 207 TB 135 TB 65% -3%
    Uncheck Thin/Dedup   387 TB 135 TB 35% 207 TB 135 TB 65% +30%
    Uncheck Snaps   369 TB 135 TB 36% 207 TB 150 TB 72% +36%

    However, just because I care (and I wanted to see what happened) I decided to say "Screw the EMC Guarantee" and threw all caution to the wind and decided to compare a pure-play SAN v SAN scenario, just to see how it’d look.

    clip_image007[4]

    I swapped out the numbers to be Database Data, Email/Collaboration Data – The results don’t change (Eng Data seems to have a minor 7TB Difference.. Not sure why that is, – feel free to manipulate the numbers yourself though, it’s negligible)

    clip_image008[4]

    And I got this rocking result! (Yay, right?!) 202TB seems to be my requirement with all the checkboxes checked! But this is Exchange and Sharepoint data (or notes.. I’m not judging what email/collab means ;))… I’m being honest and realistic with myself, so I’m not going to thin provision or Dedup it any way, so how does that change the picture?

    clip_image009[4]

    It looks EXACTLY the same [as before]. Well, that’s cool, at least it is consistent, right?

    However, doing the same thing on the EMC side of the house.

    I want to note a few differences in this configuration – I upgraded to a 480 because I used exclusively 600GB FC drives as I’m not even going to lie to myself that I’m humoring my high IO workloads on 2TB SATA Disks – If you disagree you let me know, but I’m trying to keep it real :)

    RAID5 is good enough with FC disks (If this was SATA I’d be doing best practice and assigning RAID6 as well, so keeping it true and honest) And it looks like this:

    image

    (Side Note: It looks like this SAN Calculation has only 1 hot spare declared instead of the 6 used above in the other configuration – I’m not sure why that is, but I’m not going to consider 5 disks as room for concern so far as my calculations go – it is not reflected in my % charts below – FYI!  I fixed the issue and introduced 6 Spare disks.  I also changed the system from 14+1 R5 sets to 4+1 and 8+1 R5 sets which seems to accurately reflect more production like workloads :))

    Whoa, 200TB Raw Capacity to get me 135TB Usable? Whoa, now wait a second. (says the naysayers) You’re comparing RAID5 to RAID6 – that’s not a fair configuration because there is definitely going to be a discrepancy! And you have snapshots enabled too for this workload. (Side note: I do welcome you to compare RAID6 in this configuration, you’ll be surprised :))

    I absolutely agree – so in the effort of equalization – I’m going to uncheck the Double Disk Failure Protection from the NetApp side (Against best practices, but effectively turning the NetApp configuration into a RAID4 config) and I’ll turn off Snapshot copies to be a fair sport.

    clip_image011[4]image

    There, it’s been done. The difference is.. That EMC RAW Capacity has stayed the same(200TB) while NetApp raw capacity has dropped considerably by 30TB from 387TB to 357TB. (I do like how it reports "Total Storage Savings – 0%" :))

    So, what does all of this mean? Why do you keep taking screen caps, ahh!!

    This gives you the opportunity to sit down, configure what it is YOU want, get a good feel for what configuration feels right to you and be open and honest with yourself and said configuration.

    No matter how I try to swizzle it, I end up with EMC coming front and center on capacity utilization from RAW to Usable – Which down right devastates anything in comparison. I do want to qualify this though.

    The ‘guarantee’ is that you’ll get 20% savings with both SAN and NAS. Apparently if I LIE to my configuration and say ‘Eh, I don’t care about that’ I still get OMG devastatingly positive results of capacity utilization. – So taking the two scenarios I tested in here and reviewing the math..

    Configuration NetApp RAW NetApp Useable Raw v Useable % EMC RAW EMC Useable Raw v Useable % Difference
    FILE+DB                
    Default Checkboxes   197 TB 135 TB 68% 207 TB 135 TB 65% -3%
    Uncheck Thin/Dedup   387 TB 135 TB 35% 207 TB 135 TB 65% +30%
    Uncheck Snaps   369 TB 135 TB 36% 207 TB 150 TB 72% +36%
                     
    EMAIL/Collab                
    Default Checkboxes   202 TB 135 TB 67% 200 TB 135 TB 68% +1%
    Uncheck Thin/Dedup   387 TB 135 TB 35% 200 TB 135 TB 68% +33%
    Uncheck RAID6/Snaps   357 TB 135 TB 38% 200 TB 151 TB 76% +38%

    When we’re discussing apples for apples – We seem to be meeting the guarantee whether NAS, SAN or Unified.

    If we were to take things to another boundary, out the gate I get the capacity I require – If I slap Virtual Provisioning, Compression, FAST Cache, Auto-Tiering, Snapshots and a host of other benefits that the EMC Unified line brings to solve your business challenges… well, to be honest it looks like you’re coming out on top no matter what way you look at it!

    I welcome you to ‘prove me wrong’ based upon my calculations here (I’m not sure how that’s possible because I simply entered data which you can clearly see, and pressed little calculate buttons… so if I’m doing some voodoo, I’d really love to know)

    I also like to try to keep this as realistic as possible and we all know some people like their NAS only or SAN only configurations. The fact that the numbers in the calculations are hitting it out of the ballpark so to speak is absolutely astonishing to me! (Considering where I worked before I joined EMC… well, I’m as surprised as you are!) But I do know the results to be true.

    If you want to discuss these details further, reach out to me directly (christopher.kusek@emc.com) – or talk to your local TC (Or your TC, TC Manager and me in a nicely threaded email ;)) – They understand this rather implicitly.. I’m just a conduit to ensure you folks in the community are aware of what is available to you today!

    Good luck, and if you can find a way to make the calculations look terrible – Let me know… I’m failing to do that so far :)

    !UPDATE! !UPDATE! !UPDATE! :)  I was informed apparently every thing is not as it seems? (Which frankly is a breath of relief, whew!)

    Latest news on the street is, apparently there is a bug in the NetApp Efficiency Capacity Calculator – So after that gets corrected, things should start to look a little more accurate, let me breathe a sigh of relief around that, because apparently (after being heavily slandered for ‘cooking the numbers’) the only inaccuracy going on there [as clearly documented] was in the source of my data.

    However, being that I’m not going to go through and re-write everything I have above again, I wanted to take things down to their roots, lets get down into the dirt, the details, the raw specifics so to speak.  (If any thing in this chart below is somehow misrepresented, inaccurate or incorrect, please advise – This is based upon data I’ve collected over time, so hash it out as you feel :))

    NetApp Capacity GB TB EMC Capacity GB TB GB Diff TB Diff % Diff
                       
    Parity Drives 4000 3.91   Parity Drives 4000 3.91 0 0  
    Hot Spares 1000 0.98   Hot Spares 1000 0.98 0 0  
    Right Sizing 3519 3.44   Right Sizing 1822.7 1.78 1696.3 1.66  
    WAFL Reserve 2045.51 2   CLARiiON OS 247.87 0.24 1797.64 1.76  
    Core Dump Reserve 174.35 0.17   Celerra OS 60 0.06 114.35 0.11  
    Aggr Snap Reserve 863.06 0.84     0 0 863.06 0.84  
    Vol Snap Reserve – 20% 3279.62 3.2   Check/Snap Reserve 20% 3973.89 3.88 -694.27 -0.68  
    Space Reservation 0 0     0 0 0 0  
                       
    Usable Space 13118.5 12.8 Usable Space 16895.54 16.49 -3777.04 -3.69 +23%
    Raw Capacity 28000 27.34 Raw Capacity 28000 27.34 0 0  

    What I’ve done here is take the information and tried to ensure each one of these apples are as SIMILAR as possible.

    So you don’t have to read between the lines either, let me break down this configuration – This assumes 28 SATA 1TB Disks, with 4 PARITY drives and 1 SPARE – in both configurations.

    If you feel that I somehow magically made numbers appear to be or do something that they shouldn’t – Say so.   Use THIS chart here, don’t create your own build-a-config workshop table unless you feel this is absolutely worthless and that you truly need that to be done.   

    You’ll notice that things like Parity Drives and Hot Spares are identical (As they should be) Where we start to enter into discrepancy is around things like WAFL Reserve, Core Dump Reserve and Aggr Snap Reserve – Certainly there are areas of overlap as shown above and equally the same can be said of areas of difference, which is why in those areas on the EMC side I use that space to define the CLARiiON OS and the Celerra OS.    I did have the EMC Match the default NetApp Configuration of a 20% vol snap reserve (on the EMC side I call it Check/Snap Reserve) [Defaults to 10% on EMC, but for the sake of solidarity, what’s 20% amongst friends, right?]    (On a side note, I notice that my WAFL Reserve figures might actually be considerably conservative as a good friend gave me a dump of his WAFL Reserve and the result of his WAFL Reserve was 1% of total v raw compared to my 0.07% calculation I have above, maybe it’s a new thing?)

    So, this is a whole bunch of math.. a whole bunch of jibber jabber even so to speak.   But this is what I get when I look at RAW numbers.   If I am missing some apparent other form of numbers, let it be known, but let’s discuss this holistically.     Both NetApp and EMC offer storage solutions.    NetApp has some –really- cool technology.  I know, I worked there.   EMC ALSO has some really cool technology, some of which NetApp is unable to even replicate or repeat.   But before we get in to cool tech battles, as we sit in a cage match watching PAM duel it out with FAST-Cache, or ‘my thin provisioning is better than yours’ grudge matches.    We have two needs we need to account for.

    Customers have data that they need to protect.   Period.

    Customers have requirements of a certain amount of capacity they expect to get from a certain amount of disks.

    If you look at the chart closely, there are some OMFG ICANTBELIEVEITSNOTWAFL features which NetApp brings to bear, however they come at a cost.   That cost seems to exist in the form of WAFL Reserve, and Right sizing (I’m not sure why the Right Sizing is coming in a considerably fat consideration when contrasted with how EMC does it, but it apparently is?)  So while I can talk all day long about each individual specific feature NetApp has, and equivalent parity which EMC has in that same arena; I need to start somewhere.  And strangely going back to basics, seems to come to a 23% realized space savings in this scenario (Which seems inline with the EMC Unified Storage Guarantee) Which frankly, I find to be really cool.  Because like has been resonated by others commenting on this ‘guarantee’, what the figures appear to be showing is that the EMC Capacity utilization is more efficient even before it starts to get efficient (through enabling technologies).

    Obviously though, for the record I’m apparently riddled with Vendor Bias and have absolutely no idea what I’m talking about! [disclaimer: I have no idea what I’m talking about when I define and disclose I am in this post and others ;)]   However, I’d like to go on record based upon these mathematical calculations, were I not an employee of EMC, and whether I did or did not work for NetApp in the past, I would have come to these same conclusions independently when presented with these same raw figures and numerical metrics.   I continue to welcome your comments, thoughts and considerations when it comes to a Capacity bound debate [Save performance for another day, we can have that battle out right ;)]  Since this IS a Pureplay CAPACITY conversation.

    I hope you found this as informative as I did taking the time to create, generate, and learn from the experience of producing it.  Oh, and I hope you find the (unmoderated) comments enjoyable. :)   I’d love to moderate your comments, but frankly… I’d rather let you and the community handle that on my behalf.   I love you guys, and you too Mike Richardson even if you were being a bit snarky to me. {Hmm, a bit snarky or a byte snarky… Damn binary!}  Take care – And Thank you for making this my most popular blog-post since Mafia Wars and Twitter content! :)

    The latest update to this is included here in the Final Reprise! EMC 20% Unified Storage Guarantee: Final Reprise

    Everything I need to know I learned at EMC World 2010

    I got asked a –great- question the other day by my boss (Props to my boss for giving me some great blogging material!)

    “What are the Top 5 things you learned or thought valuable at EMC World 2010”

    I cannot guarantee I’ll give you 5 top things… or that I won’t even exceed 5 things, but at the very least I’ll outline what I found as the most valuable things, as well as discussions I had with various others as to what was the most valuable to be had/found!   If there are things you found particularly useful – Please do share it here so we can all benefit as well! :0

    Introducing EMC VPLEX - Access Data Anywhere in the Private Cloud

    Yes, this was a big one for me, top of mind! So many useful things to talk solely about VPLEX alone! To share all of it would be a challenge! So let me break down some of the value points!

    EMC has a great portal page around VPLEX which shares lots of useful information, collateral, white papers and more! (I’ll publish a nice one later with PowerLink things soon)

    My good friend Bas Raayman collated all of the blog posts covering VPLEX from the general purpose Blogosphere and Media analysts too!  EMC VPLEX – Introduction and link overview

     

    So to wrap up my own personal experience around the VPLEX from the sessions, prior knowledge and playing with the live VPLEX LAB – This is really cool – Live data access on both sides of a metro link = rockstar event! Check it out! Check it out! :)

    See for yourself how new advanced storage efficiencies and EMC Unisphere can expedite your journey to the private cloud.

    After you pick yourself up after drooling all over VPLEX, it’s time to sign up for the Unisphere train, the Highlander of getting you to the Private Cloud “There can only be one… Private Cloud” and apparently, you’ll manage that with Unisphere!

    You can listen to and watch Chad present a video on using Unisphere to get you kicked off on things! Certainly, if you’re at some of those non-youtube type locations, ask your local EMC to get you a demo :)      Though once you get yourself settled on a single interface for fun and managing off this rocking system, well.. you likely want it to do a little more!

    This video actually takes you through some of the new features discussed at EMC World – Sub-Lun FAST, FAST Cache, and Primary block storage compression!

    1) Sub-LUN FAST = automated tiering of storage (think "VMware DRS for Storage") – lowers cost of storage by between 30-50% on average.
    2) FAST Cache = up to 2TB of low cost read/write cache – get more performance for the same $. In some cases, like VMware View – it can deliver same performance at 10x lower cost.
    3) Primary block storage compression = save between 30- 50% on average on your storage consumption.

    All sweet features which are pretty damn cool if you ask me! (Feel free to ask me.. we could discuss for hours on end ;))

    Another cool feature which was announced at EMC World which most people don’t usually think about, but I’m gonna get you to start thinking about it now!

    Unified Storage Guarantee - EMC Unified Storage is 20% more efficient. Guaranteed. 

    It’s easy to be efficient with EMC.

    EMC® unified storage brings efficiency to a whole new level. We’ve even created a capacity calculator so you can configure efficiency results for yourself. You’ll discover that EMC requires 20% less raw capacity to achieve your unified storage needs. This translates to superior storage efficiency when compared to other unified storage arrays—even those utilizing their own documented best practices.

    If we’re not more efficient, we’ll match the shortfall

    If for some unlikely reason the capacity calculator does not demonstrate that EMC is 20% more efficient, we’ll match the shortfall with additional storage. That’s how confident we are.

    The guarantee to end all guarantees

    Storage efficiency is one of EMC’s fundamental strengths. Even though our competitors try to match it by altering their systems, turning off options, changing defaults or tweaking configurations—no amount of adjustments can counter the EMC unified storage advantage.

    Here’s the nitty-gritty, for you nitty-gritty types
    • The 20% guarantee is for EMC unified storage (file and block—at least 20% of each)
    • It’s based on out-of-the-box best practices
    • There’s no need to compromise availability to achieve efficiency
    • There are no caveats on types of data you must use
    • There’s no need to auto-delete snapshots to get results

    This guarantee is based on standard out-of-the-box configurations. Let us show you how to configure your unified storage to get even more efficiency. Try our capacity calculator today.

    So, hmm.. What does this mean exactly?! Well, if you read above, that gives you a break down of what the deal offers.  But like you, I want specific use-cases which I can relate to!  

    This basically says, by using the EMC Unified Storage Calculator – Without enabling features (like thin provisioning, etc) you know.. ‘magic’ features, out-right the system will save you 20%, and once you start to enable other features you’ll see it grow above and beyond that – Sounds pretty compelling when it comes to the value of your solutions right?!image

    This is actually pretty damn sweet because it lets you use YOUR figures to determine what your space and layout concerns are, instead of generating off some spreadsheet with a bunch of jibber jabber and you go ‘eh, I still don’t believe it’  – But hey, don’t let me be the judge, load up the EMC Unified Storage Calculator and give it a try your self :)

    Oh, did someone say “Hey, Wasn’t there some cool Ionix UIM 2.0 and Vblock things going on?!”

    But we cannot discuss the rocking unified world of Vblock’s and vSphere without at the very least mentioning the Next Generation vCenter Plugins!

    So, while there were a REAL LOT OF COOL THINGS both Announced and able to be seen at this years EMC World 2010 – You’re going to come away from it, whether you attended or missed the opportunity with some questions! So let me help you address these questions in tangible ways you can take it to your TC/SE and Sales Guy so you can make them work for YOU and your Organization.

    Pick up the phone/email – Reach out to your EMC Guys and tell them you want to hear about:

    • EMC VPLEX – And how it’s ease of use and whether the various use cases, Local, Metro and beyond will be a fit for your organization
    • EMC Unisphere – Taking next generation unified up a notch and enabling your datacenter to grow and scale with the Cloud
      • But seriously, the last one is really cool (Check it out, get a briefing with your Unified Storage guys as soon as you can because…)
      • Oh, and don’t forget to ask about FAST Cache! I’ll be covering this fairly deeply soon in some blog posts, including direct use cases, and my deep technical considerations and assaults on engineering I already have in the works to make sure that it works as I architect and not as marketing tells me to :)
    • EMC Unified Storage 20% Guarantee – This isn’t 90 pages of caveats, this is heads up, you will save period.  Ask your TC to help you save! ;)
    • EMC Ionix UIM 2.0 – If you remember all of my coverage on UIM 1.0, 2.0 is actually gonna be the cats meow!
    • EMC vCenter Plugins! – These are pretty sweet, and will help manage your experience!

    I’m sure there are numerous other things I did not cover (You be the one to tell me!) but the ones above I found particularly cool – So go check them out, ask your TC or Sales Rep (Let me know if you do… I’m pushy… :) But I particularly want to hear your experiences and impressions!)

    Enjoy the new world, and share share share! :)

    Every marathon starts and ends with a 1000 mile journey – Boston Marathon wrap-up

    (sorry for the delay getting this out.. I’ve been running non-stop since I got back, figuratively.. but it was important to get this out, so here it is!)

    I had the great honor and privilege to run as part of an amazing team (Team EMC – Run as One) part of an Amazing Organization (EMC) to run on behalf of an excellent cause (Michael Carter Lisnow Respite Center) and let me tell you, I wouldn’t change anything about what I went through in order to get there.

    But I’m getting ahead of myself…  Let’s go back many months and start at the beginning, shall we?

    The day was September 14th, my birthday! Oh, and it also happened to be the first day I started at EMC as a Technology Evangelist in the Enterprise sector of Chicago.     My feet were not even the slightest bit wet from dipping my toe in the water, and then I see an email come out “Apply to be a part of TEAM EMC to run in the Boston Marathon!” (I’m paraphrasing.. I cannot find the original email ;))   I thought.. What do I have to lose? I’m new here, the worst thing that can happen is, by entering into this lottery, filling out an application and an essay; I don’t get accepted – Right, no skin off my feet (that would certainly come later ;)) – So I went and applied! And less than a month after I start, October 9th, 2009 I receive notice that:

    Congratulations!

    You have won an invitational, non-qualifying official number for the 2010 Boston Marathon.

    Wow! Can I say wow here? I’ll try it again.  Wow!!! I did it! I’m part of the team! I’m on board to be part of something bigger than myself at my newly joined organization! Where do I start?!

    At this point, my schedule is wracked filled with so many different things, but I took it upon myself to begin my [unorthodox] training regime as I would need to.   While my training did include some running with my personal trainer, it also included learning to breathe appropriately while running (a challenge of mine) as well as ensuring I was mentally fit for this task before me.

    I began regularly attending Power Yoga classes at my favorite Yoga Studio – Sky Yoga Studio, which would help ensure I was limber, fit, and breathing constantly.   And also took up TaeKwanDo with my son, because it’s good for us, we needed to get our kung-fu on so to speak, and the type of activity you get in there is rather complimentary as training!

    All the while, letting life happen – winter happen… and then… as if no time had passed at all, the fateful week was here!

    Unlike others who had the extreme luxury of a cast of characters to cheer them on while they run the course, my options were a little more limited than that (Though I did receive a wonderful sound off and a hug at the start line by my good friend stv, so that was nice! :))  But before I was ever going to get to this point.. I needed to get to Boston!

    I spent a fair amount of time on the road… more than would be needed, but I was tired and didn’t leave earlier in the morning as I had planned, so this is the challenge we bring upon ourselves!   I tweeted pictures along the way, but more importantly… I regularly tweeted updates of my status, including questions and whatnot to my friends and community on twitter!   The support I received while driving was amazing!   I was given updates of what the ‘appropriate’ speed to drive in areas, what to watch out for, and general ‘be safe’ type of messages.  The marathon drive was off to a great start!   And in the blink of an eye (of a very tired person….) I had arrived in the general Massachusetts area! very late on a Friday night! (I left Thursday evening)

    Unable to make the kick-off festivities on Friday at 3PM, I was ready, able and capable of making it to the event at EMC 176 to meet/greet with all of those we are running on behalf of! and let me tell you, this was a beautiful experience!

    The turnout on a rainy saturday afternoon was almost overwhelming! you can see from the photos above just a small sampling of those who turned out.    The number of people coming up to us, shaking our hands and sincerely expressing their gratitude for what is they’re doing for their loved ones – It’s beyond heart warming, I’m tearing up just typing this, and it’s been over a month!

    The last photo there taken in front of 176, me in front of my car was taken by one of those gracious attendees who expressed how much we are able to do for her daughter who visits the Respite Center.  Beautiful is a word which cannot even grasp the sheer feeling or atmosphere here.    And this video, sums up the extents and effort they went to for us, as we were going to for them.

    So, with all that behind us.. all that is left is to get my bib and get ready!

    The sense of community in boston resonates in so many ways and the support you get – which I helped contribute to when on the day before the marathon I met a lost Vancouverian who had just barely gotten their Bib before the convention center closed.   Using the power of the iPhone and Google Maps, I was able to help her get the pre-race carb-loading dinner on, and through active directions, a way to get her back to her hotel on the T! Yay power of the internets!    But, the time is finally here… the fateful morning would be before us!

    That monday morning I woke up early! (3am – read that to be 2am chicago time!) and started EATING! Yea, that pre-load.. my body knew what it was doing.  And then I headed on the road to Boston to drive in, park and take the busses back to Hopkinton!    It was a BRISK COLD MORNING.  For those of you who know me.. I’m a VERY cold person even in the best of environments.. but even given the circumstances I wasn’t nearly as cold as I should have been for the situation – Yay me? :)    But after the hearty long wait in line at a time I’d never normally be up.. I was boarded onto a bus and headed to Hopkinton and Athletes village. (Did I mention tweeting along the way entirely?   All of the photos I’m using are taken from my Tweet Stream, so the internet KNEW what I was doing at all times :))

    Once in the Athletes village I encounter the great challenge of “Where do I change… why didn’t I change in the first place before showing up” These are definitely the lessons learned you don’t normally hear about considering this is my first marathon!   But I did ensure I was appropriately pre-stretched in the pre-race massage area – And then like a lark I headed down to the corrals ready to race! (thanks stv for taking my iPhone and taking that pre-race shot for me! :))

    And… I ran! Taking my picture by myself in front of as many mile/km markers as I possibly could!   Let me tell you a little bit about the support both on the ground and in the clouds.    I’m sure some of you have run a 5k or other types of events, where it’s often you against yourself running, keeping and making a pace – And that is quite a feat to accomplish.  But a Marathon, and I cannot speak for other marathons.. but this marathon – You cannot ‘practice’ or ‘prepare’ for.    The sheer on the ground support you receive in the form of liquid (water, gatorade) food-stuffs (people giving you oranges, grapes, twizzlers) and even moreso importantly, the morale boost you get from the cheering, the hand slapping, people calling out your name – genuinely caring and cheering for you individually.   You CANNOT (not easily) replicate that in your own practice runs.   This is an enabler making you capable of things you otherwise would feel not possible.

    (I do recall a certain bout/area whereby I was HURTING like you wouldn’t believe! and then a whole row of ~50 people cheered and had their hands up, which sent me into a power-sprint, slapping the hands of every one of them along that run!   If the entire race were lined that way I’m not sure people would ever stop!)

    But when you’re not on the ground and have your head in the clouds like I do, you get support which transcends the mere physical.   I had friends, colleagues who were supporting and cheering me on via Twitter, Facebook, even LinkedIn!   People who were sharing my donation/sponsorship link (I raised an additional $195 DURING the race from their efforts!).   These were good friends who would inform me what the official BAA site was reporting as my current pace time, and offering up tips and guidance as I ran.   So as I said.. both of these channels of support… they cannot be replicated – only experienced and truly to give you that kind of extended feeling of gratification as you cross the finish line.

    I was fortunate to not have suffered some of the grave and life complicating challenges which others had experienced – I simply had a host of blisters (encountered in the first few miles due to getting splashed by water) and not to mention the chaffing… oh that was the worst, that shall be my bane!    But I was fine for the most part.    Shortly after the race, and I finally made it to my car to drive back home, my body temperature started having problems being regulated… and I’d encounter extreme shivers for the next 12-18hours.  But once I had that aside and settled in, albeit my body EXTREMELY sore.    I got back in the car… and headed on my way home! (After all I had work to get back to! :))

    The road wasn’t too bad, I was stressed and strained moving around, but I spent a fair amount of time on cruise control thankfully!  Honestly the worst part of the journey home could be summed up into the experience of the rest stop terror on the right…. Those are not the kind of post-marathon stairs I want to spend my time climbing up/down (I did make it ;))

    In the end, I made it home successfully, safely.. and in recovery for some time, but it was an amazing experience which I would do again and again if I could and have the opportunity.    The thought of would I drive again – Definitely.  I’d prefer to have someone drive with me, so I can go nearly-nonstop instead of continuous stopping and sleeping as I had, but these are the burdens I place upon myself.    For my marathon did not start and finish within 26.2 portions of one mile… No…  My marathon begins and ends with a 1000 mile journey, with a little running in the middle – And it is determination like that which will raise your mental fortitude every time – It’s not about me, nor the race; it is about the children and helping others.

    I thank all of you my supporters – And gladly still take your sponsorship dollars to help those of the Michael Carter Lisnow Respite Center, even still!  Thank you all!