EMC Unified Storage – Now community sized! Celerra and CLARiiON all grown up!

I recently was given insight into something which was up and coming (and well, clearly out at this point :)) The EMC Unified Storage Community!

What is this community though?  This is the mother load so to speak, YOUR source for Information in the Unified… and it’s available without a login! (Although I encourage a login so you can partake in questions, answers and more!)   So, let’s take this opportunity to take a little tour!

EMC Unified Storage Community

If you look at this massive page with so much to offer… it happens to break it ALL down for you with one key point to start with! – Notice on the right side of the screen..

image Wow, does this take all of the effort out of it! Making it so you can get started quick and easy! here are some details!

Getting Started in the Unified Storage community

VERSION 7  Click to view document history

Created on: Aug 16, 2010 3:51 PM by cornwk – Last Modified:  Sep 1, 2010 10:30 AM by cornwk

Thank you for visiting the Unified Storage community.

We hope you will be an active, participating member. But if you just want to view information only, that’s OK too.

Here are some suggestions for how you can get started in this community:
  1. To make sure that you can fully participate in this community, be sure to Login (or Register if this is your first visit to ECN).
    Click Login/Register at the top of the screen in the area that looks like this:
    https://community.emc.com/servlet/JiveServlet/showImage/102-7531-9-17181/logintext.png
  2. In the Unified Storage community, visit Breaking the Ice and tell us a little bit about yourself.
  3. Scan the current discussions and jump in – post a new discussion of your own or respond to somebody else’s question.
  4. Have fun!
If this is your first venture into ECN, visit the Quick Tour and learn how to put the communities to work for you!

75 Views Tags: midrange_storage , community , getting_started , welcome

What I particularly like about this Community though, is the fact that is it organic… driven by the community FOR the community!   It’s young right now (hey, it just launched!) and it will only continue to mature!   And I particularly want to thank the infamous cornwk @kcornwall for everything she did to make this a reality and to continue driving this forward!

Now, while it may all be relative because content changes regularly… the current threads out there are WOW ON TARGET! To things I see come up in meetings practically every day  – so I wanted to bring specific light and clarity to them!

Every time I sit down with folks and we discuss Celerra, CLARiiON, and the whole of the Unified stack… they say “I want to know more information about….” and the …. is ALL of these threads!

So, without further adieu, here are some of the hot links here in the community to get you going!

Obviously I advise you to defer to the actual community for new content and more! but I thought I’d high-light what are often heavily discussed items where people say “Hey! I want more info on this!” Not that I’m saying anyone in particular (hi mike! :)) Should check these out, but if you find these useful, and/or your name is Mike… Definitely check it out!

Oh, and if you’re looking into downloading the UBER VSA which I’ve referenced in the past … Definitely here is the link to v3.2! Your best friend in virtualization!

Play it again, Sam: Celerra UBER v3.2

Thanks guys, check out the community… grow, and learn, question and learn…. and communitize yourselves!

EMC Unisphere in your pocket! Announcing the UBER VSA 3! (Now with low sodium!)

You heard it here! Time to cut your blood pressure in half! (I apologize for those of you who already have low blood pressure.. this may put you over the top!)

UBERTastic : Celerra UBER VSA v3 – Unisphere -  Be sure to click this link here to get to the download links to pull down the OVA or Workstation version!

EMC Unisphere, now with less sodium!

So, Roxanne… what is new in this version of the VSA? As it appears that I’m practically stealing Nick’s entire post (which I’m cool with… ;))

  • DART is now 6.0.36.4
  • Unisphere management console (rocks!)
  • The Celerra VSA is now 64 bit! This means you can throw RAM at it for bigger setups and it will use it. Over 8GB becomes less beneficial without code changes to simulation services. Future updates will fix this from the Celerra VSA engineering teams.
  • The biggest and most difficult change to construct is that the configuration is now adaptive depending on the virtual machine setup. This version is now intelligent in seeing how many resources you have given it.
  • The new Celerra UBER VSA uses this intelligence to now allow *Thin* mode. If you give the VSA under 2GB of RAM it will automatically size the memory limits, processes, and management interface settings to allow it to run with as low as 1024MB of RAM. You won’t do replication or host a ton of VM’s but you can use this mode to host a few and fully demonstrate/test the new Unisphere interface on even a 2GB laptop.
  • The new VSA also uses this intelligence to automatically allow the configuration of single or dual Data Mover version based on the memory assigned. If you give the VSA more than 4GB of memory you will be given the option to enable an additional Data Mover for use as a standby or load balancing experimentation. This means this single appliance can be a small lightweight NFS unit at 1024MB of RAM or can be a 2 Data Mover powerhouse at 8GB of RAM. All automatically configured on first boot through the wizard.
  • Automatic VMDK/Storage additions have been adjusted for new 64 bit OS. This means this still works. Shutoff the VM, add VMDK(s), turn on and you have more space. Automagic
  • Since automagic is so cool, I have changed the Data Mover Ethernet binding to be automatic also. The VM starts with 1 interface for management and 1 interface for the Data Movers. If you want more for the DM(s), just shutoff the VM, add NIC cards (up to 6 additional), and turn back on. It will automatically bind the Data Mover (yes it works with the 2 DM mode also) to the new interfaces and virtual slots. Just go back into Unisphere and assign away. This allows scale up for the bigger 2 Data Mover 8GB of RAM versions easily.
  • Configuration is now Perl/Bash based instead of just Bash to keep things cleaner and slicker and allow for some coolness later on ;)
  • NTP from the configuration portion of the wizard works correctly. It sets both the Control Station and all Data Movers and enables NTP as a running service. Make sure your NTP server is valid.

    So let’s summarize:

    1. New Unisphere
    2. 64 Bit
    3. Automatic sizing
    4. Thin Mode
    5. Optional 2 Data Mover mode
    6. Automatic Data Mover Ethernet adding (along with fixed Storage [VMDK] adding)
    7. NTP works now

    Wow! That’s a whole lot! Where do I sign up to download?!?  UBERTastic : Celerra UBER VSA v3 – Unisphere – No signup required, just go click and download!  Because Nick has so many other vital details about the differences of THIS Uber VSA compared to Uber VSA’s in the past, I am referring you to his page so you can read the ‘technical’ details and stuff!  So go download the UBER VSA TODAY! (I am downloading it right now, literally.. )

    OMFG IM DOWNLOADING IT TOO

    I look forward to your feedback… and enjoyment of this tool, I know I’ve been waiting for some time for this myself!

    Time is running out; EMC Forum Toronto Canada! Sep 9 – (Keynote by EMC CMO Jeremy Burton)

    Register Now for EMC Forum 2010 - Join us at this complimentary one-day event to connect with your peers and learn about the challenges that organizations from all sizes are facing today. 

    Exhausted from Post-VMworld? Not sure what the status of your Private Cloud has been? Wondering where the Cloud meets the road up in Canada?   Come on down to EMC Forum hitting Toronto on September 9th, 2010!   Register now!

    I thought you might find the actual agenda useful of what tracks happen to be available, so here is the breakdown removing all of the other details like breaks and keynotes:

    Track 1 Track 2 Track 3 Track 4 Track 5
    10:00 – 11:00 Simple and Efficient Storage for IT Production Backup and Recovery Redesign with Deduplication – Why The Move Is On Virtualize Mission Critical Applications for Business Production Secure the Journey to the Private Cloud EMC Ionix and VMware vCenter: Management for Virtual Environments
    11:15 – 12:15 Executing a Private Cloud Strategy: Building on the Key Learnings of EMC & Cisco IT Get More out of Storage with Deduplication Storage Systems Transform Your Desktops With VMware View Securing Storage in a Virtualized Environment – Best Practices and Techniques Infrastructure as a Service: The Future of Management
    1:15 – 2:15 Virtual Storage for Business Production Backup for Virtual Environments – How to Maximize Your Virtualization Investments Accelerate your Virtualization Journey with the Virtual Computing Environment Coalition (VCE) Why Archive? Apply Intelligence to Your Information – Then Take Action Sharepoint, Email and File System Archiving
    2:30 – 3:15 Panel Discussion: Storing in the Cloud Panel Discussion: Backup and Recover Redesign with Deduplication Panel Discussion: Making Virtual Environments More Secure Than Physical Environments   Panel Discussion: Managing in the Cloud

    Now.. you know how I love free stuff.. so here is some additional incentive! :)

    · EMC Forum Facebook Page: http://www.facebook.com/EMCForum

    · Onsite Facebook “Best Pic” photo contest. Attendees are asked to submit a picture from the event to our Facebook account. The best picture wins an iPad at the end of the event. Check out the best pic submissions on Facebook: http://www.facebook.com/EMCForum?v=photos

    So, I definitely encourage you to check this out and be sure to Register now! before it is too late!

    Data Longevity, VMware deduplication change over time, NetApp ASIS deterioration and EMC Guarantee

    Hey guys, the other day I was having a conversation with a friend of mine that went something like this.

    How did this all start you might say?!? Well, contrary to popular belief, I am a STAUNCH NetApp FUD dispeller.  What that means is, if I hear something said about NetApp by a competitor, peer, partner or customer which I feel is incorrect or just sounds interesting; I task it upon myself to prove/disprove it because well frankly… People still hit me up with NetApp questions all the time :) (And I’d like to make sure I’m supplying them with the most accurate and reflective data! – yea that’s it, and it has nothing to do with how much of a geek I am.. :))

    Well, in the defense of the video it didn’t go EXACTLY like that.   Here is a little background on how we got to where that video is today :)   I recently overheard someone say the following:

    What I hear over and over is that dedupe rates when using VMware deteriorate over time

    And my first response was “nuh uh!”, Well, maybe not my FIRST response.. but quickly followed by; “Let me try and get some foundational data”  because you know me… I like to blog about things and as a result collect way too much data to try to validate and understand and effectively say whatever I say accurately :)

    The first thing I did was engage several former NetApp folks who are as agnostic and objective as I am to get their thoughts on the matter (we were on the same page!)Data collection time!  

    For Data Collection… I talked to some good friends of mine regarding how their Dedupe savings have been over time because they were so excited when we first enabled it in the first place (And I was excited for them!)   This is where I learned some… frankly disturbing things (I did talk to numerous guys named Mike interestingly enough, and on the whole all of those who I talked with and their data they shared with me reflected similar findings)

    Disturbing things learned!

    Yea I’ve heard all the jibber jabber before usually touted as FUD that NetApp systems will deteriorate over time in general (whether it be Performance, whether it be Space Savings) etc etc. 

    Well some of the disturbing things learned actually coming from the field on real systems protecting real production data was:

    • Space Savings are GREAT, and will be absolutely amazing in the beginning! 70-90% is common… in the beginning. (Call this the POC and the burn-in period)
    • As that data starts to ‘change’ ever so slightly as you would expect your data to change (not sit static and RO) you’ll see your savings start to decrease, as much as 45% over a year
    • This figure is not NetApp’s fault.  Virtual machines (mainly what we’re discussing here) are not designed to stay uniformly the same no matter what in accordance to 4k blocks, so the very fact that they change is absolutely normal so this loss isn’t a catastrophe, it’s a fact of the longevity of data.
  1. Virtual Machine data which is optimal for deduplication typically amounts to 1-5% of the total storage in the datacenter.   In fact if we want to lie to ourselves or we have a specific use-case, we can pretend that it’s upwards of 10%, but not much more than that.  And this basically accounts for Operating System, Disk Image, blah blah blah – the normal type of data that you would dedupe in the first place.
    • I found that particularly disturbing because after reviewing the data from these numerous environments… I had the impression VMware data would account for much more!   I saw a 50TB SAN only have ~2TB of data residing in Data stores and of that only 23% of it was deduplicating (I was shocked!)
    • I was further shocked that when reviewing the data that over the course of a year on a 60TB SAN, this customer only found 12TB of data they could justify running the dedupe process against and of that they were seeing less than 3TB of ‘duplicate data’ coming in around 18% space savings over that 12TB.    The interesting bit is that the other 48TB of data just continued on un-affected by dedupe.   (Yes, I asked why don’t they try to dedupe it… and they did in the lab and, well it never made it into production)

    At this point, I was even more so concerned.   Concerned whether there was some truth to this whole NetApp starts really high in the beginning (Performance/IO way up there, certain datasets will have amazing dedupe ratios to start) etc. and then starts to drop off considerably over time, while the EMC equivalent system performs consistently the entire time.

    Warning! Warning Will Robinson!

    This is usually where klaxons and red lights would normally go off in my head.    If what my good friends (and customers) are telling me is accurate, it is that not only will my performance degrade just by merely using the system, but my space efficiency will deteriorate over time as well.    Sure we’ll get some deduplication, no doubt about that!  But the long term benefit isn’t any better than compression (as a friend of mine had commented on this whole ordeal)    With the many ways of trying to look at this and understand I discussed it with my friend Scott who had the following analogy and example to cite with this:

    The issue that I’ve seen is this:

    Since a VMDK is a container file, the nature of the data is a little different than a standard file like a word doc for example.

    Normally, if you take a standard windows C: – like on your laptop, every file is stored as 4K blocks.  However, unless the file is exactly divisible by 4K (which is rare), the last block has just a little bit of waste in it.  Doesn’t matter if this is a word doc, a PowerPoint, or a .dll in the \windows\system32 directory, they all have a little bit of waste at the end of that last block.

    When converted to a VMDK file, the files are all smashed together because inside the container file, we don’t have to keep that 4K boundary.  Kind of like sliding a bunch of books together on a book shelf eliminating the wasted space.  Now this is one of the cool things about VMware that makes the virtual disk more space efficient than a physical disk – so this is a good thing.

    So, when you have a VMDK and you clone it – let’s say create 100 copies and then do a block based dedupe – you’ll get a 99% dedupe rate across those virtual disks.  That’s great – initially.  Netapp tends to calculate this “savings” into their proposals and tell customers that require 10TB of storage, that they can just buy 5TB and dedupe and then they’ll have plenty of space.

    What happens is, that after buying ½ the storage they really needed the dedupe rate starts to break down. Here’s why:

    When you start running the VMs and adding things like service packs or patches for example – well that process doesn’t always add files to the end of the vmdk.  It often deletes files from the middle, beginning, end and then  replaces them with other files etc.  What happens then is that the bits shift a little to the left and the right – breaking the block boundaries. Imagine adding and removing books of different sizes from the shelf and making sure there’s no wasted space between them.

    If you did a file per file scan on the virtual disk (Say a windows C: drive), you might have exactly the same data within the vmdk, however since the blocks don’t line up, the block based dedupe which is fixed at 4K sees different data and therefore the dedupe rate breaks down.

    A sliding window technology (like what Avamar does ) would solve this problem, but today ASIS is fixed at 4K. 

    Thoughts?

    If you have particular thoughts about what Scott shared there, feel free to comment and I’ll make sure he reads this as well; but this raises some interesting questions.   

    We’ve covered numerous things in here, and I’ve done everything I can to avoid discussing the guarantees I feel like I’ve talked about to death (linked below) so addressing what we’ve discussed:

    • I’m seeing on average 20% of a customers data which merits deduping and of that I’m seeing anywhere from 10-20% space saved across that 20%
    • Translation: 100TB of data, 20TB is worth deduping reclaiming about 4TB of space in total; thus on this conservative estimate you’d get about 4-5% space saved!
    • Translation: When you have a 20TB data warehouse and you go to dedupe it (You won’t) you’ll see no space gained, with a 100% cost across it.
    • With the EMC Unified Storage Guarantee, that same 20TB data warehouse will be covered by the 20% more efficient guarantee (Well, EVERY data type is covered without caveat)   [It’s almost like it’s a shill, but it really bears repeating because frankly this is earth shattering and worth discussing with your TC or whoever]

    For more great information on EMC’s 20% Unified Storage Guarantee – check out these links (and other articles I’ve written on the subject as well!)

    EMC Unified Storage is 20% more efficient Guaranteed

    I won’t subject you to it, especially because it is over 7 minutes long, but here is a semi funny (my family does NOT find it funny!) video about EMCs Unified Storage Guarantee and making a comparison to NetApp’s Guarantee.   Various comments included in the description of the video – Don’t worry if you never watch it… I won’t hold it against you ;)

    Be safe out there, the data jungle is a vicious one!   If you need any help driving truth out of your EMC or NetApp folks feel free to reach out and I’ll do what I can :)

    SPOILERS!!!

    Read More

    What is the difference between orange and blue fiber cable

    Hey guys! I bet you’re wondering the very same thing a lot of people wonder.  And chances are you did a search (as I had in the past) with the very same query as this title.   And chances are.. you’d end up with a page filled with OMFG irrelevant data, when all you want to know is; WILL ORANGE CABLE DO 10G, WHY ARE ALL MY CABLES BLUE NOW?!’  Etc.. :)

    So, in the effort of full disclosure and thanks to my good friend Stuart Miniman, I’m publishing this post here, containing everything you actually care about!

    Published in this lovely slide share; Converged Networks: FCoE, iSCSI and the Future of Storage Networking is a beautiful slide like this, but the real core and important bit is.

    image

    Orange Fiber Cable DOES 10g, but only up to 82m

    Blue (Aqua) Fiber Cable is ALSO 10g, at lengths of up to 300m

    By the looks of things though, when it comes to the days of 40g or 100g, Orange cable will not exactly be your best friend in those scenarios; a good problem to have at that point.

    So, when you’re going through your datacenter and doing short-haul in a cabinet and saying “doh! we need to rip out all of this orange so we can to 10g”.. Don’t.   If you’re running long haul or to a cabinet pretty far away, sure make sure it’s blue all day long but you will be fine in the short term if you’re doing short runs.

    Hopefully you found this useful (let me know if so) because I’m sick and tired of trying to remember what this is and end up at a million pages telling me EVERY detail but this! :)