I have been working on a slidedeck lately that explains how to build a hyper-converged platform using VMware technology. Of course it is heavily focusing on Virtual SAN as that is one of the core components in the stack. I created the slidedeck based on discussions I have had with various customers and partners who were looking to architect a platform for their datacenters that they could easily repeat. A platform which had a nice form factor and allowed them to scale out and up. Something that could be used in a full size datacenter, but also in smaller SMB type environments or ROBO deployments even.
I guess it makes sense to start with explaining what hyper-converged means to me, although I already wrote various articles on this topic a while back. I explicitly say “to me” as I am sure many folks will have a different opinion on this. A hyper-converged platform is an appliance type of solution where a single box provides a platform for virtual machines. This box typically contains multiple generic x86 hosts (trying to avoid using the word commodity) on which a hypervisor is installed, local storage which is aggregated in to a large shared pool, and network ports. Note that typically network switching is not included in the box itself, well except for virtual switches. In order to aggregate storage in to a large shared pool an additional piece of software is required. Typical examples of hyper-converged platforms which are out there today are Nutanix, SimpliVity and Pivot .
The question than arises if these are “just” x86 boxes with hypervisors installed and storage software, what are the benefits over a regular environment? Those benefits in my opinion are:
- Time to market is short, < 4hrs to install / deploy (probably much faster for the majority)
-
Easy of management and integration
-
Scale out, both capacity and performance wise
-
Typically more affordable (results will vary)
Sounds like a done deal right? Easier, cheaper and faster… It is fair to say that these are great solutions for many companies as they provide you with one throat to choke. With that meaning it is a single SKU offering and which includes a single point of contact for support in most cases. Only downside I have heard from some partners and customers is that these solutions are typically tied to hardware and specific configurations, which is not always the same as your preferred supplier and probably not the exact configuration you prefer. This could lead to operational challenges when it comes to updating / patching, which probably doesn’t make the operational team happy. On top of that there is the “trust” issue. Some people swear by HP and would never ever want to touch any other brand, while others won’t come close to it. That is a matter of experience and personal taste I guess. Where is all of this leading to? Well here is, in my opinion, where Virtual SAN / VSAN comes in.
Virtual SAN / VSAN as available in beta today allows you to pick and choose your own hardware vendor, your own components and size it so that it meets your requirements and also your budget. (Before anyone asks, pricing and packaging of VSAN has not been announced yet.) So where do you start if you want to build a hyper-converged platform for your customers as a partner or as a customer for your own consumption? Well lets start with which VMware products and features one can leverage!
- ESXi 5.5
- vCenter Server 5.5
- vSphere HA
- vSphere DRS + vMotion
- VASA + VM Storage Policies
- Distributed Switch + Network IO Control
- Virtual SAN
That would be the core of your stack from a software point of view, but you can include things like vCenter Orchestrator, VC Ops, vSphere Replication, vSphere Data Protection and so on but lets focus on the core and how to build a solution. I am hoping all of you are familiar by now with the components listed so I am not going to explain them in-depth. I will however briefly touch on why I feel these are recommended. vSphere HA is a no-brainer, availability is key in the majority of environments out there. vSphere DRS and vMotion will be used to ensure that virtual machines will receive the resources they are demanding and will allow us to do things like placing a host in maintenance mode without any disruption to the workload. The Distributed Switch and Network IO Control will provide us an easy to manage virtual switching solution that includes Quality of Service based on shares assigned to traffic streams. This can be useful with Virtual SAN as we want to make sure VSAN receives the resources it requires from a networking perspective. VSAN itself would be needed to aggregate the various local disk resources in to one large pool.
But how does all of that relate to hardware? I posted an article a week ago that explains how to get from an average VM disk size to the number of disks required for VSAN per host. For an in-depth explanation I would like to refer to that. Lets take a random scenario and look at our options.
- 1.5 vCPU on average per VM
- 5 GB of memory per VM
- 50 GB of disk per VM
- Failures to tolerate = 1
We would like to run 100 VMs. This means we require a total of:
- 100 x 1.5 vCPUs = 30 cores
- 500GB of memory
- 11.8 TB of disk space
How did I get there? From a vCPU perspective I want to have 5 vCPUs per core at most. This means that I divide 150 by 5 and the result is a minimum of 30 cores. From a memory perspective the 500GB is straight forward, I am not going to take TPS in to account in this case just to keep the calculation easy. And there is the 11.8 TB disk capacity requirement. How did I get there? I used the equation I developed in my sizing blog post:
(((Number of VMs * Avg VM size) + (Number of VMs * Avg mem size)) * FTT+1) + 10%
In my case that results in: (((100 * 50) + (100 * 5)) * 2) = (5000 + 500) * 2 = 12100 GB + 10% / 1024 = 11.8TB
(I used this VSAN Datastore Calculator)
Now that we know what we roughly need in terms of capacity, how do we tie things together? I will discuss that in Part 2 of this series!
Antone says
I’ve thought about this same model using vsan but the concerns I have is that it’s still in beta and I’m clear when it will be ga or how the licensing will be structured.
Duncan Epping says
I can’t comment on that, but hopefully more news soon.
Locca says
Hi Duncan, Thanks for the post. I saw that virsto has been end of life from 2014. Could you please shed some light on this?
Duncan Epping says
Virsto technology will be integrated in to VSAN.
Totie Bash says
I have the same idea with using OpenIndiana ZFS vm passthrough and Hadoop HDFS but I am still on the researching part. I am not sure if HDFS can handle vm and handle it well. OpenNutanix type of thing. For sure if VSAN licensing is priced good then no need to explore in HDFS.
Duncan if VSAN is already redundant by design and requres min 3 host, do you recommend RAID 0 on the local store vice RAID 10? RAID0 will also benefit on capacity and performance.
Duncan Epping says
VSAN is an object store solution leveraging RAIN/RAID. In other words: VSAN will replicate objects for you. No need for Raid 0 or Raid 10. Just pass the disks to VSAN and VSAN will take care of it.
Totie Bash says
Also interested if they are going to bundle the license for VMware View customer.
Duncan Epping says
Can’t comment on licensing or packaging.
VMrandy says
Great post! Glad to see your thinking about this.
When I was looking at “getting small” which sorta translates to “hyperconverged” I ran into the same question about sizing. My approach went the opposite way, in that instead of looking to hardware for more compute, I would like to see services become less compute intensive. Meaning, create VMs or services with less compute requirements. There is so much waste in the appliances created these days. Disk space can easily be saved by making linked clones more robust within vSphere. I thought this was why VMware bought Visrto actually. If you implement all the VMs and services needed for a vSphere env you quickly start running up the resource price tag. If you created linked clones for all the plugins (VIN, Support, AppHA, VUM and any other 3rd party to take advantage of) you would need far less resources than you do now.
At last count 18 VMs were recommended for vCloud Director for example.
The OS is irrelevant. The services are what matter most.
Hyperconverged VMs would be less to back up, easier to harden, create a known restore point that doesn’t require an OS rebuild.
Much like containers are being used today.
The concept is not new, it’s just not leveraged in the VMware ecosystem like it could be.
Similar arguments exist for CPU and memory. Smaller VMs use less resources giving great flexibility to the user to consume them more efficiently with greater control over how they are consumed.
I did case study with a company called FastScale a few years ago where we looked at resource availability with JeOS’d VMs vs Thick VMs. The results were obvious.
So it’s good to try and shrink things down, but I believe you can attain better results shrinking both hardware and software.
Duncan Epping says
Good point, but completely different discussion 🙂
VMrandy says
I think it fits in nicely. Your initial hardware requirements would be lower, which means less of your own dollars to get started.
They supermicro boxes are nice. Lots if options. They fit the VSAN model very well. Nice choice.
Historically it’s integrators like AMAX who have taken the lead and qualified VMware Ready solutions on Supermicro gear. Would be great to see a partnership with VMware where Supermicro would come to play.