VMware VSAN 6.2 Performance & Storage savings

Just a quick post to share some very interesting performance stats observed on my home lab VSAN cluster (Build details here). The VSAN datastore is in addition to a few NFS datastores also mounted on the same hosts using an external Synology SAN.

I had to build a number of Test VMs, a combination of Microsoft Windows 2012 R2 Datacenter and 2016 TP4 Datacenter VMs on this cluster and I placed all of them on the VSAN datastore to test the performance. See below the storage performance stats during the provisioning (cloning from template) time. Within the Red square are the SSD drive performance stats (where the new VM’s being created) Vs Synology’s NFS mount’s performance  stats (where templates resides) in the Yellow box.

Provisioning Performance

Pretty impressive from all Flash VSAN running on a bunch of white box servers with consumer grade SSD drives (officially unsupported of course but works!), especially relative to the performance of the Synology NFS mounts (RAID1/0 setup for high performance), right??

Imagine what the performance would have been if this was on enterprise grade hardware in your datacentre?

Also caught my eye was the actual inline deduplication and compression savings immediately available on the VSAN datastore after the VM’s were provisioned.

Dedupe & Compression Savings

As you can see, to store 437GB of raw data, with a FTT=1 (where VSAN keeping redundant copies of each vmdk file), its only consuming 156GB of actual storage on the VSAN cluster, saving me 281GB of precious SSD storage capacity. Note that this is WITHOUT Erasure Coding RAID 5 or RAID 6 that’s also available with VSAN 6.2 which, had that been enabled, would have further reduced the actual consumed space more.

The point of this all is the performance and the storage savings available in VSAN, especially all flash VSAN is epic and I’ve seen this in my own environment. In an enterprise datacenter, All Flash VSAN can drastically improve your storage performance but at the same time, significantly cut down on your infrastructure costs for all of your vSphere storage environments. I personally know a number of clients who have achieved such savings in their production environments and each and every day, there seem to be more and more demand from customers for VSAN as their preferred storage / Hyper-Converged technology of choice for all their vSphere use cases.

I would strongly encourage you to have a look at this wonderful technology and realise these technical and business benefits (summary available here) for yourself.

Share your thoughts via comments below or feel free to reach out to discuss what you think via email or social media



VMware Home Lab – New ESXi WhiteBox

I have a VMware Home Lab with 3 ESXi whiteboxes which is kind of the life blood of most of my VMware related studying and new product deployment experience.  As I’m in a presales role, I don’t often get to go out and deploy every single product I talk about in an enterprise IT environment everytime (I do limited deployment work every now and then) and having a decent lab where I can simulate an enterprise IT infrastructure where I can deploy any VMware (or any other part of the SDDC for that matter) is absolutely essential for me to function successfully in my job.  So, due of this necessity, I maintain my own lab in my little garage at home and currently I have 3 ESXi servers as follows

  • Dedicated Management cluster – 1 ESXi whitebox
    • 1 x Intel Xeon E3-1230 4C CPU @ 3.20GHz with HT (8 threads), 32 GB RAM, SuperMicro X8SIL motherboard, 1 x Dual port Intel 1000mbps NIC card

Mgmt Cluster

  • A Compute cluster – 2 ESXi whiteboxes
    • 1 x Intel Xeon X3450 4C CPU @ 2.67GHz with HT (8 threads), 32 GB RAM, Gigabyte Z68AP-D3 motherboard, 1 x Dual port Intel 1000mbps LOM
    • 1 x Intel Core i7 950 4C CPU @ 3.10 GHz with HT (8 threads), 32 GB RAM, Gigabyte motherboard, 1 x Dual port Intel NIC card

Compute Cluster

I also have a Synology 412+ as a shared SAN (iSCSI & NFS – in VMware’s HCL also) and a Cisco 3560G Gigabit switch -L3 enabled for storage and networking

As you can see, my management cluster is jam packed with lots of VM’s some of which have large resource assignments (i.e. vCops VM’s are 8GB each). There’s heavy ballooning happening already and that was even after I’ve had to shut some VM’s down temporarily (I don’t like to keep key VM’s powered down, I’d like them to be running 24×7 to pretty much simulate a real production behaviour).

Mgmt Cluster Memory Util

When I deploy new VMware products, I usually deploy them in a distributed architecture rather than the default all in one server kinda way to simulate a enterprise scalable deployment and this is obviously sucking up lot of resources, mainly memory in my poor 1 server management cluster. Add on top, things like VMware vCAC and NSX demanding dedicated management clusters, its obvious that I need to increase the resources within this cluster ASAP.

So, time to add a new ESXi host to the management cluster….

I’ve been looking around a lot and doing some research in to the options I have available, I’ve considered buying a OEM server (HP and Dell has few SMB servers for reasonable prices) but whats obvious is while the initial server within a minimum spec seems cheap enough, the moment you wanna add some memory to make it useful, the costs jump up massively (nice try HP & Dell…!!)

So, similar to the existing boxes I’ve decided to build another whitebox…. Having done some research, I’ve come up with the following key components as the best fit for a reasonable cost… I will aim to justify why for each item

  • Intel Xeon E5-2620 V2 (Ivy Bridge EP) – Cost around £300
    • Justification: The most important thing this gives me is the thread count. Its a 6c CPU with HT, giving me a total of 12 threads which is great for more VM’s. And its not badly priced compared to other options available. I did look at the Intel core i7’s again (which is also in the VMware HCL) but none (recent models) comes with more than 4 cores which potentially limit my VM density. Core i7 extreme would have been an option but the price and the age ruled that out. Xeon E3 were also limited to 4 cores and the E7’s were astronomically expensive for a Home Lab so no go there. Xeon E5 seemed to be the best option available and E5 version 2 processors seems to strike the best balance between the core count & core speed & the cost. The one I’ve chosen, E5-2620 V2 has the best mix and a very low price. If you are concerned about the power usage, its got one of the lowest TDP (Thermal Design Power) requirement which is 80W.
  • MSI X79A-GD65 (8D) Motherboard – Cost around £160
    • This was a key part of the system and I settled for a single socket motherboard but the most important requirement was that it had the memory scalability up to 64GB (which this does thanks to the 8 DIMM slots). The Intel X79 chipset has built in RAID (AHCI) is great as it enables vSAN support (ESXi 5.5 U2 has AHCI drivers) so I can test that out too.
  • 64GB RAM (non ECC) – Cost around £340
    • This was just easy. I needed as much ram as I could get in for a reasonable cost and non ECC was obviously cheaper. Would have preferred to have got 128GB ram but I felt the cost was a little too high.

I initially spent quite a lot of time looking at the inter-compatibility of the components the hard way by reading the documentation for each components before accidentally coming across the site http://pcpartpicker.com/. This site was brilliant in that you can start of with, say the CPU and it will automatically show you what other components are compatible with it so you can select from a pre-vetted compatible list of components. Once you select an item it also show you various prices for each item from a multiple of online sources that you can directly jump to if you wanna order (I did search for pricing outside of the listing given by the site to double check and most of the ones were cheaper through the sources suggested through this site.

If you wanna see the complete build of my Whitebox, see the link below where it let you publish the full configs and the approximate cost based on the cheapest available online (which was pretty accurate) – All for under £1000 which wasn’t too bad in my case.


Whitebox build & cost

I think this site is awesome and hopefully would help you quite easily build your own whitebox if you need it. There’s a US, UK, Australia and few other countries version of this site available so it uses local suppliers for costs calculations.

Note: I bough my case from elsewhere though as I needed a rack mount server case and I’d bought the same before for my existing servers. That can be found here.

Hopefully this was useful if you are planning to build your own ESXi whitebox.

Also, have a look at the Frank Denneman’s blog about his Home Lab here, I think the motherboard he’s used is pretty awesome with the built in 10gbe ports and memory scalability up to 128GB