• Skip to primary navigation
  • Skip to main content
  • Skip to primary sidebar

Yellow Bricks

by Duncan Epping

  • Home
  • Unexplored Territory Podcast
  • HA Deepdive
  • ESXTOP
  • Stickers/Shirts
  • Privacy Policy
  • About
  • Show Search
Hide Search

vSAN

#VMworld #STO1378 – vSAN Management Today and in the Future

Duncan Epping · Aug 30, 2017 ·

#STO1378 “vSAN Management Today and in the Future” by Junchi Zhang and Christian Dickmann. Christian did this session by himself last year, and it was one of the coolest sessions of last year, so the expectations are set high. In this session Junchi and Christian will be discussing what vSAN management looks like today, but more importantly they will demo what it will look like in the future.

Christian started with the basics. As Christian stated: when we started with vSAN we developed a storage solution, but we evolved to HCI, which means deep integration with vSphere and hardware. As Christian stated: it is our mission to provide Radically simple HCI with choice. This means building an “appliance-ike” experience on your choice of hardware. Deep integration with vSphere and vSphere workflows. All of this for environments with 2 hosts or 1000s of hosts.

In the past releases vSAN provided a lot of new functionality to realize our mission. Just as an example: Health Checks in 6.0, Capacity Reporting in 6.2, Performance Analytics in 6.6. Going forward however the focus will be: Easier Liefecycle Management, Rich Monitoring, Infrastructure Analytics.

First demo up: HTML5 vSphere Client Demo. In the demo the HTML5 interface is shown, and what stood out to me is the responsiveness. As Christian said, we kept what worked well, but we also have new ways of presenting info. For instance the Health Check view has changed. But also the Virtual Object view has changed, we will be showing swap and for instance iSCSI object. On top of that it will provide filters to make it easier to find what you are looking for. Biggest change in the H5 UI is the SPBM interface. New screens, new layout, simpler, fast and more intuitive!

Up next: Patching ESXi / Firmware. In 6.6 we launched the capability to upgrade disk controllers using Config Assist. But what about VUM? And will it be able to not just do a disk controller but also the SAS Expander, BIOS, NICs etc? And that is what the next demo is about. First of all, they removed 1 required reboot. So down from 2 to just 1! Which will easily save 10-15 minutes per host. The demo next shows the current version of the iDRAC firmware, and the version of vSphere used. Leveraging VUM a new baseline is created which not only contains the latest version of vSphere but also a baseline for firmware. Which in this demo updates the iDRAC firmware and the bios. VUM places the host in maintenance mode, upgrades the firmware and BIOS, then updates the hosts itself and reboots the host and the host is done. As Christian said: One click for you as an admin, but many different steps happening in the back-end. The next thing the team will be looking at is not just doing per host updates, but remediating full clusters at a time. This will include full vSAN Health integrations, which will help validating the hosts if they are “safe / healthy” to run vSAN/Workloads again. In other words: update firmware>> update software >> validate health >> bring back in to cluster.

Next demo was around “HCI Cluster Creation and Extension”. How can we make the creation of clusters easier. Those who have seen the VxRail demoes know how simple things can be. But what if have a vSAN Ready Node? In this demo a simplified UI is shown. During the creation of the demo it is shown how for instance all hosts are added to a cluster at the same time. Before the hosts are added though the hosts are validated/verified. Yes, from a vSAN Health point of view. And when they are safe to add, they are added in to the cluster in “maintenance mode”. This is something I have personally been asking for for a long time now. Great to see this is finally added. Also the health check before adding it to the cluster is great, no point in adding a host which is not healthy. Not just adding hosts, but also configure the network as part of the cluster configuration and apply all the vSAN best practices are applied (yes you can change the settings as you go), not just vSAN but also HA/DRS/NTP/Security settings etc. Great demo, great work!

“Rich Infra Analytics” is up next. Christian explained that we already have some cool analytics tools in vSAN 6.6 and vSAN 6.6.1 with the Cloud Connected Health Check and the Performance Analytics tool. This demo is showing an internal tool. So this is not showing what the customers will be using, but showing how engineers and GSS can leverage internal tooling to make the support experience better! In the demo it is shown how GSS can instantly view your environment and details leveraging the “Phone Home” data. More or less replicating the H5 UI. It will allow support to validate if you set advanced settings, if health checks have failed, if you have contention etc. Not just that, also go deep on performance of your environment. They can even go back in time, everything worked fine  day ago? Did anything change? What did the environment look like? Simply by going by that “point in time”. Not only when you phone us up, but we can also provide pro-active support. We have the data, we can contact you when something is about to fail, or configuration settings could cause issues in the future. (Depending on the bought support level of course) In order for this to work though, CEIP will need to be enabled.

Up next: Junchi talking about VROps Dashboards and what we are doing in this space. As it stands right now, there are different interfaces to monitor vSAN: VROps and vCenter. But can we collapse these when VROps is installed and provide all the info you need under the H5 Client? We have a great number of dashboards in VROps and in the future that data will automatically be exposes in the H5 Client when you have VROps installed.

What about hardware topology? What if you need to swap a disk or add a disk. In the demo it is shown how the H5 will show you the hardware topology in a physical overview. You add a disk and the UI will instantly show you the new disk in the chassis. That was a very “simple” change, but very welcome!

Great session, highly recommended to watch the recording when it becomes available. I will add a link when it is published (DONE!) Thanks Junchi and Christian for showing what we are working on and what the future of vSAN/HCI looks like!

#VMworld #STO1794BU – Evolution of vSAN

Duncan Epping · Aug 30, 2017 ·

First session of the day at 08:30: “The Evolution of vSAN”. Vijay Ramachandran and Christos Karamanolis will be talking where vSAN came from and where it will be going towards in the upcoming years. Vijay started out laying the foundation, talking about what HCI is, what the benefits are and where it fits in.

I think Vijay was spot on, HCI simplifies operating storage / infrastructure as the various layers have converged and the single management pane simplifies your life. Especially when using an appliance model. I liked how Vijay explained the difference between vSAN and competitive solutions. The big difference is the fact that for competitors the storage piece is a separate entity(virtual appliance), with vSAN your storage stack comes as part of vSphere.

Next Vijay discussed all the different use cases and customer types. He showed some new survey data, and it showed that 65% of our customers use vSAN for business critical apps (Oracle, SQL, Exchange, SAP etc). With the two other big use cases being Horizon View/Citrix Xen Desktop but also for Management Clusters and DMZ. Some of the newer use cases we are starting to see are NoSQL solutions.

Next Performance was discussed, shocking how vSAN 6.6 bumped performance with 60% even when all data services are enabled. 60% is a lot higher than I expected, especially as this wasn’t just an increase in IOPS but 6.6 also brought a decrease in latency! Before Vijay handed it over, he mentioned the latest vSAN Customer Count: 10k!

Up now Christos, this will be the forward looking part of the session. Christos kicks off talking about three key tends. First, the new storage devices out in the market: NVMe and NVDIMM. These faster devices in most cases will require a re-architecture or at a minimum think about how you leverage these devices. NVMe devices are becoming more affordable and provide extremely low latency and high IOPS, and non-volatile memory even tops that. How will this change the architecture for vSAN? Secondly, the growth of public cloud, is there a way HCI can leverage that? And thirdly, cloud native apps. How do these new application architectures and delivery methods change the storage requirements?

Christos briefly discussed a potential future change for vSAN which is the convergence of the caching tier and the capacity tier… Providing essentially a single tier of storage devices in your vSAN environment, where vSAN will automatically tier data between devices. You could imagine meta-data residing on NVDIMM or NVMe while “cold data” resides on the “lower tiers”. No longer should it be needed to define a caching layer and a capacity layer. Which should result in better performance and better capacity utilization.

Next point being discussed is the operational aspect. Functionality introduced recently have made the life of customers easier, things like vSAN and VUM, S.M.A.R.T support, PowerCLI cmlets, vRA / VROps and LogInsight advanced support and integration. Next Christos demoed one of the latest addition, the Performance Diagnostic feature that is part of vSAN 6.6.1. If you haven’t seen it yet, watch my demo on youtube.

Next Christos discusses how vSAN is used in AWS and how hybridity is achieved leveraging Hyper-Converged architectures. Enterprise grade storage is needed in AWS to provide the ability to use DRS, HA and other essential vSphere functionality. This is where vSAN came in to play. This provides you a very similar experience in the cloud as on-premises. But of course there’s a big benefit of leveraging VMware Cloud on AWS. In the future you can expect an extremely easy scalable storage stack. Compute already scales in a simple manner, but what if you do not need extra compute and just want to add storage, this will be possible soon. Not only will you be able to add storage, but you will also be able to connect to all native AWS Services. You can imagine that a great use case would be “Disaster Recover”. What you can expect from VMware in the near future is a DR as a Service solution, allowing you to failover from your local datacenter in to VMware Cloud on AWS. This offering would be leveraging SRM and vSphere Replication. But on top of that a new snapshotting mechanism is developed, which will allow even for a more flexible architecture. Not just the ability to replicate to vSAN, NFS or something like data domain. But also store your data in S3 in AWS, and when a recovery needs to occur, without having an environment pre-provisioned in VMware Cloud on AWS, trigger a fail-over when it needs to occur and have a VMware Cloud on AWS infrastructure being spun out and your VMs being recovered from S3 in to that environment.

Last topic of the session was Cloud Native Apps. Christos discussed Project Hatchway. Project Hatchway provides the Docker Volume Service and the vSphere Cloud Provider for Kubernetes. On top of that vSAN will be optimized for “share-nothing” applications. As mentioned in my summary of STO1490, these enhancements will allow you to co-locate data and compute when you have “FTT=0” configured even. Also, sequential I/O will be improved in future releases, as in many cases these share-nothing applications are highly sequential workloads.

Great session, not sure it was recorded, but if it was… watch it online when it is released!

vSAN 6.6.1 Performance Diagnostics Demo

Duncan Epping · Aug 14, 2017 ·

I was playing in my lab this morning and figured I would record a demo of a new feature which is part of vSAN 6.6.1. The Performance Diagnostics feature is aimed to help those running benchmarks to optimize their benchmarks, or optimize their vSAN configuration to reach their expected goals. Note that it is a “cloud connected” feature, and in order to use this you need to participate in the Customer Experience Improvement Program. This however is enabled by default in the latest releases of vSphere. This means that data is send up to the VMware cloud, anonymous, then analyzed and the results are send back to the Web Client. Anyway, enough said, just watch the demo.

Unbalanced Stretched vSAN Cluster

Duncan Epping · Jul 30, 2017 ·

I had this question a couple of times the past month so I figured I would write a quick post. The question that was asked is: Can I have an unbalanced stretched cluster? In other words: Can I have 6 hosts in Site A and 4 hosts in site B when using vSAN Stretched Cluster functionality? Some may refer to this as asymmetric or uneven. Either way, the number of hosts in the two locations differ.

Is this supported? In short: Yes.

The longer answer: Yes you can do this, this is fully supported but you will need to keep your selected PFTT (Primary Failures To Tolerate), SFTT (Secondary Failures To Tolerate) and FTM (Failure Tolerance Method) in to account. If PFTT=1 and SFTT=2 and FTM=Erasure Coding (RAID5/6) then the minimum number of hosts per site is 6. You could however have 10 hosts in Site A while having 6 hosts in Site B.

Some may wonder why anyone would want this, well you can imagine you are running workloads which do not need to be recovered in the case of a full site failure. If that is the case you could stick these to 1 site. (You can set PFTT=0 and SFTT=1, which would result in this scenario for that particular VM.)

That is the power of Policy Based Management and vSAN, extremely flexibility and on a per VM/VMDK basis!

vSphere 6.5 U1 is out… and it comes with vSAN 6.6.1

Duncan Epping · Jul 28, 2017 ·

vSphere 6.5 U1 was released last night. It has some cool new functionality in there as part of vSAN 6.6.1 (I can’t wait for vSAN 6.6.6 to ship ;-)). There are a lot of fixes of course in 6.6.1 and U1, but as stated, there’s also new functionality:

  • VMware vSphere Update Manager (VUM) integration
  • Storage Device Serviceability enhancement
  • Performance Diagnostics in vSAN

So those using vSAN 6.2 who upgraded to vSphere 6.0 U3, here’s your chance to upgrade to get all vSAN 6.5 functionality, and more!

The VUM integration is pretty cool if you ask me. First of all, when there’s a new release the Health Check will call it out from now on. And on top of that, when you go to VUM then also things like async drivers will be taken in to consideration. Where you would normally have to slipstream drivers in to the image and make that image available through VUM, we now ensure that the image used is vSAN Ready! In other words, as of vSphere 6.5 U1 Update Manager is fully aware of vSAN and integrated as well. We are working hard to bring all vSAN Ready Node vendors on-board. (With Dell, Supermicro, Fujitsu and Lenovo leading the pack.)

Then there’s this feature called “Storage Device Serviceability enhancement”. Well this is the ability to blink the LEDs on specific devices. As far as I know, in this release we added support for HP Gen 9 controllers.

And last but not least: Performance Diagnostics in vSAN. I really like this feature. Note that this is all about analyzing benchmarks. It is not (yet?) about analyzing steady state. So in this case you run your benchmark, preferably using HCIBench, and then analyze it by selecting a specific goal. Performance will be analyzed using “cloud analytics”, and at the end you will get various recommendations and/or explanations for the results you’ve witnessed. These will point back to KBs, which in certain cases will give you hints how to solve your (if there is) bottleneck.

Note that in order to use this functionality you need to join CEIP (Customer Experience Improvement Program), which means that you will upload info in to VMware. But this by itself is very valuable as it allows our developers to solve bugs / user experience issues and getting a better understanding of how you use vSAN. I spoke with Christian Dickmann on this topic yesterday as he tweeted the below, and he was really excited, he said he had various fixes going in the next vSAN release based on the current data set. So join the program!

A huge THANK YOU to anyone participating in Customer Experience Improvement Program (CEIP)! Eye-opening data that we are acting on.

— Christian Dickmann (@cdickmann) July 27, 2017

For those how can’t wait, here’s the release notes and downloads links:

  • vCenter Server 6.5 u1 release notes: https://docs.vmware.com/en/VMware-vSphere/6.5/rn/vsphere-vcenter-server-651-release-notes.html
  • vCenter Server 6.5 u1 downloads: https://my.vmware.com/web/vmware/details?downloadGroup=VC65U1&productId=614&rPId=17343
  • ESXi 6.5 release notes: https://docs.vmware.com/en/VMware-vSphere/6.5/rn/vsphere-esxi-651-release-notes.html
  • ESXi 6.5 download: https://my.vmware.com/web/vmware/details?downloadGroup=ESXI65U1&productId=614&rPId=17342
  • vSAN 6.6.1 release notes: https://docs.vmware.com/en/VMware-vSphere/6.5/rn/vmware-vsan-661-release-notes.html

Oh and before I forget, there’s new functionality in the H5 Client for vCenter in 6.5 U1. As mentioned on this vSphere blog: “Virtual Distributed Switch (VDS) management, datastore management, and host configuration are areas that have seen a big increase in functionality”. And then some of the “max config” items had a big jump. 50k powered on VMs for instance is huge:

  • Maximum vCenter Servers per vSphere Domain: 15 (increased from 10)
  • Maximum ESXi Hosts per vSphere Domain: 5000 (increased from 4000)
  • Maximum Powered On VMs per vSphere Domain: 50,000 (increased from 30,000)
  • Maximum Registered VMs per vSphere Domain: 70,000 (increased from 50,000)

That was it for now!

  • « Go to Previous Page
  • Page 1
  • Interim pages omitted …
  • Page 32
  • Page 33
  • Page 34
  • Page 35
  • Page 36
  • Interim pages omitted …
  • Page 71
  • Go to Next Page »

Primary Sidebar

About the Author

Duncan Epping is a Chief Technologist and Distinguished Engineering Architect at Broadcom. Besides writing on Yellow-Bricks, Duncan is the co-author of the vSAN Deep Dive and the vSphere Clustering Deep Dive book series. Duncan is also the host of the Unexplored Territory Podcast.

Follow Us

  • X
  • Spotify
  • RSS Feed
  • LinkedIn

Recommended Book(s)

Also visit!

For the Dutch-speaking audience, make sure to visit RunNerd.nl to follow my running adventure, read shoe/gear/race reviews, and more!

Do you like Hardcore-Punk music? Follow my Spotify Playlist!

Do you like 80s music? I got you covered!

Copyright Yellow-Bricks.com © 2026 · Log in