For those who weren’t able to attend the VMware Online Technology Forum live, the recordings are available for replay now including the Q&A sessions. If you are interested in what’s new for vSphere, VSAN, NSX, vRealize Operations, View and much much more make sure to tune in. Great way to get an overview and a better understanding of what vSphere has to offer. Click the image below to go to the OTF website, register (if you haven’t yet) and get access to a lot of free cool content!
When I joined VMware and had read a white paper on memory reclamation techniques a dozen times. I was left with a bunch of questions still and I emailed the engineer who authored it back in the days. I asked him a couple of “simple” questions and received a one pager email full with answers. Even the email I had to read twice. Not because it is insanely complex, but because there was so much information in there that it was impossible to digest at all. Carl Waldspurger was that engineer. I’d seen some of his talks when he was still at VMware but he has gone “dark” for a while.
Carl joined CloudPhysics in the early stages of the company. He has been working on various projects, and one of those projects is called SHARDS. I had not seen the result yet, and a couple of weeks ago I watched the presentation. Excellent presentation skills, but more importantly amazing research with a very important result. Some people may have been wondering what you can do with a platform like CloudPhysics and what you can harvast from the data, well I think it is fair to say that this is one of the results of all the hard data mining work that has been done over the last years. Here is the abstract with a link to the online presentation. I didn’t want to share everything here to drive some traffic to Usenix as support. Before you watch the video, a warning…. this isn’t a high level overview, serious deep dive.
Reuse-distance analysis is a powerful technique for characterizing temporal locality of workloads, often visualized with miss ratio curves (MRCs). Unfortunately, even the most efficient exact implementations are too heavyweight for practical online use in production systems.
We introduce a new approximation algorithm that employs uniform randomized spatial sampling, implemented by tracking references to representative locations selected dynamically based on their hash values. A further refinement runs in constant space by lowering the sampling rate adaptively. Our approach, called SHARDS (Spatially HashedApproximate Reuse Distance Sampling), drastically reduces the space and time requirements of reuse-distance analysis, making continuous, online MRC generation practical to embed into production firmware or system software. SHARDS also enables the analysis of long traces that, due to memory constraints, were resistant to such analysis in the past.
We evaluate SHARDS using trace data collected from a commercial I/O caching analytics service. MRCs generated for more than a hundred traces demonstrate high accuracy with very low resource usage. MRCs constructed in a bounded 1 MB footprint, with effective sampling rates significantly lower than 1%, exhibit approximate miss ratio errors averaging less than 0.01. For large traces, this configuration reduces memory usage by a factor of up to 10,800 and run time by a factor of up to 204.
You can find the slide/paper and video below as a download.
I found out there is another way to fix your non compliant host profile problems with vSphere 6.0 when you have SAS drives which are detected as shared storage while they are not. This method is a bit more complicated though and there is a command line script that you will need to use: /bin/sharedStorageHostProfile.sh. It works as follows:
- Run the following to dump all your local details in a folder on your first host
/bin/sharedStorageHostProfile.sh local /folder/youcreated1/
- Run the following to dump all your local details in a folder for your second host, you can do this on your first host if you have SSH enabled
/bin/sharedStorageHostProfile.sh remote /folder/youcreated2/ <name or ip of remote host>
- Copy the outcome of the second host to folder where the outcome of your first host is stored. You will need to copy the file “remote-shared-profile.txt”.
- Now you can compare the outcomes by running:
/bin/sharedStorageHostProfile.sh compare /folder/youcreated1/
- After comparing you can run the configuration as follows:
/bin/sharedStorageHostProfile.sh configure /folder/youcreated1/
- Now the disks which are listed as cluster wide resources but are not shared between the hosts will be configured as non-shared resources. If you want to check what will be changed before running the command you can simply do a “more” of the file the info is stored in:
esxcli storage core device setconfig -d naa.600508b1001c2ee9a6446e708105054b --shared-clusterwide=false
esxcli storage core device setconfig -d naa.600508b1001c3ea7838c0436dbe6d7a2 --shared-clusterwide=false
You may wonder by now if there isn’t an easier way, well yes there is. You can do all of the above by running the following simple command. I preferred to go over the steps so at least you know what is happening.
/bin/sharedStorageHostProfile.sh automatic <name-or-ip-of-remote-host>
After you have done this (first method or second method) you can now create your host profile of your first host. Although the other methods I described in the post of yesterday are a bit simpler, I figured I would share this as well as you never know when it may come in handy!
A couple of years ago I wrote an article titled “Host Profile noncompliant when using local SAS drives with vSphere 5?” I was informed by one of our developers that we actually solved this problem in vSphere 6. It is not something I had see yet so I figured I would look at what we did to prevent this from happening and it appears there are two ways to solve it. In 5.x we would solve it by disabling the whole tree, which is kind of a nasty workaround if you ask me. In 6.0 we fixed it in a far better way.
When you create a new host profile and edit is you now have some extra options. One of those options being able to tell if a disk is a shared cluster resource or not. By disabling this for your local SAS drives you avoid the scenario where your host profile shows up as noncompliant on each of your hosts.
There is another way of solving this. You can use “esxcli” to mark your devices correctly and then create the host profile. (SSH in to the host.)
First list all devices using the following command, I took a screenshot of my outcome but yours will look slightly different of course.
esxcli storage core device list
Now that you know your naa identifier for the device you can make the change by issueing the following command and setting “Is Shared Clusterwide” to false:
esxcli storage core device setconfig -d naa.1234 --shared-clusterwide=false
Now you can create the host profile. Hopefully you will find the cool little enhancement in esxcli and host profiles useful, I certainly do!
It has been a year since the last version. This weekend Rob de Vey finally released RVTools 3.7. With over 400.000 downloads so far, this is for sure the most used free health check tool that exists. And there is a good reason for it, it is most definitely also one of the bests tools that can provide valuable insights. I’ve personally been using it since the very first version and have discovered a bunch of potential problems at customers sites. Make sure to follow RVTools on twitter for any updates.
Version 3.7 (March, 2015)
- VI SDK reference changed from 5.0 to 5.5
- Extended the timeout value from 10 to 20 minutes for really big environments
- New field VM Folder on vCPU, vMemory, vDisk, vPartition, vNetwork, vFloppy, vCD, vSnapshot and vTools tabpages
- On vDisk tabpage new Storage IO Allocation Information
- On vHost tabpage new fields: service tag (serial #) and OEM specific string
- On vNic tabpage new field: Name of (distributed) virtual switch
- On vMultipath tabpage added multipath info for path 5, 6, 7 and 8
- On vHealth tabpage new health check: Multipath operational state
- On vHealth tabpage new health check: Virtual machine consolidation needed check
- On vInfo tabpage new fields: boot options, firmware and Scheduled Hardware Upgrade Info
- On statusbar last refresh date time stamp
- On vHealth tabpage: Search datastore errors are now visible as health messages
- You can now export the csv files separately from the command line interface (just like the xls export)
- You can now set a auto refresh data interval in the preferences dialog box
- All datetime columns are now formatted as yyyy/mm/dd hh:mm:ss
- The export dir / filenames now have a formated datetime stamp yyyy-mm-dd_hh:mm:ss
- Bug fix: on dvPort tabpage not all networks are displayed
- Overall improved debug information