• Skip to primary navigation
  • Skip to main content
  • Skip to primary sidebar

Yellow Bricks

by Duncan Epping

  • Home
  • Unexplored Territory Podcast
  • HA Deepdive
  • ESXTOP
  • Stickers/Shirts
  • Privacy Policy
  • About
  • Show Search
Hide Search

Startup News Flash part 10

Duncan Epping · Nov 29, 2013 ·

There we are, part 10 of the Startup News Flash. Someone asked me on Twitter last week why Company XYZ was never included in the news flash. Let it be clear that I am not leaving anyone out (unless I feel they aren’t relevant to this newsletter or my audience), I have limited time so typically do not do briefings… Which means that if the marketing team doesn’t sent me the details via email and I haven’t somehow stumbled across the announcement it will not appear on here. If you want your company to be listed, make sure they sent their press releases over.

Some new models announced by Nutanix. Funny to see how they’ve been pushing hard from a marketing perspective to remove the “pure VDI play” label they had and now launch a VDI focused model called the 7000 series. (Do not get me wrong, I think this is a brilliant move!) The 7000 series offers you the option to include NVIDIA K1 or K2 Grid cards. Primarily intended to accelerate graphics, so if you are for instance doing a lot of 3D rendering or just are a heavy graphical VDI user these could really provide a benefit over their (and other vendors) normal offerings. On top of that the 3000 and 6000 series has been overhauled. The NX-3061 and NX-3061 with 10 Core (2.8GHz) Ivy Bridge have been introduced and the NX6060 and NX6080 10 Core (2.8 and 3.0GHz respectively) have been introduced. Haven’t seen anything around pricing, so can’t comment on that.

No clue what it is exactly these guys do to be honest. I find their teaser video very intriguing. Not much detail to be found around what they are doing other than “re-imagine enteprise computing”. Hoping to hear more from these guys in the future as their teaser did make me curious.

I don’t care much about benchmarks, but it is always nice to see a smaller (or the underdog) company beat the big players. Kaminario managed to outperform Oracle, IBM and Fujitsu with their SPC-2 Performance Benchmark using their scale-out all flash array K2 v4. Just a couple of weeks after breaking the SPC-1 Benchmark World Record again. Like I said, I don’t care much about benchmarks  as it doesn’t typically say much about the operational efficiency etc. Still it is a nice indication of what can be achieved, though your results may vary depending on your IO pattern of course.

VSAN VDI Benchmarking and Beta refresh!

Duncan Epping · Nov 26, 2013 ·

I was reading this blog post on VSAN VDI Benchmarking today on Vroom, the VMware Performance blog. You see a lot of people doing synthetic tests (max iops with sequential reads) on all sorts of storage devices, but lately more and more vendors are doing these more “real world performance tests”. While reading this article about VDI benchmarking, and I suggest you check out all parts (part 1, part 2, part 3), there was one thing that stood out to me and that was the comparison between VSAN and an All Flash Array.

The following quotes show the strength of VSAN if you ask me:

we see that VSAN can consolidate 677 heavy users (VDImark) for 7-node and 767 heavy users for 8-node cluster. When compared to the all flash array, we don’t see more than 5% difference in the user consolidation.

Believe me when I say that 5% is not a lot. If you are actively looking at various solutions, I would highly recommend to include the “overhead costs” to your criteria list as depending on the solution chosen this could make a substantial difference. I have seen other solutions requiring a lot more resources. But what about response time, cause that is where the typical All Flash Array shines… ultra low latency, how about VSAN?

Similar to the user consolidation, the response time of Group-A operations in VSAN is similar to what we saw with the all flash array.

Both very interesting results if you ask me. Especially the < 5% in user consolidation is what stood out to me the most! Once again, for more details on these tests read the VDI Benchmarking blog part 1, part 2, part 3!

Beta Refresh

For those who are testing VSAN, there is a BETA refresh available as of today. This release has a fix for the AHCI driver issue… and it increases the disk group limit from 6 to 7. From a disk group perspective this will  come in handy as many servers have 8, 16 or 24 disk slots allowing you to do 7HHDs + 1 SSD per group. Also some additional RVC commands have been added in the storage policy space, I am sure they will come in handy!

Nice side affect of the number of HDDs going up is increase in max capacity:

(8 hosts * (5 diskgroups * 7 HDDs)) * Size of HDD = Total capacity

With 2 TB disks this would result in:

(8 * (5 * 7)) * 2TB = 560TB

Now keep on testing with VSAN and don’t forget to report feedback through the community forums or your VMware rep.

Virtual SAN and maintenance windows…

Duncan Epping · Nov 25, 2013 ·

After writing the article that “4 is the minimum number of hosts for VSAN” I received a lot of questions via email and on twitter etc about the cost associated with it and if this was a must. Let me start with saying that I wrote this article to get people thinking about Sizing their VSAN environment. When it comes to it, Virtual SAN and maintenance windows can be a difficult topic.

I guess there are a couple of things to consider here. Even in a regular storage environment you typically do upgrades in a rolling fashion meaning that if you have two controllers one will be upgraded while they other handles IO. In that case you are also at risk. The thing is though, as a virtualization administrator you have a bit more flexibility, and you expect certain features to work as expected like for instance vSphere HA. You need to ask yourself what is the level of risk I am willing to take, the level of risk I can take?

When it comes to placing a host in to Maintenance Mode, from a VSAN point of view you will need to ask yourself:

  • Do I want to move data from one host to another to maintain availability levels?
  • Do I just want to ensure data accessibility and take the risk of potential downtime during maintenance?

I guess there is something to say for either. When you move data from one node to another, to maintain availability levels, your “maintenance window” could be stretched extremely long. As you would potentially be copying TBs over the network from host to host it could take hours to complete. If your ESXi upgrade including a host reboot takes about 20 minutes, is it acceptable to wait for hours for the data to be migrated? Or do you take the risk, inform your users about the potential downtime, and as such do the maintenance with a higher risk but complete it in minutes rather than hours? After those 20 minutes VSAN would sync up again automatically, so no data loss etc.

It is impossible for me to give you advice on this one to be honest, I would highly recommend to also sit down with your storage team. Look at what their current procedures are today, what they have included in their SLA to the business (if there is one), and how they handle upgrades / periodic maintenance.

 

Startup News Flash part 9

Duncan Epping · Nov 18, 2013 ·

There we are, part 9 of the Startup News Flash. As mentioned last time, last week was “Storage Field Day” so typically a bit more news then normally this time of year. I would highly recommend to watch the videos. Especially the Coho video is very entertaining.

The original founders of Fusion IO (David Flynn and Rick White) just received 50 Million in funding for their new startup called Primary Data. I mentioned them briefly in Startup News Flash part 3 when they announced they started a new company and it seems that they have something on their hands! They haven’t revealed what they are working on, they are aiming to come out stealth around the second quarter of 2014. In the WSJ the following was mentioned in terms of the space these guys will be playing in: “The company is developing software–though it actually will come bundled on standard server hardware–that essentially connects all those pools of data together, offering what Flynn calls a “unified file directory namespace” visible to all servers in company computer rooms–as well as those “in the cloud” that might be operated by external service companies.” Indeed, something with storage / caching / software defined / scale-out…

I guess scale-out hypervisor based storage solutions are hot… Maxta just officially announced their new product called MxSP. Some rumors had already been floating around but now the details are out their. Marcel v/d Berg did a nice article on them which I recommend reading if you like to get some more details. Basically Maxta created a Virtual Storage Appliance which pools all local storage and presents it as NFS to your hypervisor. Today VMware vSphere is fully supported and KVM / Hyper-V in a limited fashion. It offers functionality like VM-level snapshots and zero-copy clones, Thin provisioning, Inline deduplication and more. It looks like licensing is capacity based but no prices have been mentioned.

When first looking at Avere is was intrigued by their solution but somehow it didn’t really click. Primary focus was a caching layer in between your NFS storage and your hosts… But I wondered why I would want an extra box for that and not just use something host local. Last week Avere made announcement around a solution that allows you to pool local and cloud storage resources and present them via a common namespace and move data between these tiers. FlashCloud is what Avere calls it. Their paper describes it best, so a shameless copy of that: “FlashCloud software running on Avere FXT Edge filers addresses this challenge by storing cold data on cost-effective cloud storage at the core of the network and automatically and efficiently moving active data to the edge near the users.” I like the concept… If you are interested, check out their site here.

Far from a startup, but cool enough to be listed here… The release of the X-Brick aka XtremIO by EMC. The XtremIO solution is a brand new all-flash array which delivers screaming performance in a scale-out fashion. Although there are limitations from a scaling point of view today, it is expected that these will be lifted soon. One of the articles I enjoyed reading is this one by Jason Nash. What is most interesting about the product is the following, and I am going to quote Jason here as he is spot on: “There is no setup and tuning of XtremIO.  No LUNs.  No RAID Groups.  No pools.  No stripe sizes.  No tiering.  Nothing.  You have a pool of very fast storage.  How big do you want that LUN to be?  That’s all you really need to do!”

Another round of funding for SimpliVity, Series C… 58 Million led by Kleiner Perkins Growth Fund and DFJ Growth with contributions by Meritech Capital Partners and Swisscom Ventures. I guess this GigaOM quote says it all: “CEO Doron Kempel, an EMC veteran, said the cash infusion will enable the company to execute on plans to triple its staff and boost sales growth five fold in 2014”.

VMworld session on vSphere Metro Storage Cluster on youtube!

Duncan Epping · Nov 16, 2013 ·

I didn’t even realize this, but just found out that the session Lee Dilworth and I did at VMworld on the subject of vSphere Metro Storage Clusters can actually be viewed for free on youtube!

There are some more sessions up on youtube, so make sure you have a look around!

  • « Go to Previous Page
  • Page 1
  • Interim pages omitted …
  • Page 162
  • Page 163
  • Page 164
  • Page 165
  • Page 166
  • Interim pages omitted …
  • Page 492
  • Go to Next Page »

Primary Sidebar

About the Author

Duncan Epping is a Chief Technologist and Distinguished Engineering Architect at Broadcom. Besides writing on Yellow-Bricks, Duncan is the co-author of the vSAN Deep Dive and the vSphere Clustering Deep Dive book series. Duncan is also the host of the Unexplored Territory Podcast.

Follow Us

  • X
  • Spotify
  • RSS Feed
  • LinkedIn

Recommended Book(s)

Advertisements




Copyright Yellow-Bricks.com © 2025 · Log in