I posted a week or so ago about the ESXCLI UNMAP process with vSphere 5.5 on the Pure Storage FlashArray here and came up with the conclusion that larger block counts are highly beneficial to the UNMAP process. So the recommendation was simply use a larger block count than the default to speed up the UNMAP operation, something sufficiently higher than the default of 200 MB. I received a few questions about a more specific recommendation (and had some myself) so I decided to dive into this a little deeper to see if I could provide some guidance that was a little more concrete. In the end a large block count is perfectly fine–if you want to know more details–read on!
One of the many VMware integration pieces that I have been impressed with since I’ve joined Pure Storage was the vSphere Web Client Plugin. While not only being one of the first storage vendor plugins released for the Web Client but also one of the simplest ones that I have used.
One of the main things I have been doing in my first few weeks at Pure Storage (which has been nothing but awesome so far by the way) is going through all of our VMware best practices and integration points. Testing them, seeing how they work or can they be improved etc. The latest thing I looked into was Dead Space Reclamation (which from here on out I will just refer to as UNMAP) with the Pure Storage FlashArray and specifically ESXi 5.5. This is a pretty straight forward process but I did find something interesting that is worth noting.
Ah my first official post during my tenure at Pure and it couldn’t have happened at a better time! Just in time for the Purity 4.0 release which we just announced today. While there are plenty of under-the-cover enhancements I am going to focus on the two biggest parts of the release: new hardware and replication. There are other features such as for example hardware security token locking but I am not going to go into those in this post. So first let’s talk about the advancement in hardware!
This is a topic I have posted about in the past but this time I am going to speak about it with the Pure Storage FlashArray. Anyone familiar with the VMware Native Multipathing Plugin probably knows about the Round Robin “IOPS” value which I will interchangeably also refer to as the IO Operation Limit. This value dictates how often NMP switches paths to the device–after a configured number of I/Os NMP will move to a different path. The default value of this is 1,000 but can be changed to as low as 1. For the highest performance Pure recommends changing this setting to 1 for all devices. The tricky thing is that it has to be done for every device on every host and doing this in a simple way isn’t immediately obvious. But here is the procedure.
Short FYI post here. As some of you know I had my last day at EMC Corporation recently and started at Pure Storage this week. I worked at EMC from 2008 (right out of college) and felt it was time for me to personally and professionally to make a change and try something new and different. A great opportunity at Pure came up and I jumped on it!
Finally procured a pass to EMC World 2014 and I am very excited to attend–this will be my 5th one in a row! Once again EMC World is in Las Vegas–you still have time to sign up–so if you haven’t please do! This post is going to center around session suggestions, I will do another concerning the goings-on at EMC World that are of note closer to the start of the conference. Continue reading EMC World 2014: Some Session Suggestions!
Just wanted to take a moment to congratulate all of the 2014 VMware vExperts! I am honored to be counted amongst their number for the second year in a row. A very nice recognition of being active in the VMware community that comes with a lot of perks at conferences, through technology previews/betas and much more.
As you might have read on my blog a few days ago, EMC released an updated version of the Virtual Storage Integrator tool for vSphere Web Client that supports direct provisioning and some management of VNX and VMAX storage. The previous version supported ViPR-only provisioning. If you didn’t see that post you can check it out here. Inevitably when a product involves cross-application and importantly cross-server integration many customers ask the question about what are the firewall requirements to get this thing to work? Let’s take a look.
Today the long-awaited update to Virtual Storage Integrator for the vSphere Web Client as been released! Six months or so ago EMC released the first iteration of the VSI Web Client (version 6.0) that supported provisioning of storage but only for environments enabled with ViPR. The latest release (version 6.1) now adds support for direct provisioning of storage from a VMAX or VNX array.