With VMFS-6, space reclamation is now an automatic, but asynchronous process. This is great because, well you don’t have to worry about running UNMAP anymore. But since it is asynchronous (and I mean like 12-24 hours later asynchronous) you lose the instant gratification of reclamation.
So you do find yourself wondering, did it actually reclaim anything?
Besides looking at the array and seeing space reclaimed, how can I see from ESXi if my space was reclaimed?
Continue reading Monitoring Automatic VMFS-6 UNMAP in ESXi
EnableBlockDelete is a setting in ESXi that has been around since ESXi 5.0 P3 I believe. It was initially introduced as a way to turn on and off the automatic VMFS UNMAP feature introduced in 5.0 and then eventually canned in 5.0 U1.
The description of the setting back in 5.0 was “Enable VMFS block delete”. The setting was then hidden and made defunct (it did nothing when you turned it off or on) until ESXi 6.0. The description then changed to “Enable VMFS block delete when UNMAP is issued from guest OS”. Continue reading In-Guest UNMAP, EnableBlockDelete and VMFS-6
As you might have noticed vSphere 6.5 Update 1 just came out (7/27/2017) and there are quite a few enhancements and fixes. I will be blogging about these in subsequent posts, but there is one that I wanted to specifically and immediately call out now.
Round Robin and IO Operations Limit of 1 is now default in ESXi for the Pure Storage FlashArray! This means that you no longer need to create a custom SATP rule when provisioning a new host or adding your first FlashArray into an existing environment. Continue reading NMP Multipathing rules for the FlashArray are now default
Another how-to post on iSCSI. Essentially another “for the good of the order post” here. iSCSI is becoming increasingly common, so figured I would put a post together that covers the ins and outs of port binding with standard vSwitches.
For information on distributed switches (which I highly recommend using over standard vSwitches) check out this post here:
Setting up Software iSCSI Multipathing with Distributed vSwitches with the vSphere Web Client
So on to Standard vSwitches. Continue reading Setting up iSCSI Port Binding with Standard vSwitches in the vSphere Web Client
Sorry the title is a bit of a mouthful.
I have written some posts on iSCSI in the past, around setup:
Setting up iSCSI with VMware ESXi and the FlashArray
Configuring iSCSI CHAP in VMware with the FlashArray
Another look at ESXi iSCSI Multipathing (or a Lack Thereof)
These have been on various parts, but primarily the setup around standard vSwitches, which generally, in at least in larger environments, is not the norm. Distributed vSwitches are. I have seen a few posts on doing this with the old C# client, but not the vSphere Web Client. Reference those posts here:
So with the amount of questions I have received on it, it is probably worth putting pen to paper on it. Nothing profound here, basically a walkthrough.
This is of course assuming you are doing port binding. If you are not, then just the standard software iSCSI setup (as described in the above 1st post) is needed.
Continue reading Setting up Software iSCSI Multipathing with Distributed vSwitches with the vSphere Web Client
Quick post. I updated my PowerShell GUI tool I maintain for VMware and FlashArray management and added some new features. This time mainly around protection group management. Download it from my GitHub page here:
Continue reading VMware and FlashArray PowerShell GUI tool v2.7
I posted a few months back about ESXi queue depth limits and how it affects performance. Just recently, Pure Storage announced our upcoming support for vSphere Virtual Volumes. So, this begs the question, what changes with VVols when it comes to queuing? In a certain view, a lot. But conceptually, actually very little. Let’s dig into this a bit more.
Continue reading Queue Depth Limits and VVol Protocol Endpoints
This is a blog post I have been waiting to write for quite some time. I cannot even remember exactly how long ago I saw Satyam Vaghani present on this as a concept at VMworld. Back when the concept of what is now called a protocol endpoint (more on that later) was called an I/O Demultiplexer. A mouthful for sure. Finally it’s time! With pleasure, I’d like to introduce VVols on the FlashArray!
Continue reading Introducing vSphere Virtual Volumes on the FlashArray
So over the past two years or so I have been talking up vRealize Orchestrator quite a bit. And a fair amount of that conversation was based on the eventual usage of vRealize Automation. While I certainly feel vRA is a GREAT use case for vRO, the usefulness of vRO does not in any way require vRA.
A common question I get is, “hey can you add this feature to the official FlashArray Plugin?”. The answer is often “maybe” or “eventually” but sometimes even “no”. The plugin is centered at the satisfying the majority and therefore sometimes does not exactly meet your requirements.
So with these two things in mind, what is the connection? Well, using vRO (which is FREE when you have vCenter) you can easily build your own. Especially when you install the FlashArray vRO plugin.
I see a couple advantages here:
- Start learning vRO. Using default workflows so you don’t have to “code” anything. Then start with some more customization as you become familiar.
- Provide tailored workflows in the vSphere Web Client
- Interface-agnostic workflows. As you move forward and use the HTML-5 interface, or vRA you don’t have to redo your work.
Continue reading Building your own Web Client Plugin with vRO
A customer pinged me the other day and said they could not see a volume on their ESXi host. Running ESXi version 6.5. All of the normal stuff checked out, but the volume was nowhere to be seen. What gives? Well it turned out to be the LUN ID was over 255 and ESXi couldn’t see it. Let me explain.
The TLDR is ESXi does not support LUN IDs above 255 for your average device.
*It’s not actually aliens, it is perfectly normal SCSI you silly man.
Continue reading ESXi and the Missing LUNs: 256 or Higher