Tag Archives: VMware

Virtual Volumes and Array Snapshots Part I: Managed Snapshots

One of the first technical benefits users can enjoy around VVols is the use of snapshotting. Snapshots created through VMware of VMs have always been a point of contention which as severely limited their usability (see a post I did around the performance impact of them here).

With VVols, when you right-click on a VM and choose take snapshot, VMware does not create the performance-impacting delta VMDK files that were traditionally used, but instead VMware entirely offloads this process to the array. So the array creates the snapshots and VMware just tracks them.

But since VMs are now a collection of individual volumes on the array (a VVol is just an array volume) you can also snapshot and restore individual virtual disks as well directly on the array.

So what does all of this mean?

Continue reading Virtual Volumes and Array Snapshots Part I: Managed Snapshots

Site Recovery Manager and ActiveCluster Part II: Configuring SRM

In my last post, I walked through configuring ActiveCluster and your VMware environment to prepare for use in Site Recovery Manager.

Site Recovery Manager and ActiveCluster Part I: Pre-SRM Configuration

In this post, I will walk through configuring Site Recovery Manager itself.  There are a few pre-requisites at this point:

  • Everything that was done in part 1.
  • Site Recovery Manager installed and paired
  • Inventory mappings in SRM are complete (network, folders, clusters, resource pools etc).
  • Downloaded and installed the FlashArray SRA 3.x or later on both SRM servers.

Continue reading Site Recovery Manager and ActiveCluster Part II: Configuring SRM

What’s New in Core Storage in vSphere 6.7 Part IV: NVMe Controller In-Guest UNMAP Support

vSphere 6.7 core storage “what’s new” series:

Another feature added in vSphere 6.7 is support for a guest being able to issue UNMAP to a virtual disk when presented through the NVMe controller.

Continue reading What’s New in Core Storage in vSphere 6.7 Part IV: NVMe Controller In-Guest UNMAP Support

What is the latency stat QAVG?

I wrote a blog post a year or so ago about ESXi and storage queues which has received a lot of wonderful feedback (thank you!!) and I eventually turned it into a VMworld session and other engagements:

So in the past year I have had quite a few discussions around this. And one part has always bothered me a bit.

In ESXI, there are a variety of latency metrics:

  • GAVG. Guest average. Sometimes called “VM observed latency”. This is the amount of time it takes for an I/O to be completed, after it leaves the VM. So through ESXi, through the SAN (or iSCSI network) and committed to the array and acknowledged back.
  • KAVG. Kernel average. This is how long an I/O is spending in the ESXi kernel. If this is anything but zero, there is some kind of bottleneck (often a maxed out queue)
  • DAVG. This is how long it takes for the I/O to be sent from host, through the SAN and to the array and acknowledged back.

Continue reading What is the latency stat QAVG?

FlashArray Plugin 2.0 for vRealize Orchestrator

We have published the FlashArray plugin 2.0 for vRealize Orchestrator on the VMware Solutions Exchange! Download it here:

https://marketplace.vmware.com/vsx/solutions/pure-storage-flasharray-plugin-for-vmware-vrealize-orchestrator-2-0-0

We put a lot of work into this one and I am quite excited for customers and partners to start using it.

There are three primary enhancements:

  1. New workflows
  2. New actions
  3. New scriptable objects

Continue reading FlashArray Plugin 2.0 for vRealize Orchestrator

VMware Capacity Reporting Part IV: VVol Capacity Reporting

Storage capacity reporting seems like a pretty straight forward topic. How much storage am I using? But when you introduce the concept of multiple levels of thin provisioning AND data reduction into it, all usage is not equal (does it compress well? does it dedupe well? is it zeroes?).

This multi-part series will break it down in the following sections:

  1. VMFS and thin virtual disks
  2. VMFS and thick virtual disks
  3. Thoughts on VMFS Capacity Reporting
  4. VVols and capacity reporting
  5. VVols and UNMAP

Let’s talk about the ins and outs of these in detail, then of course finish it up with why VVols makes this so much better.

NOTE: Examples in this are given from a FlashArray perspective. So mileage may vary depending on the type of array you have. The VMFS and above layer though are the same for all. This is the benefit of VMFS–it abstracts the physical layer. This is also the downside, as I will describe in these posts. Continue reading VMware Capacity Reporting Part IV: VVol Capacity Reporting

VMware Storage Capacity Reporting Part II: VMFS and Thick Virtual Disks

Storage capacity reporting seems like a pretty straight forward topic. How much storage am I using? But when you introduce the concept of multiple levels of thin provisioning AND data reduction into it, all usage is not equal (does it compress well? does it dedupe well? is it zeroes?).

This multi-part series will break it down in the following sections:

  1. VMFS and thin virtual disks
  2. VMFS and thick virtual disks
  3. Thoughts on VMFS Capacity Reporting
  4. VVols and capacity reporting
  5. VVols and UNMAP

Let’s talk about the ins and outs of these in detail, then of course finish it up with why VVols makes this so much better. Continue reading VMware Storage Capacity Reporting Part II: VMFS and Thick Virtual Disks

ActiveCluster and vMSC Implementation Guide

Quick post. I have published my ActiveCluster implementation guide for vSphere Metro Storage Cluster (vMSC). You can find it here:

https://support.purestorage.com/Solutions/VMware_Platform_Guide/ActiveCluster_with_VMware/Implementing_vSphere_Metro_Storage_Cluster_With_ActiveCluster

ActiveCluster, if you are unfamiliar, is the Pure Storage FlashArray implementation of fully Active/Active replication. Meaning a volume can exist and take write simultaneously from two arrays. No additional hardware or licenses required. Comes with Purity 5.0.0. A main focus, like everything else Pure, is simplicity–that goal has definitely been achieved in my view.

Continue reading ActiveCluster and vMSC Implementation Guide

VMware Storage Capacity Reporting Part I: VMFS and Thin Virtual Disks

Storage capacity reporting seems like a pretty straight forward topic. How much storage am I using? But when you introduce the concept of multiple levels of thin provisioning AND data reduction into it, all usage is not equal (does it compress well? does it dedupe well? is it zeroes?).

This multi-part series will break it down in the following sections:

  1. VMFS and thin virtual disks
  2. VMFS and thick virtual disks
  3. Thoughts on VMFS Capacity Reporting
  4. VVols and capacity reporting
  5. VVols and UNMAP

Let’s talk about the ins and outs of these in detail, then of course finish it up with why VVols makes this so much better.

Continue reading VMware Storage Capacity Reporting Part I: VMFS and Thin Virtual Disks

Announcing Pure Storage FlashArray VVol GA

This is a blog post I have been waiting to write for a long time. We at Pure Storage are pleased to announce that vSphere Virtual Volume support on the FlashArray is officially GA!

The FlashArray now supports running VVols in Purity 5.0.0 and later. The cool thing about the FlashArray is the flexibility of the Purity Operating Environment–so VVols are supported on all FA 400 models (405, 420, and 450), //M models (m10, m20, m50, m70) and FlashArray//X. Continue reading Announcing Pure Storage FlashArray VVol GA