Posts Tagged ‘vSAN’

Minisforum MS-A2 – The Ultimate #Homelab Server for VMware vSphere, VVF, and VCF?

Monday, June 30th, 2025

Lately, it feels like every VMware vExpert has been posting photos of their compact lab servers — and I’ll be honest, I was starting to feel left out.

So, I joined the club.

I picked up the new Minisforum MS-A2, and I’ve not looked back. This isn’t just another NUC alternative — it’s a serious powerhouse in a tiny chassis, perfect for VMware enthusiasts building or upgrading their vSphere, VVF, or VCF test environments.

Let’s dig into what makes this little beast a perfect addition to any #homelab setup in 2025.

Hardware Highlights – Not Your Average Mini PC
The MS-A2 isn’t just punching above its weight — it’s redefining what’s possible in a compact lab node.

Key Specs:
CPU: AMD Ryzen™ 9 9955HX – 16 cores / 32 threads of Zen 5 power

Memory: Dual DDR5-5600MHz SODIMM slots – up to 96GB officially, but…

Storage:

3× M.2 PCIe 4.0 slots (22110 supported)

Supports U.2 NVMe – great for enterprise-grade flash

Networking:

Dual 10Gbps SFP+ LAN

Dual 2.5GbE RJ45 ports

Wi-Fi 6E + Bluetooth 5.3 (going to replace this with more NVMe storage!)

Expansion:

Built-in PCIe x16 slot (supports split mode – ideal for GPUs, HBAs, or NICs)

This is homelab gold. It gives you the raw compute of a full rack server, the storage flexibility of a SAN box, and the network fabric of a modern datacenter — all under 2L in size.

How I Configured Mine – still sealed in box as I write – video incoming!
I purchased mine barebones from Amazon, and — as of writing — it’s still sealed in the box. Why? I’m waiting for all the parts to arrive.

Most importantly, I’ll be upgrading it with:
128GB of Crucial DDR5-5600 SODIMMs (2×64GB) — pushing beyond the official spec to see just how much performance this little box can handle.

Once everything’s here, I’ll be unboxing and assembling it live on a future episode of Hancock’s VMware Half Hour. Stay tuned if you want a front-row seat to the full setup, testing, and VMware lab deployment.

Perfect for VMware Labs: vSphere 8/9, VVF, and VCF
Whether you’re testing ESXi on bare metal or running full nested labs, this spec ticks every box.

ESXi Bare Metal Capable
The Ryzen 9 9955HX and AMD chipset boot vSphere 8.0U2 and 9.0 Tech Preview cleanly with minimal tweaks. Use community networking drivers or USB NIC injectors if needed.

VVF / VCF in a Box
If you’re exploring VMware Validated Foundation (VVF) or want a self-contained VCF lab for learning:

16C/32T lets you run nested 3-node ESXi clusters + vCenter + NSX-T comfortably

128GB RAM gives breathing room for resource-heavy components like SDDC Manager

PCIe 4.0 + U.2 = blazing fast vSAN storage

Dual 10Gb SFP+ = NSX-T overlay performance lab-ready

Community Validation – I Was Late to the Party
Fellow vExpert Daniel Krieger was ahead of the curve — writing about the MS-A2 months ago in his excellent blog post here:
sdn-warrior.org/posts/ms-a2

Then vExpert William Lam added his voice to the conversation with a guide to running VMware Cloud Foundation (VCF) on the MS-A2:
williamlam.com/2025/06/vmware-cloud-foundation-vcf-on-minisforum-ms-a2.html

Seeing both of them validate the MS-A2 pushed me over the edge — and I’m glad I jumped in.

Setup Tips (Soon!)
Once the unboxing is done, I’ll share:

BIOS tweaks: SVM, IOMMU, PCIe bifurcation

NIC setup for ESXi USB fling and 10GbE DAC

Storage layout for vSAN and U.2/NVMe configs

Full nested VCF/VVF deployment guide

Considerations
Still not officially VMware HCL — but community-tested

Ryzen platform lacks ECC memory — standard for most mini-PC builds

PCI passthrough needs thoughtful planning for IOMMU groupings

Ideal Use Cases
Nested ESXi, vSAN, vCenter, NSX labs

VVF deployment simulations

VCF lifecycle manager testing

Tanzu Kubernetes Grid

NSX-T Edge simulations on 10GbE

GPU or high-speed NIC via PCIe slot for advanced lab scenarios

Final Thoughts
The Minisforum MS-A2 with Ryzen 9 9955HX is a serious contender for the best compact homelab system of 2025. Whether you’re diving into vSphere 9, experimenting with VVF, or simulating a full VCF environment, this mini server brings serious firepower.

It may still be in the box for now —
—but soon, it’ll be front and center on Hancock’s VMware Half Hour, ready to power the next chapter of my lab.

Join the Conversation
Got an MS-A2 or similar mini-monster? Share your specs, test results, or VMware experience — and tag it:

#VMware #vSphere #VCF #VVF #homelab #MinisforumMSA2 #10GbE #vExpert

Part 56: HOW TO: Manually remove a failed vSAN disk group from a VMware vSphere vSAN cluster using ESXCLI

Monday, November 25th, 2024

In this video presentation which is part of the Hancock’s VMware Half Hour I will show you HOW TO: Manually remove a failed vSAN disk group from a VMware vSphere vSAN cluster using ESXCLI.

 

The VMware vSphere vCenter Server web client has difficulty in performing some vSAN actions, so we need to connect via SSH to the bash shell of the ESXI host to perform this action using the following command

esxcli vsan storage remove -u <VSAN Disk Group UUID>

see here

How to manually remove and recreate a vSAN disk group using esxcli

Part 48. HOW TO: Add a VMware vSphere vSAN license to a VMware vSphere vSAN Cluster

Saturday, October 19th, 2024

In this video presentation which is part of the Hancock’s VMware Half Hour HOW TO Video Series I will show you HOW TO: Add a VMware vSphere vSAN license to a VMware vSphere vSAN Cluster.

The storage devices we are using in this video are the Intel® Optane™ SSD DC P4800X Series 375GB, 2.5in PCIe x4, 3D XPoint™, but this procedure can be use to add any compatible storage devices in ESXi to a vSAN datastore.

This video follows on from the follow video in this series

Part 36: HOW TO: Select an inexpensive HCL Certified 10GBe network interfaces for vSphere ESXi 7.0 and vSphere ESXi 8.0 for VMware vSphere vSAN

Part 37: HOW TO: Change the LBA sector size of storage media to make it compatible with VMware vSphere Hypervisor ESXi 7.0 and ESXi 8.0.

Part 39: HOW TO: Create a VMware vSphere Distributed Switch (VDS) for use with VMware vSphere vSAN for the VMware vSphere vSAN Cluster.

If you are creating a design for VMware vSphere vSAN for a Production environment, please ensure you read the  VMware Cloud Foundation Design Guide 01 JUN 2023 – this should be regarded as The Bible!

References

HOW TO: FIX the Warning System logs on host are stored on non-persistent storage, Move system logs to NFS shared storage.

WHAT’S HAPPENING WITH INTEL OPTANE? – Mr vSAN – Simon Todd

Matt Mancini blog

VMware vSAN 8.0 U1 Express Storage Architecture Deep Dive

VMware vSAN 7.0 U3 Deep Dive Paperback – 5 May 2022

VMware vSphere vSAN Licensing Guide

VMUG Advantage

Part 47: HOW TO: Perform storage performance tests on VMware vSphere vSAN, using the VMware Hyper-converged Infrastructure Benchmark fling (HCIBench)

Saturday, October 19th, 2024

In this video presentation which is part of the Hancock’s VMware Half Hour HOW TO Video Series I will show you HOW TO:  Perform storage performance tests on VMware vSphere vSAN, using the VMware Hyper-converged Infrastructure Benchmark fling (HCIBench).

HCIBench is a storage performance testing automation tool that simplifies and accelerates customer Proof of Concept (POC) performance testing in a consistent and controlled way. VMware vSAN Community Forum provides support for HCIBench.

HCIBench

The storage devices we are using in this video are the Intel® Optane™ SSD DC P4800X Series 375GB, 2.5in PCIe x4, 3D XPoint™, but this procedure can be use to add any compatible storage devices in ESXi to a vSAN datastore.

Benchmarks obtained in this video

FIO Benchmarks

4K/70%Read/100%Random

4K/100%Read100%Random

8K/50%Read/100%Random

256K/100%Write/100%Sequential

This video follows on from the follow video in this series

Part 36: HOW TO: Select an inexpensive HCL Certified 10GBe network interfaces for vSphere ESXi 7.0 and vSphere ESXi 8.0 for VMware vSphere vSAN

Part 37: HOW TO: Change the LBA sector size of storage media to make it compatible with VMware vSphere Hypervisor ESXi 7.0 and ESXi 8.0.

Part 39: HOW TO: Create a VMware vSphere Distributed Switch (VDS) for use with VMware vSphere vSAN for the VMware vSphere vSAN Cluster.

If you are creating a design for VMware vSphere vSAN for a Production environment, please ensure you read the  VMware Cloud Foundation Design Guide 01 JUN 2023 – this should be regarded as The Bible!

References

HOW TO: FIX the Warning System logs on host are stored on non-persistent storage, Move system logs to NFS shared storage.

WHAT’S HAPPENING WITH INTEL OPTANE? – Mr vSAN – Simon Todd

Matt Mancini blog

VMware vSAN 8.0 U1 Express Storage Architecture Deep Dive

VMware vSAN 7.0 U3 Deep Dive Paperback – 5 May 2022

Part 41: HOW TO: Increase the capacity of an existing VMware vSphere vSAN datastore by adding new storage devices

Saturday, October 12th, 2024

In this video presentation which is part of the Hancock’s VMware Half Hour HOW TO Video Series I continue setting up VMware vSphere vSAN.

I will show you HOW TO: Increase the capacity of an existing VMware vSphere vSAN datastore by adding new storage devices.

The storage devices we are using in this  VMware vSphere vSAN datastore are Intel Optane devices.

If you are creating a design for VMware vSphere vSAN for a Production environment, please ensure you read the  VMware Cloud Foundation Design Guide 01 JUN 2023 – this should be regarded as The Bible!

References

HOW TO: FIX the Warning System logs on host are stored on non-persistent storage, Move system logs to NFS shared storage.

WHAT’S HAPPENING WITH INTEL OPTANE? – Mr vSAN – Simon Todd

Matt Mancini blog

VMware vSAN 8.0 U1 Express Storage Architecture Deep Dive

VMware vSAN 7.0 U3 Deep Dive Paperback – 5 May 2022

Part 40: HOW TO: Create a VMware vSphere 7.0.3 vSAN datastore using Intel Optane NVMe storage devices

Saturday, October 12th, 2024

In this video presentation which is part of the Hancock’s VMware Half Hour HOW TO Video Series I will show you HOW TO: Create a VMware vSphere vSAN using Intel Optane NVMe storage devices.

The storage devices we are using in this video are the Intel® Optane™ SSD DC P4800X Series 375GB, 2.5in PCIe x4, 3D XPoint™, but this procedure can be use to add any compatible storage devices in ESXi to a vSAN datastore.

This video follows on from the follow video in this series

Part 36: HOW TO: Select an inexpensive HCL Certified 10GBe network interfaces for vSphere ESXi 7.0 and vSphere ESXi 8.0 for VMware vSphere vSAN

Part 37: HOW TO: Change the LBA sector size of storage media to make it compatible with VMware vSphere Hypervisor ESXi 7.0 and ESXi 8.0.

Part 39: HOW TO: Create a VMware vSphere Distributed Switch (VDS) for use with VMware vSphere vSAN for the VMware vSphere vSAN Cluster.

If you are creating a design for VMware vSphere vSAN for a Production environment, please ensure you read the  VMware Cloud Foundation Design Guide 01 JUN 2023 – this should be regarded as The Bible!

References

HOW TO: FIX the Warning System logs on host are stored on non-persistent storage, Move system logs to NFS shared storage.

WHAT’S HAPPENING WITH INTEL OPTANE? – Mr vSAN – Simon Todd

Matt Mancini blog

VMware vSAN 8.0 U1 Express Storage Architecture Deep Dive

VMware vSAN 7.0 U3 Deep Dive Paperback – 5 May 2022

Part 39: HOW TO: Create a VMware vSphere Distributed Switch (VDS) for use with VMware vSphere vSAN for the VMware vSphere vSAN Cluster.

Saturday, October 12th, 2024

In this video presentation which is part of the Hancock’s VMware Half Hour HOW TO Video Series I will show you HOW TO: Create a VMware vSphere Distributed Switch (VDS) for use with VMware vSphere vSAN for the VMware vSphere vSAN Cluster.

VMware vSphere Distributed Switch (VDS) provides a centralized interface from which you can configure, monitor and administer virtual machine access switching for the entire data center. The VDS provides:

  • Simplified virtual machine network configuration
  • Enhanced network monitoring and troubleshooting capabilities
  • Support for advanced VMware vSphere networking features

As my 10GBe switch in this VMware vSphere Lab has LACP functionality I have decided to demonstrate how we configure the vDS for a LACP LAG. Link Aggregation Control Protocol (LACP) is one elements of an IEEE specification (802.3ad) that provides guidance on the practice of link aggregation for data connections, it’s used on trunks or port channels, to bond two ethernet ports together. It is only supported using a VMware vSphere Distributed Switch (VDS) , it is not supported on a VMware vSphere Standard Switch (VSS).

This video covers the following

  • Creation of the VMware vSphere Distributed Switch (VDS).
  • Creation of Portgroups with vLANs for Management, vMotion and vSAN.
  • Creation of the LACP LAG.
  • Adding vDS to hosts in the vSphere Cluster.
  • Migration of existing VMKernel portgroups from VSS to VDS.
  • Testing the VMware vSphere Distributed Switch (VDS).

If you are creating a design for VMware vSphere vSAN for a Production environment, please ensure you read the  VMware Cloud Foundation Design Guide 01 JUN 2023 – this should be regarded as The Bible!

Part 36: HOW TO: Select an inexpensive HCL Certified 10GBe network interfaces for vSphere ESXi 7.0 for VMware vSphere vSAN

Saturday, October 12th, 2024


In this video presentation which is part of the Hancock’s VMware Half Hour HOW TO Video Series I explore two inexpensive 10Gbe network interfaces suitable for the #homelab for use with VMware vSphere vSAN.

 

  • Dell 0Y40PH Broadcom 57810S Dual Port 10GbE SFP+ Network Card Low Profile Dell P/N: 0Y40PH
  • Dell 0XYT17 Intel X520-DA2 Dual Port 10GB SFP+ NIC with SFP

Full details of the part numbers can be found here on my blog – Inexpensive HCL Certified 10GBe network interfaces for vSphere ESXi 7.0 and vSphere ESXi 8.0 #homelab

The Hidden Truth: How to Use Non-Approved SFP+ Modules on Lenovo Gear!

Monday, September 23rd, 2024

In your #homelab, if you’re experimenting with VMware vSphere vSAN, having one or two 10GbE switches is essential for optimal performance. A cost-effective option is the RackSwitch G8124E, often available on eBay. Originally produced by Blade, this switch was later acquired by IBM and eventually became part of Lenovo’s networking portfolio.

BLADE Rackswitch G8124

If you own a Blade, IBM, or Lenovo RackSwitch G8124 10GbE network switch and encounter the frustrating “UNAPPROVED – SR SFP+ is DISABLED” message when using non-Blade, IBM, or Lenovo SFP+ modules, you’re likely searching for a way to bypass this issue. You’re in the right place!

I have no idea why Lenovo is so precious about this information and keeps it a secret. Officially, Lenovo didn’t want people tinkering with the maintenance prompt for fear of causing damage and unnecessary service calls. In reality, there’s a lot of money to be made in optics/modules, so for most vendors, including Lenovo, that’s likely the main reason they’ve kept this information private.

Personally, I do not understand why Lenovo is so secretive. The switches are currently available as scrap value on eBay, and Lenovo does not receive any money from the sale of these switches. These devices are out of warranty and will not be used in production, and Lenovo hasn’t sold these switches for over 14 years. Other compatible SFP+ modules are available that work with this SFP+ override, but based on my experience, the effort was a waste of time since modern SFP+ modules are not compatible with older switches from 2010!

Additionally, I uploaded photos of Blade SFP+ modules available on eBay for as little as £1 each to the Lenovo Forums, which resolved my issue without needing the SFP+ override mentioned above. However, as a result of trying to help others in the community by suggesting these purchase options, I was permanently banned from the Lenovo Forums! It’s frustrating when the goal is to assist fellow users, yet vendor lock keeps people from finding practical solutions. I’m not sure if they thought I was selling SFP+ modules! I didn’t post any direct links to eBay!


While this method can unlock non-vendor SFP+ modules, it’s crucial to note that your experience may vary depending on the optics used. We do not recommend using dual-rate optics with these older switches, as they may not offer optimal compatibility or performance. Always ensure the optics meet the required specifications and thoroughly test them in a controlled environment before deploying in live networks to avoid potential hardware or network disruptions.

Procedure to Unlock Non-Vendor SFP+ Modules:

For early 8.X code (8.3.X and lower) or older firmware versions (7.X, 6.X, 5.X), the following steps must be done through the serial console port (mini-USB, RS232 signaling). You need to have a Serial RS232 cable with a Mini-USB connector on the end; please note this is NOT USB!

  1. Reload the switch (this will interrupt traffic).
  2. Watch the serial console (9600, N, 8, 1). During the memory test (indicated by dots on the display), repeatedly press “Shift + M” until you see the “=>” prompt.
  3. At the “=>” prompt, input these commands exactly as shown (case-sensitive):
 
setenvsfp Override
saveenv
printenv
boot OR reset OR reload

Command Breakdown:

  • setenv sfp Override: Sets the environment variable “sfp” to “Override”.
  • saveenv: Saves the override setting to NVRAM.
  • printenv: Verifies the variable has been set; confirm you see “sfp=Override”.
  • boot, reset, or reload: Restarts the switch with the new environment variable.

Important Notes:

  • This is primarily for lab use or urgent situations where a valid transceiver is unavailable. In production environments, always use IBM or Lenovo-branded optics to avoid issues.
  • After software upgrades, the variable may reset, so keep these instructions handy to reapply the override as needed.

Thank You Lenovo !

🙂

VMware vExpert Sub Programs

Tuesday, April 23rd, 2024

So I thought it was time to apply for the VMware vExpert Sub Programs, of which there are several. I’ve been a vExpert PRO for several years, which is one of the sub programs, so this year I also applied for the VMware vSphere vSAN and vSphere sub programs.

and I’m happy to report that I’ve been excepted into the VMware vSphere vSAN and vSphere sub programs for 2024, as well as a vExpert PRO 2024.