Posts Tagged ‘Compute’

Minisforum MS-A2 HOW TO: Fix the Failed to Update CPU#0 Microcode PSOD with ESXi 8.0.3i ESXi 8.0u3i

Wednesday, March 11th, 2026

Minisforum MS-A2 HOW TO: Fix the Failed to Update CPU#0 Microcode PSOD with ESXi 8.0.3i ESXi 8.0u3i

 

HOW TO: Fix “Failed to update CPU#0 microcode” PSOD on ESXi 8.0.3i (Minisforum MS-A2)

In this guide from Hancock’s VMware Half Hour, I demonstrate how to fix the
Purple Screen of Death (PSOD) error:

“Failed to update CPU#0 microcode”

This issue occurs when installing or booting VMware ESXi 8.0 Update 3i (ESXi 8.0.3i)
on the Minisforum MS-A2 mini workstation.

ESXi 8.0.3i, released on 2 March 2026, includes security fixes for
CVE-2025-15467, an OpenSSL vulnerability. If you are running ESXi and have not
patched yet, you should update as soon as possible.

However, this update also includes new AMD CPU microcode updates, which currently
appear to trigger a PSOD during boot on the Minisforum MS-A2 platform.


The Problem

When booting the ESXi 8.0.3i installer (for example from Ventoy) on the
Minisforum MS-A2, the system may fail during boot with the following PSOD message:

The system has found a problem on your machine and cannot continue.

Failed to update CPU#0 microcode

This prevents ESXi from completing the boot process or installer launch.


Why This Happens

The ESXi 8.0.3i update includes newer AMD microcode updates intended to improve
security and stability. Unfortunately, these updates currently appear to be
incompatible with the MS-A2 platform, which results in the microcode update failing
during boot.

When the microcode update fails, ESXi halts the boot process and displays the PSOD.


The Workaround

Until VMware releases a permanent fix, the issue can be worked around by using a
kernel boot option during ESXi startup.

Steps to Fix

  1. Boot the ESXi 8.0.3i installer.
  2. When the ESXi boot screen appears, press Shift + O.
  3. This opens the ESXi boot options.
  4. Add the required kernel option shown in the video.
  5. Press Enter to continue booting.

With the boot option applied, ESXi should boot successfully on the Minisforum MS-A2.


Video Walkthrough

Watch the full walkthrough below where I demonstrate the issue and apply the workaround.


What You Will Learn

  • What changed in ESXi 8.0.3i
  • Why AMD microcode updates trigger a PSOD
  • How to reproduce the issue during boot
  • The Shift + O ESXi boot workaround
  • How to confirm the fix works successfully

Credits

Huge thanks to members of the VMware community who investigated and documented this issue:

  • Stephen Wagner
  • Patrick Kernstock
  • vAndu
  • Martin Gustafsson

Their research and testing helped identify the workaround shown in this video.

Resources:


Support My Work

? Enjoying the content? Support the channel with an espresso!

Buy Me an Espresso Coffee

Donate via PayPal


Follow Hancock’s VMware Half Hour


Support British Beekeeping ?

If you’d like to support British beekeeping, you can purchase
raw pure honey from my apiaries.

Meltonby Honey – Raw Pure Honey

UK shipping currently only.

Part 7: DIY UNRAID NAS “BY THE POWER OF UNRAID” THE SCRET REVEALED

Saturday, December 6th, 2025

By The Power Of UnRAID – The Secret Reveal Of ESXi And Windows 11 VMs

For the last few episodes of Hancock’s VMware Half Hour, we have been quietly building something a little different.
On the surface it looked like a simple DIY UNRAID NAS project and a couple of Windows 11 P2V demonstrations.
In reality, everything was running inside virtual machines on an UnRAID host.

In Part 7 of the DIY UNRAID NAS series, we finally pull back the curtain and reveal what has really been powering the lab:
UnRAID running nested ESXi and Windows 11 VMs, complete with PCI passthrough.
This post walks through the idea behind the episode, how it ties back to earlier parts, and why I keep saying,
“By the power of UnRAID.”

Recap: Parts 6, 100 and 101

If you have been following along you will have seen:

  • Part 6 – Installing and testing Samsung 990 PRO NVMe drives in the Intel NUC based NAS.
  • Part 100 – Performing P2V migrations of Windows 11 systems.
  • Part 101 – Continuing the Windows 11 P2V work and refining the process.

In those episodes the star of the show appeared to be a physical Windows 11 machine and a separate ESXi host called ESXi052.
In Part 7 we reveal that this was deliberately misleading. Both the Windows 11 system and the ESXi host were in fact virtual machines.

The Secret: Everything Was A Virtual Machine

Part 7 opens by jumping back to those previous episodes and then revealing the twist:

  • The “physical” Windows 11 machine you saw on screen was actually a Windows 11 VM.
  • The ESXi host ESXi052 that we used for P2V work was also a VM.
  • The same VM was used in Part 6 when we installed and tested the NVMe drives.

In other words, the entire recent run of content has been driven by virtual machines on UnRAID.
The NVMe upgrades, the Windows 11 P2Vs, and the ESXi demonstrations were all happening inside VMs, not on bare metal.

Windows 11 With PCI Passthrough

One of the key enabling features in this setup is PCI passthrough on UnRAID.
By passing through hardware devices such as NVMe controllers or GPUs directly into a Windows 11 VM,
we can test and demonstrate “bare metal like” performance while still keeping everything virtual.

In the video we show Windows 11 running with PCI passthrough on UnRAID, giving the VM direct access to the hardware.
This is ideal for lab work, testing, and for scenarios where you want to push a homelab system without dedicating separate physical machines.

Nested ESXi 8.0 On UnRAID

The next part of the reveal is nested virtualization.
UnRAID is hosting a VMware vSphere Hypervisor ESXi 8.0 VM which in turn can run its own VMs.
This gives an incredibly flexible environment:

  • UnRAID manages the storage, cache, parity and core virtual machine scheduling.
  • ESXi runs nested on top for VMware specific testing and lab work.
  • Windows 11 runs as another VM on the same UnRAID host, with PCI passthrough as needed.

With this approach a single Intel NUC based NAS can simulate a much larger lab
while still being compact and power efficient.

By The Power Of UnRAID

To celebrate the reveal I created a fun meme inspired by the classic “By the power of Grayskull” scene.
In our version, “By the power of UnRAID” raises ESXi and Windows 11 high above the NUC,
showing that UnRAID is the platform empowering the whole setup.

Whether you are running nested ESXi, Windows 11 with PCI passthrough, or a mixture of containers and VMs,
UnRAID makes it straightforward to combine storage flexibility with powerful virtualization features.

The Power Of UnRAID In The Homelab

The big takeaway from Part 7 is simple:

  • A single UnRAID host can consolidate multiple roles: NAS, hypervisor, and workstation.
  • You can experiment with ESXi 8.0, Windows 11, and PCI passthrough without building a large rack of servers.
  • By keeping everything virtual you gain snapshots, flexibility, and the ability to rebuild or clone systems quickly.

For homelab enthusiasts, students, and anyone who wants to learn VMware or Windows 11 in depth,
this approach offers a lot of power in a very small footprint.

Watch The Episode

If you want to see the full walkthrough, including the moment the secret is revealed,
watch Part 7 of the DIY UNRAID NAS series on Hancock’s VMware Half Hour.
You will see exactly how the Windows 11 VM, the nested ESXi host, and UnRAID all fit together.

Conclusion

Part 7 closes the loop on a long running lab story.
What looked like separate physical systems were really virtual machines,
carefully layered on top of an UnRAID powered NAS.
By the power of UnRAID, we have been able to demonstrate NVMe upgrades, Windows 11 P2Vs, and ESXi 8.0 testing
all on a single compact platform.

If you are planning a new homelab or want to refresh an existing one,
consider what UnRAID plus nested ESXi and Windows 11 VMs could do for you.

Part 1: Building a DIY NVMe NAS with the Intel NUC 11 Extreme (Beast Canyon)

Saturday, November 15th, 2025

 

Part 1: The Hardware Build

Welcome to AndysWorld.org.uk! Today, we’re diving into a project that’s perfect for anyone looking to build a powerful, yet compact, DIY Network-Attached Storage (NAS) solution. In this post, I’ll walk you through the first part of building a ‘MEGA’ NVMe NAS using the Intel NUC 11 Extreme (Beast Canyon). This mini-PC packs a punch with its powerful hardware, making it a great choice for a NAS build, especially when combined with UnRAID to handle storage and virtualization.


Why Choose the Intel NUC 11 Extreme for a NAS?

If you’ve been looking into NAS setups, you know the balance between power, size, and expandability is crucial. The Intel NUC 11 Extreme (Beast Canyon) checks all the right boxes, offering:

  • Compact Form Factor: It’s a small but powerful solution that doesn’t take up much space.

  • High-Performance NVMe Support: NVMe drives provide incredibly fast data transfer speeds—perfect for a NAS that needs to handle heavy workloads.

  • Flexibility for Virtualization: With UnRAID, you can set up multiple virtual machines, containers, and storage arrays, making it a versatile solution for any home or small office.

For this build, we’re focusing on using NVMe storage for high-speed access to files and a 64GB Kingston Fury DDR4 RAM kit to ensure smooth performance under load.


What You’ll Need for This Build:

  • Intel NUC 11 Extreme (Beast Canyon)

  • 64GB Kingston Fury DDR4 RAM

  • 2 x 512GB XPG GAMMIX NVMe SSDs

  • UnRAID Operating System

  • A few basic tools for assembly (screwdriver, anti-static mat, etc.)

If you’ve never worked with the Intel NUC before, don’t worry! I’ll guide you through every step of the assembly process. Let’s get into it!


Step-by-Step Build Process:

1. Unboxing the Intel NUC 11 Extreme

First things first, let’s unbox the Intel NUC 11 Extreme (Beast Canyon). When you open the box, you’ll find the compact, sleek chassis, which packs quite a punch for such a small form factor. This NUC is equipped with an 11th Gen Intel Core i7 processor and can support a variety of high-speed storage options, including NVMe SSDs.

2. Installing the RAM and NVMe Drives

With the NUC unboxed, the next step is to install the Kingston Fury RAM and XPG GAMMIX NVMe SSDs. Be careful during installation—especially with the tiny NVMe screws! The NUC has an easy-to-access compute board where both the RAM and NVMe drives will fit.

  • Installing the RAM: Simply slot the 64GB Kingston Fury DDR4 RAM sticks into the dedicated slots, making sure they’re fully seated.

  • Installing the NVMe SSDs: These go directly onto the motherboard and can be secured using small screws. Be sure to handle them gently as the connectors are quite delicate.

3. Reassembling the NUC

Once the RAM and NVMe drives are installed, it’s time to reassemble the NUC. This involves:

  • Reattaching the fan tray and shroud

  • Reinstalling the side and back panels

At this stage, everything should feel secure and ready for the next steps.


Why NVMe Storage for a NAS?

NVMe drives are game-changers when it comes to NAS storage. Here’s why:

  • Speed: NVMe offers lightning-fast read/write speeds compared to SATA SSDs or traditional HDDs. For anyone who works with large files or needs to serve data quickly, NVMe is a must.

  • Future-Proofing: With more applications and data being handled in the cloud, having NVMe in your NAS ensures your storage solution is ready for the future.

  • Reliability: NVMe drives are more reliable than traditional spinning hard drives, with less moving parts and faster data recovery times.


What’s Next?

Now that we’ve completed the hardware installation, in the next post, we’ll dive into setting up UnRAID on the NUC. UnRAID will allow us to easily configure our storage arrays, virtual machines, and containers—all from a user-friendly interface. Stay tuned for Part 2, where we’ll cover configuring the software, optimizing the NAS, and making sure everything runs smoothly.


Helpful Resources:

To help you along the way, I recommend checking out the blog posts from two experts in the field:


Wrapping Up

This build was just the beginning! The Intel NUC 11 Extreme provides an excellent foundation for a fast, reliable NAS. With NVMe storage and the flexibility of UnRAID, you can build a high-performance system that’s both versatile and compact.

What do you think of this build? Have you used the Intel NUC for similar projects? Drop a comment below or connect with me on social media—I’d love to hear about your experiences!


Follow Andy’s World for More DIY Tech Projects
Don’t forget to check out the latest posts and tutorials on AndysWorld.org.uk to keep up with all things tech and DIY. Happy building!


Exploring ESXi ARM Fling v2.0 with the Turing Pi Mini ITX Board

Tuesday, November 26th, 2024

As an avid enthusiast of VMware’s innovations, I’m diving headfirst into the ESXi ARM Fling v2.0, which is built on the robust VMware vSphere Hypervisor ESXi 8.0.3b codebase. The ARM architecture has always intrigued me, and with this latest version, VMware has pushed the boundaries of what’s possible with ESXi on ARM devices. It’s a playground full of potential for anyone experimenting with lightweight, power-efficient infrastructures.

 

The Turing Pi Journey

After much anticipation, my Turing Pi Mini ITX boards have arrived! These boards are compatible with the Raspberry Pi Compute Module 4, offering a modular, scalable setup perfect for ARM experimentation. With a few Compute Module 4s ready to go, I’m eager to bring this setup to life. However, finding a suitable case for the Turing Pi board has proven to be a bit of a challenge.

Case Conundrum

While Turing Pi has announced an official ITX case for their boards, it’s currently on preorder and comes with a hefty price tag. For now, I’ve decided to go with a practical and versatile option: the Streamcom Mini ITX OpenBench case. Its open-frame design is functional, and it’ll keep the board accessible during testing and configuration.

I’m also considering crafting my own custom case. Using laser-cut wood or acrylic is an appealing option, offering the opportunity to create something unique and tailored to my specific requirements. But for now, the OpenBench case will do nicely as I explore the ESXi ARM Fling.

Why ESXi ARM Fling v2.0?

The ESXi ARM Fling project is an exciting venture for anyone who loves to experiment with virtualization. Running ESXi on ARM hardware offers several advantages:

  • Energy efficiency: ARM boards consume far less power compared to traditional x86 systems.
  • Cost-effectiveness: Affordable hardware like the Raspberry Pi Compute Module 4 makes it accessible to a wider audience.
  • Flexibility: The compact form factor of ARM devices is ideal for edge computing, IoT, or even small-scale home labs.

The v2.0 update introduces enhanced support, better performance, and bug fixes, making it an excellent choice for exploring the ARM ecosystem.

What’s Next?

With the hardware in hand and the ESXi ARM Fling v2.0 ready to install, I’m planning to dive into:

  1. Setting up and configuring the Turing Pi board with ESXi.
  2. Testing the system’s stability, performance, and scalability using multiple Raspberry Pi Compute Modules.
  3. Exploring practical use cases, such as lightweight Kubernetes clusters or edge computing applications.

I’ll share updates on the build process, challenges, and performance insights in future posts. For now, I’m excited to get started and see what this setup can achieve.

Stay tuned for more! If you’ve experimented with the ESXi ARM Fling or have tips for working with the Turing Pi board, I’d love to hear from you.