When Distance Beats Defaults

Fixing ESXi Not Responding States in a WAN-Connected Raspberry Pi 5 Homelab

During the Christmas holidays, I turned a family visit into a distributed homelab setup.

I deployed two ESXi on ARM hosts, each running on a Raspberry Pi 5, placed in two different houses of my family. Both sites were connected back to my central VMware vCenter Server via IPsec site-to-site VPNs.

Each remote site was intentionally minimal but practical:

  • One Pi-hole VM per site for local DNS filtering
  • Local NFS storage on a Synology NAS for encrypted backup sync
  • Some reserved compute resources for future workloads

No production systems. Just learning, experimenting, and building something realistic.


The Setup at a Glance

Per site:

  • Raspberry Pi 5 running VMware ESXi on ARM
  • Local Synology NAS
  • NFS datastore mounted to ESXi
  • Encrypted backups synced from my home
  • IPsec S2S VPN to the main site hosting vCenter

This design had a clear goal:
👉 Local services stay local, WAN is used only for management and sync.


The Problem: Hosts Randomly “Not Responding”

Several times per day – mostly in the evenings both Raspberry Pi ESXi hosts suddenly appeared in vCenter as:

Not Responding

Yet every time I checked:

  • SSH access to ESXi worked
  • Pi-hole kept resolving DNS
  • NFS datastores stayed mounted
  • Backups continued syncing
  • No crashes, no reboots

From the ESXi and Synology side, everything was always up.


The Clue: Latency During Peak Hours

Once the pattern became obvious, I stopped troubleshooting ESXi and started measuring latency.

During peak hours:

  • ISP uplinks were saturated
  • IPsec VPN latency increased
  • Jitter became inconsistent

A quick test from ESXi confirmed it:

vmkping -I vmk0 <vcenter-ip>

The hosts were reachable – but not fast enough.

That’s when it became clear:

vCenter wasn’t losing the hosts.
It was losing patience.


VMware Documents This Exact Behavior

This issue is officially described in Broadcom / VMware KB 1005757:

“ESXi host disconnects intermittently from vCenter Server”
https://knowledge.broadcom.com/external/article?legacyId=1005757

Key points from the KB:

  • vCenter relies on regular management heartbeats
  • Heartbeats delayed by latency or jitter may be considered lost
  • Hosts can be marked Not Responding even when fully operational

That described my setup perfectly:

  • WAN distance
  • IPsec encryption overhead
  • Consumer internet links
  • ARM-based ESXi hosts doing exactly what they should

Root Cause (In One Sentence)

vCenter did not receive ESXi heartbeats within the expected timeframe, so it marked the hosts as Not Responding – even though they were alive, reachable, and serving workloads.


The Fix: One Setting, No Hacks

The correct and supported fix is adjusting exactly one vCenter advanced setting:

config.vpxd.heartbeat.notRespondingTimeout

This defines how long vCenter waits (in seconds) before declaring an ESXi host as Not Responding.


How to Apply the Fix

1. Open vCenter Advanced Settings

  1. Log in to the vSphere Client
  2. Navigate to
    Administration → vCenter Server Settings → Advanced Settings
  3. Click Edit Settings

2. Add or Modify the Setting

KeyTypeRecommended Value
config.vpxd.heartbeat.notRespondingTimeoutInteger60–120

My recommendation for WAN / VPN setups:

  • Start with 60 seconds
  • Increase to 90 or 120 seconds if latency fluctuates

3. Restart vCenter Services

Apply the change:

service-control --restart vpxd

⚠️ Do this during a maintenance window if vCenter availability is critical in your environment


What This Change Does Not Affect

  • ❌ No impact on vSphere HA failover timing
  • ❌ No masking of real host crashes
  • ❌ No change to VM runtime behavior

It only affects how long vCenter waits before panicking.


Lessons Learned

  • Default vCenter values assume LAN conditions
  • WAN + VPN + consumer ISPs require tuning
  • Not Responding does not mean Down
  • Raspberry Pi 5 ESXi hosts are surprisingly reliable (if cooled properly.. :P)
  • Local NFS storage is a huge win for distributed labs

Final Thoughts

Running ESXi on ARM on Raspberry Pi 5 across family homes – with local Synology NFS storage, encrypted backups, and IPsec VPNs – was a great reminder that:

Distance breaks defaults.

If you manage ESXi hosts over WAN links, tuning
config.vpxd.heartbeat.notRespondingTimeout isn’t optional – it’s essential.

Schreibe einen Kommentar

Deine E-Mail-Adresse wird nicht veröffentlicht. Erforderliche Felder sind mit * markiert

I’m Lukas

Herzlich willkommen auf meinem Blog!

Hier teile ich meine Leidenschaften für Technologie, Abenteuer und Autos. Viel Spaß beim Lesen!