Hypervisor and Private Cloud Platform Comparison -- Which One Fits Your Environment
Reference Guide | All Platforms | Audience: Infrastructure Architects, Platform Engineers, IT Decision Makers
Ten hypervisor and private cloud platforms are covered in this series. Each has a full end-to-end setup article. This guide is the decision layer that sits in front of all of them. It answers the question you have to answer before you start reading any of the individual articles: which platform fits your environment, your team, and your constraints?
This is not a benchmarks article. Synthetic benchmarks don't translate to production environments and vendor-sponsored comparisons don't survive contact with the real world. What this covers is architecture, operational fit, licensing economics, Veeam backup compatibility, and the specific conditions under which each platform is the right answer -- and the conditions under which it isn't.
1. Platform Landscape at a Glance
Before going deep on any individual platform, it helps to understand how they group. These ten platforms fall into four categories based on their fundamental architecture and target use case.
| Platform | Category | License Model | Hypervisor | Native HA Storage | Veeam v13 Support |
|---|---|---|---|---|---|
| VMware vSphere 8 | Enterprise HCI / Traditional | Commercial (Broadcom) | ESXi | vSAN (optional) | Full -- CBT, HotAdd, CDP |
| VMware VCF 9 | Full-stack private cloud | Commercial (Broadcom) | ESXi | vSAN (included) | Full -- VCF-aware deployment |
| Microsoft Hyper-V 2025 | Enterprise / Windows-native | Included with Windows Server | Hyper-V | S2D (Storage Spaces Direct) | Full -- RCT, HotAdd. No CDP. |
| Windows Server + Hyper-V cluster | SMB / Mid-market | Included with Windows Server | Hyper-V | S2D or external | Full -- RCT, HotAdd. No CDP. |
| Proxmox VE 8 | Open source / SMB | Free + paid support subscription | KVM + LXC | Ceph (built-in) | Full -- agentless native plugin (VMs). Agent-based for LXC containers. |
| Nutanix AHV | Enterprise HCI | Commercial (included with Nutanix) | AHV (KVM-based) | NDFS (native) | Full -- dedicated Veeam-Nutanix integration |
| oVirt / RHV | Open source enterprise | Free (oVirt) / Commercial (RHV) | KVM | GlusterFS or external | Full -- agentless native plugin |
| Scale Computing HyperCore | Edge / SMB HCI | Commercial | KVM-based (SCRIBE) | SCRIBE (native) | Full -- agentless native plugin |
| Harvester HCI | Cloud-native HCI | Open source (SUSE) | KubeVirt | Longhorn (native) | Agent-based (no native VBR integration) |
| OpenStack | Open source private cloud | Free + support contracts | KVM (Nova/libvirt) | Ceph via Cinder | Agent-based (no hypervisor-level integration) |
2. The Platform Profiles
Each platform profile below covers what it is, who it's built for, where it fits well, and where it doesn't. Read the profiles for the platforms you're seriously considering, then use the decision framework in Section 3 to cut to a shortlist.
VMware vSphere 8
vSphere is the incumbent enterprise virtualization platform. Most organizations running VMs at any scale have vSphere in their history if not their present. The 2024 Broadcom acquisition changed the licensing model fundamentally: the previous a-la-carte model with free ESXi is gone. ESXi is now bundled into VCF or VVF (VMware vSphere Foundation) subscription tiers. There is no standalone free hypervisor path from VMware anymore.
What vSphere still has is the deepest ecosystem of any virtualization platform: every major storage vendor, every backup vendor, every monitoring tool, every security scanner has a vSphere integration. The operational toolchain (vCenter, DRS, HA, vMotion, vLCM lifecycle management) is mature and well understood. If your team has vSphere experience, that institutional knowledge has real value. The question is whether that value justifies the subscription cost at your scale.
Strong fit when
Existing vSphere estate and team expertise
Deep third-party tool integrations are required
Enterprise support SLAs are contractually mandated
Regulated environments requiring certified platform
Poor fit when
Budget pressure is high and Broadcom pricing is a hard constraint
Small VM count where per-socket cost is difficult to justify
Greenfield deployment with no existing VMware investment
VMware Cloud Foundation 9
VCF is Broadcom's full-stack private cloud platform. It bundles vSphere, vSAN, NSX networking, SDDC Manager lifecycle automation, and Aria Operations into a single licensed unit. Every component is deployed and lifecycle-managed together by SDDC Manager. You don't configure individual pieces independently -- you deploy workload domains and let SDDC Manager manage the stack.
VCF is the right answer when you want the full VMware stack under a single support contract with automated lifecycle management and you're running at a scale where that automation saves meaningful operational time. It is not the right answer if you only need basic virtualization without NSX networking or if the all-in licensing cost is prohibitive. VCF 9 introduced significant simplifications to the deployment process compared to VCF 4.x and 5.x, but it remains the most complex platform in this series to deploy from scratch.
Strong fit when
Large-scale vSphere deployments needing lifecycle automation
NSX micro-segmentation is a security requirement
Multiple sites with workload domain isolation requirements
Existing VCF licensing already in place
Poor fit when
Small or medium environments where VCF overhead exceeds benefit
NSX is not needed and vSAN Foundation licensing is sufficient
Team lacks VMware-certified architects for initial deployment
Microsoft Hyper-V 2025 (Standalone)
Hyper-V is Microsoft's hypervisor, built into Windows Server at no additional license cost. A standalone Hyper-V deployment runs Windows Server 2025 with the Hyper-V role on individual hosts. Failover Clustering adds HA and live migration between nodes. Storage Spaces Direct (S2D) enables converged or hyper-converged storage using local NVMe or SSD across cluster nodes.
The economics are the primary argument for Hyper-V: if you're already licensed for Windows Server (which most Microsoft-heavy shops are), the hypervisor has no additional per-socket cost. Management tooling has improved significantly with Windows Admin Center, though it still lags vCenter in depth for large environments. Veeam's Hyper-V integration is first-class -- RCT (Resilient Change Tracking) is the Hyper-V equivalent of CBT and performs well.
Strong fit when
Microsoft-centric shops with existing Windows Server licensing
Windows workload-heavy VM estate
Teams with strong Windows administration skills
Cost pressure makes per-socket hypervisor licensing difficult
Poor fit when
Linux-heavy workloads requiring deep Linux VM integration
Large heterogeneous environments where vCenter-level management depth is needed
Non-Microsoft storage backends with limited Hyper-V integration
Windows Server + Hyper-V Cluster (S2D)
A Windows Server Hyper-V cluster with Storage Spaces Direct is Microsoft's HCI answer. S2D pools local NVMe, SSD, and HDD across cluster nodes into a shared storage pool surfaced as Cluster Shared Volumes (CSVs). VMs live on CSVs and can live-migrate between any node in the cluster. This is Microsoft's direct competitor to vSAN and Nutanix -- a converged compute-and-storage architecture that doesn't require a separate SAN.
S2D works well when your hardware is on Microsoft's Hardware Compatibility List and you use validated configurations. It struggles with mixed or non-validated hardware where the resiliency behavior and performance characteristics are harder to predict. The two-node S2D configuration (with a witness) is a legitimate small-site HCI option that many SMBs use successfully.
Strong fit when
Microsoft shops wanting HCI without a third-party SAN
Azure Stack HCI roadmap is in scope
Two-node edge or branch office deployments
Hardware is on the Microsoft HCL
Poor fit when
Hardware is not on the HCL and validated configurations aren't possible
Team lacks deep Windows clustering expertise
Workloads require non-Windows-native storage features
Proxmox VE 8
Proxmox VE is the most popular VMware alternative in the post-Broadcom migration wave for a reason: it's genuinely capable, free to use, and has a well-designed web UI that covers the majority of operational tasks without CLI. KVM provides the VM layer. LXC containers run alongside VMs in the same cluster. Built-in Ceph support means you can deploy converged HCI without additional software. HA clustering, live migration, and SDN (Software Defined Networking with VLANs and VXLANs) are all included.
Proxmox's limitation is enterprise support breadth. Proxmox Server Solutions offers paid support subscriptions, but the ecosystem of certified third-party integrations is smaller than VMware or Nutanix. Veeam v13 includes a native Proxmox VE plugin that backs up VMs agentlessly using a Linux worker VM deployed on the Proxmox host -- no guest agent required in VMs. CBT is supported via QEMU Dirty Bitmaps at the hypervisor level. One caveat: bitmaps for disks in RAW and VMDK formats are discarded on VM reboot or restart, which forces a full re-read on the next backup run for affected VMs. Using qcow2 format avoids this. For containers (LXC), Veeam uses agent-based protection.
Strong fit when
VMware cost pressure is the primary driver of platform evaluation
Mixed VM and container workloads on the same cluster
Linux-comfortable teams who don't need a GUI for everything
Homelab, dev/test, or SMB environments
Poor fit when
Enterprise support SLAs require a commercially backed hypervisor
Hypervisor-level CBT backup performance is a hard requirement
Regulatory frameworks require certified platform documentation
Nutanix AHV
Nutanix AHV is the hypervisor included with Nutanix HCI hardware. It's a KVM-based hypervisor that Nutanix has built a full management and storage stack on top of: NDFS (Nutanix Distributed File System) provides the converged storage, Prism Central provides multi-cluster management, Flow provides micro-segmentation networking, and Leap provides automated DR runbooks to Nutanix secondary sites or cloud.
AHV is not sold or deployed independently -- it's part of the Nutanix platform. If you're buying Nutanix hardware, AHV is the right hypervisor choice for most workloads. If you're not buying Nutanix hardware, AHV is not an option. Veeam v13 has a dedicated integration for Nutanix AHV including its own backup proxy appliance and Change Region Tracking (CRT), which is the AHV equivalent of CBT. SureBackup verification works with limitations -- I/O isolation in virtual labs requires specific configuration that the dedicated SureBackup with AHV article on this site covers in detail.
Strong fit when
Nutanix hardware is already deployed or under procurement
Converged HCI with minimal separate storage management is required
Nutanix Leap DR to cloud or secondary site is in scope
Single-vendor support for compute, storage, and hypervisor is preferred
Poor fit when
Non-Nutanix hardware is the deployment target
Budget does not support Nutanix hardware and software licensing
Deep third-party ecosystem integrations are required that don't have Nutanix AHV support
oVirt / Red Hat Virtualization (RHV)
oVirt is the upstream open source project that Red Hat Virtualization (RHV) was built on. Red Hat announced end-of-life for RHV in 2026, directing customers toward OpenShift Virtualization. oVirt continues as a community project. The self-hosted engine model -- where the management VM runs inside the cluster it manages -- is an elegant architecture that eliminates the need for a separate management server, though it introduces recovery complexity if the engine VM itself fails.
oVirt's position is complicated by the RHV EOL. New deployments on oVirt make sense only if you have specific reasons to choose it -- existing KVM expertise, Red Hat ecosystem integration requirements, or infrastructure that already runs RHEL. Veeam v13 includes a native oVirt KVM plugin using the same worker-based agentless architecture as the Proxmox and AHV integrations. Workers are Linux VMs deployed inside the oVirt cluster; no guest agent is required in protected VMs. CBT is supported at the hypervisor level via the plugin.
Strong fit when
Existing oVirt or RHV deployment being maintained
Red Hat ecosystem and RHEL integration is a requirement
KVM expertise is strong in the team and a managed KVM platform is preferred over bare KVM
Poor fit when
Greenfield deployment -- RHV EOL and oVirt community trajectory make this a declining platform for new deployments
Commercial support is required (RHV EOL in 2026)
Hypervisor-level CBT backup is a requirement
Scale Computing HyperCore
Scale Computing HyperCore is purpose-built for edge deployments and distributed organizations managing many small sites. The operational model is fundamentally different from the other platforms in this list: Scale Computing ships hardware appliances with the hypervisor, storage (SCRIBE), and management plane fully integrated. You don't build a HyperCore cluster from commodity hardware. You buy Scale Computing HC3 or HC4 appliances and plug them in.
Fleet Hub is the central management plane that manages clusters across dozens or hundreds of sites from a single console. This is HyperCore's primary differentiator: the ability to manage a large fleet of small-site deployments centrally with minimal on-site IT expertise required. Veeam v13 has dedicated integration including a native Scale Computing plugin. Replication between HyperCore sites is built into the platform and can be coordinated with Veeam for layered protection.
Strong fit when
Many distributed edge or branch office sites to manage
Minimal on-site IT staff at each location
Appliance-based model preferred over DIY hardware builds
Simple VM workloads without complex networking requirements
Poor fit when
Large centralized data center deployments where appliance economics don't scale
Complex networking or storage requirements that exceed the appliance model
Commodity hardware is a requirement for cost or procurement reasons
Harvester HCI
Harvester is architecturally unique in this list. Every other platform here is a hypervisor with optional Kubernetes support bolted on. Harvester is a Kubernetes cluster with VMs running inside KubeVirt as Kubernetes pods. The storage system is Longhorn, a Kubernetes-native distributed block store. The mental model is inverted: you're running VMs on Kubernetes, not Kubernetes on VMs.
This architecture makes Harvester the right answer for a narrow but well-defined use case: organizations that are already running cloud-native Kubernetes workloads and need to also run VMs in the same infrastructure, managed by the same team, through the same toolchain. Rancher integrates with Harvester to provide a unified management plane across VM workloads and Kubernetes clusters provisioned inside Harvester. For organizations outside that use case -- particularly those running primarily traditional VM workloads -- Harvester's inverted architecture adds complexity with no benefit. Veeam protection for Harvester VMs is agent-based. There is no native VBR-to-Harvester hypervisor integration.
Strong fit when
Cloud-native teams that need to run VMs alongside Kubernetes workloads
Existing Rancher/RKE2 environment where unified management is valued
SUSE ecosystem alignment is preferred
GitOps-style VM management via kubectl is desirable
Poor fit when
Traditional VM-only workloads with no Kubernetes component
Team has no Kubernetes operational experience
Hypervisor-level CBT backup performance is a hard requirement
Mixed CPU generations in the cluster (prevents live migration)
OpenStack
OpenStack is the largest open source private cloud platform in production at scale globally. It is also the most operationally complex platform in this series by a significant margin. A properly architected OpenStack deployment can run hundreds of thousands of VMs across multiple regions with full multi-tenancy, project isolation, quota enforcement, and self-service provisioning. Getting there requires a dedicated platform engineering team and a minimum of three controller nodes plus separate compute, storage, and network node roles.
OpenStack is not a hypervisor in the traditional sense -- it's a cloud orchestration layer that manages KVM hypervisors (via Nova/libvirt) across many compute nodes. The right comparison is not "OpenStack vs Proxmox" but "OpenStack vs AWS/Azure" for private cloud use cases. Veeam protection is agent-based with no hypervisor-level integration. Organizations running OpenStack at scale typically use Ceph-native snapshots or purpose-built cloud backup tools alongside or instead of Veeam for VM-level protection.
Strong fit when
Large-scale multi-tenant private cloud is the target architecture
Self-service tenant provisioning is a requirement
Platform engineering team exists and is resourced to operate it
OpenShift on OpenStack is in scope (supported integration)
Poor fit when
Small or medium environments where OpenStack complexity exceeds the problem it solves
No dedicated platform engineering team to own it
Simple VM hosting without multi-tenancy or self-service requirements
Hypervisor-level Veeam integration is a hard backup requirement
3. Decision Framework
Work through these decision points in order. Most environments hit a definitive answer within the first two or three steps.
Decision 1 -- Scale and complexity
If you need a large-scale multi-tenant private cloud with self-service provisioning across teams or organizations, OpenStack is the only platform in this list built for that use case. Every other platform here is an infrastructure platform for running VMs, not a cloud platform for running tenants. If that's not your requirement, continue.
Decision 2 -- Cloud-native or VM-native
If your team is running Kubernetes workloads and needs to run VMs in the same infrastructure under the same toolchain, Harvester is the only platform here designed for that architecture. If your workloads are primarily traditional VMs with Kubernetes as a separate concern, continue.
Decision 3 -- Edge or distributed sites
If you're managing many small sites with minimal on-site IT staff and need central fleet management, Scale Computing HyperCore is built specifically for that model. No other platform in this list has the same edge-first fleet management capability. If you're running a centralized data center environment, continue.
Decision 4 -- Hardware is Nutanix
If your compute infrastructure is Nutanix HCI hardware, AHV is the right hypervisor. The full Nutanix stack (NDFS, Prism Central, Flow, Leap) only works on Nutanix hardware. If your hardware is commodity or a different vendor, continue.
Decision 5 -- Microsoft ecosystem depth
If you're a Windows-centric shop with existing Windows Server licensing, Windows workload-heavy VMs, and strong Windows administration skills, Hyper-V (standalone or S2D cluster) has strong economics and first-class Veeam support. Add S2D if you want converged HCI without a separate SAN and your hardware is on the HCL. If you're not Microsoft-centric or need deeper Linux VM management, continue.
Decision 6 -- VMware licensing economics
If you have an existing vSphere estate with team expertise and the post-Broadcom subscription pricing fits your budget, staying on vSphere or moving to VCF for larger deployments is the path of least operational risk. If Broadcom pricing is a hard constraint and you're evaluating alternatives, continue.
Decision 7 -- Enterprise support requirements
If your organization requires a commercially supported hypervisor with vendor SLAs for regulatory or contractual reasons, your options are vSphere, VCF, Nutanix AHV, Hyper-V, and Scale Computing HyperCore. oVirt's RHV commercial support ends in 2026. If open source with community support is acceptable, Proxmox VE is the leading VMware alternative in this category. oVirt remains viable for existing deployments but is not recommended for new ones.
Decision 8 -- Hypervisor-level Veeam CBT integration
If hypervisor-level changed block tracking for Veeam backup performance is a hard requirement, six of the ten platforms support it without a guest agent: VMware vSphere/VCF (CBT via VADP), Hyper-V (RCT), Nutanix AHV (CRT), Proxmox VE (QEMU Dirty Bitmaps -- with a caveat on RAW/VMDK disk format and reboot), oVirt (native plugin CBT), and Scale Computing HyperCore (native agentless plugin). Harvester and OpenStack do not -- both require agent-based Veeam protection with no hypervisor-level integration.
4. Veeam v13 Compatibility Summary
| Platform | Protection Method | CBT / Change Tracking | SureBackup | CDP Support | Notes |
|---|---|---|---|---|---|
| VMware vSphere 8 | Agentless (VBR native) | CBT (hypervisor-level) | Full support | Yes (CDP) | Best-in-class Veeam integration. HotAdd, NBD, Direct SAN transport modes. |
| VMware VCF 9 | Agentless (VBR native) | CBT (hypervisor-level) | Full support | Yes (CDP) | Same as vSphere. VCF-aware deployment required for proper proxy placement. |
| Hyper-V 2025 | Agentless (VBR native) | RCT (hypervisor-level) | Full support | No | First-class integration. On-host backup mode available as fallback. CDP is not supported -- Hyper-V has no equivalent to vSphere VAIO. |
| Windows Server + Hyper-V cluster | Agentless (VBR native) | RCT (hypervisor-level) | Full support | No | Identical to standalone Hyper-V. CSV-aware backup handles clustered VMs. No CDP support. |
| Nutanix AHV | Agentless (Veeam AHV proxy) | CRT (hypervisor-level) | Partial -- see AHV SureBackup article | Yes (Universal CDP) | Dedicated Veeam-Nutanix integration. Separate AHV proxy appliance required. |
| Proxmox VE 8 | Agentless (native plugin + worker VM) | QEMU Dirty Bitmaps (hypervisor-level). Bitmaps lost on reboot for RAW/VMDK disks -- use qcow2 to avoid. | Limited | No | Native Proxmox VE plugin included in VBR. Worker VM deployed per host handles data transfer. No guest agent needed for VMs. LXC containers require agent. |
| oVirt / RHV | Agentless (native plugin + worker VM) | Hypervisor-level CBT via oVirt KVM plugin | Limited | No | Native oVirt KVM plugin included in VBR. Worker VM deployed in cluster handles data transfer. No guest agent needed. |
| Scale Computing HyperCore | Agentless (native plugin + worker) | Hypervisor-level via native plugin | Limited | No | Fully agentless native plugin included in VBR. No guest agent required. Worker-based architecture same as Proxmox and oVirt integrations. |
| Harvester HCI | Agent-based (VAL / VAW) | Agent-level CBT | No | No | No native VBR-to-Harvester integration. VMs are KubeVirt pods -- treat as Linux/Windows guests. |
| OpenStack | Agent-based (VAL / VAW) | Agent-level CBT | No | No | No hypervisor-level integration. Ceph-native snapshots often used alongside or instead of Veeam at scale. |
5. Article Index
Each platform in this comparison has a full end-to-end setup article on this site. The articles cover installation, networking, storage, HA clustering, VM management, and Veeam integration where applicable. Use the decision framework above to identify your platform, then go to the corresponding article.
| Platform | Article Title | Key Topics |
|---|---|---|
| VMware vSphere 8 | VMware vSphere 8 Cluster Setup End to End | ESXi, VCSA, vSAN, DRS, HA, vLCM |
| VMware VCF 9 | VMware Cloud Foundation 9: End-to-End Setup | VCF Installer, VCF Operations, Workload Domains, Day 2 |
| Hyper-V 2025 | Windows Server 2025 and Hyper-V Cluster Setup: End to End | Failover Clustering, S2D, WAC, Live Migration |
| Windows Server + Hyper-V cluster | Beyond the Scaffold: Building a Production-Ready Hyper-V 2025/2022 Fleet | Fleet management, S2D, CSV, cluster hardening |
| Proxmox VE 8 | Proxmox VE 8: Complete Three-Node Cluster Setup End to End | KVM, Ceph, HA, SDN, PCIe passthrough |
| Nutanix AHV | Nutanix AHV End-to-End Setup | Foundation, Prism Central, Networking, Storage, Flow |
| oVirt / RHV | oVirt and RHV End-to-End Setup | Self-Hosted Engine, Storage Domains, Networking, HA |
| Scale Computing HyperCore | Scale Computing HyperCore End-to-End Setup | SCRIBE Storage, DR Replication, Fleet Hub |
| Harvester HCI | Harvester HCI End-to-End Setup | KubeVirt, Longhorn, Rancher Integration, DR |
| OpenStack | OpenStack End-to-End Setup | Kolla-Ansible, OVN Networking, Ceph Storage, Multi-Tenancy |
Key Takeaways
- No single platform is the right answer for every environment. The correct choice depends on scale, team skills, hardware constraints, licensing economics, and backup integration requirements -- in that order.
- VMware vSphere and VCF remain the platforms with the deepest ecosystem integration and best Veeam support, but the post-Broadcom licensing model has made the cost calculus for smaller environments difficult. Large enterprises with existing VMware estates are likely to stay. Greenfield deployments need a harder justification.
- Hyper-V is the right answer for Microsoft-centric shops with Windows Server licensing already paid. The Veeam integration (RCT) is first-class. Storage Spaces Direct provides a credible HCI option when hardware is HCL-validated.
- Proxmox VE is the leading open source VMware alternative for SMB and mid-market environments. Free to use, genuinely capable, active development. Veeam protection is agentless via native plugin with QEMU Dirty Bitmap CBT -- no guest agent needed for VMs. The main limitation is a smaller enterprise support ecosystem compared to VMware or Nutanix.
- Nutanix AHV is only relevant if you're buying Nutanix hardware. If you are, it's the right hypervisor -- the full NDFS/Prism/Flow stack is tightly integrated and the Veeam CRT integration is solid.
- Scale Computing HyperCore is purpose-built for edge and distributed sites. If you're managing many small locations with minimal on-site staff, nothing else in this list competes on operational simplicity at scale.
- Harvester fits a specific use case: cloud-native teams running VMs and Kubernetes workloads on the same infrastructure. Outside that use case, its KubeVirt architecture adds complexity with no benefit.
- oVirt is viable for maintaining existing deployments. Red Hat Virtualization reaches end-of-life in 2026. New deployments on oVirt carry platform longevity risk.
- OpenStack is a private cloud platform, not a hypervisor. It solves a different problem -- large-scale multi-tenancy and self-service provisioning -- at a significantly higher operational cost. Right for organizations with platform engineering teams. Wrong for everyone else.
- For Veeam-centric environments with tight backup windows, the CBT distinction matters: VMware (CBT via VADP), Hyper-V (RCT), Nutanix AHV (CRT), Proxmox VE (QEMU Dirty Bitmaps), oVirt (native plugin CBT), and Scale Computing HyperCore (native agentless plugin) all support hypervisor-level change tracking without a guest agent. Harvester and OpenStack are the exceptions -- both require agent-based Veeam protection with no hypervisor-level integration.