KubeVirt
The KubeVirt provider allows you to use KubeVirt to automatically provision virtual machines as nodes for your vClusters.
When a vCluster requests a new node, the platform creates a KubeVirt VirtualMachine based on the template defined in the NodeProvider and the
specific NodeType requested.
This enables you to offer different "flavors" of virtual machines (e.g., small, medium, large, different operating systems) as nodes for your vClusters, all managed from a central configuration.
Overview​
The KubeVirt provider works by defining a base virtualMachineTemplate and a collection of nodeTypes. Each nodeType represents a specific kind of
virtual machine you want to offer. When a NodeClaim is created for a vCluster, it references one of these nodeTypes. The platform then generates
a VirtualMachine manifest by combining the base template with any customizations defined in the chosen nodeType and applies it to the target
KubeVirt host cluster.
This approach allows for customizations like:
- Resource Overrides: Easily create node types with different amounts of CPU or memory.
- Template Merging: Modify specific parts of the base template, like adding taints or changing the disk image.
- Template Replacement: Define completely distinct virtual machine templates for specialized use cases.
How it Works: Cloud-Init Node Registration​
For a virtual machine to automatically join a vCluster as a node, it needs to be configured with the correct vCluster address and join token.
The KubeVirt provider automates this process using cloud-init.
Here's the workflow:
- When a
NodeClaimis processed, the platform generates acloud-initconfiguration containing the necessary registration scripts. - This configuration is stored in a Kubernetes
Secretin the KubeVirt host cluster. - The provider injects this
Secretinto theVirtualMachinedefinition as acloudInitNoClouddisk. - When the VM boots,
cloud-initexecutes the script from the disk, configuring the machine and registering it as a node with the vCluster.
This entire process depends on the guest OS image having cloud-init installed and enabled. Furthermore, the image must be compatible with KubeVirt's cloudInitNoCloud data source. Standard cloud images for distributions like Ubuntu are generally compatible. If you use custom images, you must ensure they meet this requirement.
Configuration​
A KubeVirt NodeProvider configuration consists of a reference to the host cluster, a base VM template, and a list of node types.
Minimal Example​
Here is a minimal configuration for a KubeVirt provider. It defines a single node type that uses the base template without any modifications.
apiVersion: management.loft.sh/v1
kind: NodeProvider
metadata:
name: kubevirt-provider-minimal
spec:
displayName: "KubeVirt Minimal Provider"
kubeVirt:
# clusterRef allows configuration of where KubeVirt is running and where VirtualMachines will be created.
clusterRef:
cluster: test-kubevirt-1 # is a name of Host Cluster already connected to Platform
namespace: vcluster-platform # is a namespace in this host cluster where VirtualMachines will be created
# Base template for all Virtual Machines created by this provider
virtualMachineTemplate:
spec:
dataVolumeTemplates:
- metadata:
name: containerdisk
spec:
source:
registry:
url: docker://quay.io/containerdisks/ubuntu:22.04
pvc:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: '20Gi'
template:
spec:
domain:
devices:
disks:
- name: containerdisk
disk:
bus: virtio
resources:
requests:
cpu: "1"
memory: "2Gi"
volumes:
- name: containerdisk
dataVolume:
name: containerdisk
# Define the types of nodes users can request
nodeTypes:
- name: "small-ubuntu-vm"
displayName: "Small Ubuntu VM (KubeVirt)"
maxCapacity: 10
Defining NodeTypes​
The real power of the KubeVirt provider comes from its flexible nodeTypes. You can define multiple types, each with specific overrides or entire template replacements.
Overriding Resources​
You can create a NodeType that inherits the base virtualMachineTemplate but specifies different CPU and memory resources. This is useful for offering different machine sizes.
nodeTypes:
- name: "small-ubuntu-vm-2"
displayName: "Small Ubuntu VM (KubeVirt)"
maxCapacity: 10
# This NodeType uses the base template's resources (1 CPU, 2Gi Memory)
- name: "mid-ubuntu-vm-2"
displayName: "Mid Ubuntu VM (KubeVirt)"
maxCapacity: 5
resources:
cpu: 4
memory: 5Gi
# This NodeType overrides the base template to provide 4 CPUs and 5Gi of memory
Merging Template Modifications​
The mergeVirtualMachineTemplate field allows you to provide a partial VirtualMachine template that will be strategically merged with the base template.
This is ideal for changing specific attributes, like adding Kubernetes taints or labels, without redefining the entire VM.
The merge logic follows standard Kubernetes strategic merge patching. Arrays are typically replaced, while maps (objects) are merged.
In this example, the high-memory-node type adds a taint and a label to the base template.
# ... (omitting provider spec for brevity)
nodeTypes:
- name: "high-memory-node"
maxCapacity: 2
displayName: "High Memory Node (KubeVirt)"
mergeVirtualMachineTemplate:
metadata:
labels:
# This label will be added to the base labels
workload: memory-intensive
spec:
template:
spec:
# Add a taint to ensure only specific pods schedule here
taints:
- key: "workload"
value: "high-memory"
effect: "NoSchedule"
Replacing the Entire Template​
For cases where a NodeType requires a fundamentally different configuration, you can use the virtualMachineTemplate field inside the NodeType definition. This completely ignores the base template and uses the specified one instead.
# ... (omitting provider spec for brevity)
nodeTypes:
- name: "different-vm-with-different-template-2"
maxCapacity: 1
displayName: "Different VM with Different Template (KubeVirt)"
# This template completely replaces the provider-level base template
virtualMachineTemplate:
metadata:
labels:
foo: baz # Note: the base label 'foo: bar' is not present
spec:
# ... (full, self-contained VirtualMachine spec)
Example: vCluster with static KubeVirt node​
This example demonstrates a realistic multi-cluster scenario. We will configure a NodeProvider on our management cluster that provisions specialized VMs onto a separate, KubeVirt-enabled target cluster.
Verify prerequisites.
Before you begin, ensure you have a KubeVirt-enabled Kubernetes cluster connected to your vCluster Platform management plane. You can verify this by listing the connected clusters.
kubectl get clustersYou should see your target cluster in the list. For this example, we'll assume it's named
test-kubevirt-1.NAME CREATED AT
loft-cluster 2025-08-21T10:43:14Z
test-kubevirt-1 2025-08-21T11:51:36Z
test-no-kubevirt-2 2025-08-21T12:34:25ZCreate the NodeProvider
Apply the
NodeProviderconfiguration to your management cluster. This configuration references our target cluster,test-kubevirt-1, and specifies that new VMs should be created in thevcluster-platformnamespace on that cluster.apiVersion: management.loft.sh/v1
kind: NodeProvider
metadata:
name: kubevirt-provider-advanced
spec:
displayName: "KubeVirt Advanced Provider"
kubeVirt:
clusterRef:
cluster: test-kubevirt-1
namespace: vcluster-platform
virtualMachineTemplate:
metadata:
labels:
provider: kubevirt
spec:
dataVolumeTemplates:
- metadata:
name: containerdisk
spec:
source:
registry:
url: docker://quay.io/containerdisks/ubuntu:22.04
pvc:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: '20Gi'
template:
spec:
domain:
devices:
disks:
- name: containerdisk
disk:
bus: virtio
interfaces:
- name: default
masquerade: {}
resources:
requests:
cpu: "2"
memory: "4Gi"
networks:
- name: default
pod: {}
volumes:
- name: containerdisk
dataVolume:
name: containerdisk
nodeTypes:
- name: "standard-node"
displayName: "Standard Ubuntu VM (KubeVirt)"
maxCapacity: 10
- name: "high-memory-node"
maxCapacity: 5
displayName: "High Memory Node (KubeVirt)"
resources:
cpu: 4
memory: 16Gi
mergeVirtualMachineTemplate:
metadata:
labels:
workload: memory-intensive
spec:
template:
spec:
taints:
- key: "workload"
value: "high-memory"
effect: "NoSchedule"Verify NodeTypes are Available
After applying the
NodeProvider, the platform processes it and creates correspondingNodeTyperesources. You can list these to confirm they are ready to be claimed.kubectl get nodetypeThe command will show that both the
standard-nodeand thehigh-memory-nodeare in theAvailablephase.NAME AVAILABLE TOTAL PROVIDER COST PHASE CREATED AT
high-memory-node 5 5 kubevirt-provider-advanced 72 Available 2025-08-25T13:17:30Z
standard-node 10 10 kubevirt-provider-advanced 28 Available 2025-08-25T13:17:30ZCreate a vCluster with Auto-Provisioning
Now, create a vCluster instance that is configured to automatically request a node upon creation. We'll do this by providing a values file that enables the
autoNodesfeature.First, create a file named
test-kubevirt-vc.yamlwith the following content. This tells the vCluster to automatically claim one node matching thestandard-nodetype.privateNodes:
enabled: true
autoNodes:
static:
- name: static-standard-nodes
provider: my-provider
requirements:
- property: vcluster.com/node-type
value: standard-node
quantity: 1Next, create the vCluster using this file.
vcluster create --driver platform --connect=false -f test-kubevirt-vc.yaml --cluster loft-cluster test-kubevirt-node-1Verify NodeClaim.
Once the vCluster is created, its
autoNodesconfiguration will automatically generate aNodeClaim. You can verify this on the management cluster.kubectl get nodeclaim -AYou will see a
NodeClaimfor thestandard-nodetype, which eventually will reachAvailablestate, indicating that the request has been fulfilled.NAMESPACE NAME STATUS VCLUSTER NODETYPE CREATED AT
p-default test-kubevirt-node-1-x49gd Available test-kubevirt-node-1 standard-node 2025-08-25T13:41:07ZVerify the VirtualMachine on the Target Cluster.
The
NodeClaimtriggers theNodeProviderto create aVirtualMachine. Switch yourkubectlcontext to the target KubeVirt cluster (test-kubevirt-1) and inspect the VM object.# Switch context to the KubeVirt host cluster
kubectx test-kubevirt-1
# Check for the running Virtual Machine
kubectl -n vcluster-platform get vmThe output confirms that the
VirtualMachinecorresponding to theNodeClaimis running.NAME AGE STATUS READY
test-kubevirt-node-1-x49gd 3m1s Running TrueVerify the Node in the vCluster.
Finally, connect to your vCluster and verify that the new node has successfully joined the cluster.
vcluster --driver platform connect test-kubevirt-node-1After connecting, check the nodes within the vCluster.
kubectl get nodesThe output will show the new node in the
Readystate, confirming the node is up and ready for workloads.NAME STATUS ROLES AGE VERSION
test-kubevirt-node-1-9xm54 Ready <none> 27s v1.33.4