At Boston Limited, our dedicated Boston Labs serves as the hub for product testing and development, giving us the ability to trial the latest hardware from across our portfolio of vendors. We don’t just benchmark hardware to verify vendor claims – we also integrate the hypervisor or operating system, software and orchestration layers to replicate real-world configurations in the way our customers would use them.
Proxmox VE is a KVM-based hypervisor that enables the deployment and management of multiple virtual machines through an intuitive web interface. Having been on the market for over a decade, Proxmox has gained significant momentum in recent years, driven by changes in the virtualisation industry and rising costs from other providers. Whether it’s escalating licence fees for existing solutions or bringing workloads back from the cloud, Proxmox VE stands out as a cost-effective, feature-rich and hardware-agnostic platform.
Virtual GPUs (vGPUs) let multiple virtual machines share a single physical GPU, transforming how you deploy graphics and compute intensive workloads. Proxmox VE’s integration of QEMU/KVM with VFIO and NVIDIA’s vGPU drivers makes it straightforward to carve up a GPU into dynamic slices that each VM can consume as if it owned the hardware.
Imagine a digital media studio wanting to spin up multiple render nodes. Using bare metal, you may have to dedicate one GPU to that task which may leave it underutilised, especially between workloads. This also extends to pass-through as you would require a separate physical GPU in a system for every VM.
Nvidia’s vGPU technology allows you to split the GPUs vRAM into pieces and separately allocate these to individual VMs. They would allow you to take Nvidia’s RTX PRO 6000 series GPU and slice it into 12, giving each VM 8GB of vRAM/framebuffer whilst sharing the full compute power between all guest VMs. As mentioned, in recent versions you no longer need all profiles to be the same, allowing you to co-locate low-demand 2/4GB profiles alongside a more demanding 16GB profile. This can optimise the utilisation levels of vRAM and compute capacity of each card.
This can be particularly advantageous with virtual workstations. Standard virtual machines have very limited graphics capability and can feel sluggish to use, and struggle to play video. Adding a vGPU profile to a VM adds can add graphical acceleration to a system. For knowledge workers this could mean you only need 1 or 2GB per profile, allowing dozens of people to share a single GPU and other server resources.
You may need to guarantee the performance of your VMs. The default job scheduler is best-effort/round-robin and gives each vGPU access to the maximum amount of compute resources. While this can provide the greatest GPU utilisation, it can suffer from “noisy neighbour syndrome”, where one VM is especially demanding and impacts the performance of the other VMs sharing that hardware.
Changing the job scheduler to equal share can ensure quality of service to each VM, however this is at the expense of GPU utilisation, and caps the max compute power available to any single vGPU profile to a percentage of the number of vGPUs running (example below would be 33%).
Here’s a streamlined view:
It’s as easy as that. At this point you can look at taking advantage of Proxmox VE’s other features, such as HCI (Hyper Converged Infrastructure) and HA (High availability). With just 3-5 nodes, you can have a cluster of Proxmox hypervisors, managed through a single interface, with scaled-out replicated storage utilising CEPH providing the redundant storage for HA mode. In the event of an entire server going down, the running VMs data is still safe and available, the VMs can spin up on a Proxmox host with spare capacity, with minimal downtime. No need for reserve of vGPU licences, as the licence server only cares about the number of currently online VMs.
I’ve been a long-time user of Proxmox and it has been possible to setup vGPU on Proxmox prior to now with their own KVM drivers, however, this was not officially supported by NVIDIA. Any bugs or technical issues means relying on your own, or the communities help, which made it unattractive as an option for the Enterprise customer that requires robust support mechanisms. With official Proxmox VE support in version 18 of the vGPU driver, this adds a compelling option against some of the more traditional Hypervisors such as VMWare or Nutanix.
At Boston Labs, we have the ability to test and benchmark the latest, greatest hardware and software solutions to ensure we can recommend the right fit for your use case. We can also perform or give access for remote testing so, for example, we could demonstrate the features of Proxmox running on a particular set of hardware. If you are interested in what we can offer, please reach out to your account manager or our sales team to see how we can guide you.
Field Application Engineer
Boston Limited
To help our clients make informed decisions about new technologies, we have opened up our research & development facilities and actively encourage customers to try the latest platforms using their own tools and if necessary together with their existing hardware. Remote access is also available
The International Conference for High Performance Computing, Networking, Storage, and Analysis