How to Run Multiple GPU KAI Schedulers in Kubernetes Using vCluster
In today's cloud-native landscape, GPU workloads are becoming increasingly critical. From training large language models to running inference APIs, organizations are investing heavily in GPU infrastructure. But with this investment comes a challenge: how do you safely test and deploy new GPU schedulers without risking your entire production environment?
Related talks: I'll be presenting this topic at SREDay Paris Q4 2025 and Conf42 Kube Native 2025. Check the talks page for more details.