Hyperstack
On-Demand Kubernetes
Hyperstack On-Demand Kubernetes provides a fast, flexible environment for deploying, scaling and managing production-ready Kubernetes clusters built for AI and cloud-native applications.
Get started with Hyperstack On-Demand Kubernetes today!
Deploy Hyperstack On-Demand K8s Clusters Your Way
Hyperstack gives you complete flexibility in how you deploy and operate your On-Demand K8s clusters. They can be created, scaled or deleted through the Console or API. Most clusters launch in 5–20 minutes, with full visibility into scaling, reconciliation, events, resource usage and billing.
Deploy via UI
Launch and manage Kubernetes clusters through an intuitive web-based interface. Configure versions, node types, worker counts and deployment mode in a few clicks, with full lifecycle management through the console.
Deploy via API
Provision production-ready Kubernetes clusters with a single API request. Define your version, node flavors and parameters and Hyperstack handles orchestration and configuration automatically.
Features of Hyperstack On-Demand Kubernetes
AI-Optimised Kubernetes
Hyperstack Kubernetes clusters are built on GPU-enabled worker flavours and NVIDIA-optimised images, making them a strong fit for AI, ML and other compute-intensive workloads. Clusters come ready for training jobs, fine-tuning pipelines, inference services and general cloud-native applications.
NVIDIA GPU Support Out of the Box
Every cluster can be deployed on GPU worker flavours and includes NVIDIA-optimised drivers and base images. This ensures seamless execution of AI/ML workloads without manual GPU configuration, driver installation or compatibility troubleshooting.
Effortless
Deployment
Launch a fully configured Kubernetes infrastructure on demand in minutes with minimal setup. Hyperstack automates node provisioning, networking, OS configuration and driver installation, so teams can focus on building, training and scaling applications rather than managing infrastructure.
High-Speed
Networking
Clusters run on high-speed, low-latency networking designed for distributed computing and AI workloads. Bastion and load balancer nodes are publicly accessible, while master and worker nodes stay on an internal network. Firewall rules and bastion access provide a secure pattern for managing SSH access, traffic flow and exposed services.
Node Groups for Flexible Workload Deployment
Hyperstack On-Kubernetes supports node groups, a collection of worker nodes that share the same flavor and configuration. This lets you run diverse workloads within a single cluster by creating multiple node groups for different GPU or CPU flavors. Teams can isolate training, inference and general-purpose workloads efficiently, with node group creation and scaling available via the API.
Enhanced for Scalable and Production-Ready Clusters
Optimised for high-performance AI and cloud-native workloads
Persistent Storage with Hyperstack CSI Driver
The Hyperstack CSI Driver supports dynamic provisioning and lifecycle management of persistent volumes directly within Kubernetes. Workloads can request storage through standard Persistent Volume Claims in a cloud native way, with support for ReadWriteOnce volumes.
Transparent Billing for On-Demand Kubernetes Clusters
How Billing Works
Hyperstack bills Kubernetes clusters based on the compute resources powering your worker nodes.
- Worker Nodes: Charged hourly based on the selected flavour. Billing only applies when the underlying VM is in the ACTIVE state.
- Public IPs: Billed at a fixed hourly rate.
- Master, Bastion and Load Balancer Nodes: Included at no additional resource cost.
Refer to the billing documentation for complete details on pricing.
Deploy On-Demand Kubernetes Clusters in Minutes
Hyperstack lets you spin up production-ready Kubernetes clusters in just minutes, with minimal setup needed.
Frequently Asked Questions
Our product support and development go hand in hand to deliver you the best solutions available.
What can I run on Hyperstack On-Demand Kubernetes?
Hyperstack On-Demand Kubernetes is built for AI, ML, and cloud-native workloads. You can run training jobs, fine-tuning pipelines, inference services, distributed training workloads, microservices and other containerised applications, all accelerated by NVIDIA GPUs and high-speed networking
How fast can I deploy a Kubernetes cluster on Hyperstack?
Clusters typically deploy in just a few minutes. Whether launched via the UI or a single API request, Hyperstack handles the rest.
Does Hyperstack support GPU-accelerated workloads in Kubernetes?
Yes. Every Hyperstack Kubernetes cluster includes NVIDIA-optimised drivers, enabling high-performance training, inference and distributed GPU workloads without manual configuration.
How does scaling work with Hyperstack Kubernetes?
Nodes can be scaled manually at any time through the API or UI and Hyperstack will also launch Cluster Autoscaler soon which automatically adds or removes worker nodes based on real-time resource demand. This gives you both full control and efficient, hands-off scaling when needed.
Can I use persistent storage with my Kubernetes clusters?
Absolutely. The Hyperstack CSI Driver allows your clusters to provision and manage persistent volumes dynamically. Stateful workloads can attach storage declaratively, making it easy to run databases, ML pipelines, or long-running applications.