<img alt="" src="https://secure.insightful-enterprise-intelligence.com/783141.png" style="display:none;">
Reserve here

NVIDIA H100 SXMs On-Demand at $2.40/hour - Reserve from just $1.90/hour. Reserve here

Reserve here

Deploy 8 to 16,384 NVIDIA H100 SXM GPUs on the AI Supercloud. Learn More

alert

We’ve been made aware of a fraudulent website impersonating Hyperstack at hyperstack.my.
This domain is not affiliated with Hyperstack or NexGen Cloud.

If you’ve been approached or interacted with this site, please contact our team immediately at support@hyperstack.cloud.

close

Deploy Hyperstack On-Demand K8s Clusters Your Way

Hyperstack gives you complete flexibility in how you deploy and operate your On-Demand K8s clusters. They can be created, scaled or deleted through the Console or API. Most clusters launch in 5–20 minutes, with full visibility into scaling, reconciliation, events, resource usage and billing.

deploy-via-ui

Deploy via UI

Launch and manage Kubernetes clusters through an intuitive web-based interface. Configure versions, node types, worker counts and deployment mode in a few clicks, with full lifecycle management through the console.

deploy-via-api

Deploy via API

Provision production-ready Kubernetes clusters with a single API request. Define your version, node flavors and parameters and Hyperstack handles orchestration and configuration automatically.

Features of Hyperstack On-Demand Kubernetes

ai-optimised-kubernetes

AI-Optimised Kubernetes

Hyperstack Kubernetes clusters are built on GPU-enabled worker flavours and NVIDIA-optimised images, making them a strong fit for AI, ML and other compute-intensive workloads. Clusters come ready for training jobs, fine-tuning pipelines, inference services and general cloud-native applications.

nvidia-gpu-support-out-of-the-box

NVIDIA GPU Support Out of the Box

Every cluster can be deployed on GPU worker flavours and includes NVIDIA-optimised drivers and base images. This ensures seamless execution of AI/ML workloads without manual GPU configuration, driver installation or compatibility troubleshooting.

effortless-deployment

Effortless
Deployment

Launch a fully configured Kubernetes infrastructure on demand in minutes with minimal setup. Hyperstack automates node provisioning, networking, OS configuration and driver installation, so teams can focus on building, training and scaling applications rather than managing infrastructure.

high-speed-networking

High-Speed
Networking

Clusters run on high-speed, low-latency networking designed for distributed computing and AI workloads. Bastion and load balancer nodes are publicly accessible, while master and worker nodes stay on an internal network. Firewall rules and bastion access provide a secure pattern for managing SSH access, traffic flow and exposed services.

node-groups-for-flexible-workload-deployment

Node Groups for Flexible Workload Deployment

Hyperstack On-Kubernetes supports node groups, a collection of worker nodes that share the same flavor and configuration. This lets you run diverse workloads within a single cluster by creating multiple node groups for different GPU or CPU flavors. Teams can isolate training, inference and general-purpose workloads efficiently, with node group creation and scaling available via the API.

Enhanced for Scalable and Production-Ready Clusters

Optimised for high-performance AI and cloud-native workloads

Persistent Storage with Hyperstack CSI Driver

The Hyperstack CSI Driver supports dynamic provisioning and lifecycle management of persistent volumes directly within Kubernetes. Workloads can request storage through standard Persistent Volume Claims in a cloud native way, with support for ReadWriteOnce volumes.

Transparent Billing for On-Demand Kubernetes Clusters

How Billing Works

Hyperstack bills Kubernetes clusters based on the compute resources powering your worker nodes.

  • Worker Nodes: Charged hourly based on the selected flavour. Billing only applies when the underlying VM is in the ACTIVE state.
  • Public IPs: Billed at a fixed hourly rate.
  • Master, Bastion and Load Balancer Nodes: Included at no additional resource cost.

Refer to the billing documentation for complete details on pricing.

transparent-billing-for-on-demand-k8s

Deploy On-Demand Kubernetes Clusters in Minutes

Hyperstack lets you spin up production-ready Kubernetes clusters in just minutes, with minimal setup needed.

manage-your-data-with-object-storage

Frequently Asked Questions

Our product support and development go hand in hand to deliver you the best solutions available.

What can I run on Hyperstack On-Demand Kubernetes?

Hyperstack On-Demand Kubernetes is built for AI, ML, and cloud-native workloads. You can run training jobs, fine-tuning pipelines, inference services, distributed training workloads, microservices and other containerised applications, all accelerated by NVIDIA GPUs and high-speed networking

How fast can I deploy a Kubernetes cluster on Hyperstack?

Clusters typically deploy in just a few minutes. Whether launched via the UI or a single API request, Hyperstack handles the rest.

Does Hyperstack support GPU-accelerated workloads 
in Kubernetes?

Yes. Every Hyperstack Kubernetes cluster includes NVIDIA-optimised drivers, enabling high-performance training, inference and distributed GPU workloads without manual configuration.

How does scaling work with Hyperstack Kubernetes?

Nodes can be scaled manually at any time through the API or UI and Hyperstack will also launch Cluster Autoscaler soon which automatically adds or removes worker nodes based on real-time resource demand. This gives you both full control and efficient, hands-off scaling when needed.

Can I use persistent storage with my Kubernetes clusters?

Absolutely. The Hyperstack CSI Driver allows your clusters to provision and manage persistent volumes dynamically. Stateful workloads can attach storage declaratively, making it easy to run databases, ML pipelines, or long-running applications.