<img alt="" src="https://secure.insightful-enterprise-intelligence.com/783141.png" style="display:none;">
Reserve here

NVIDIA H100 SXMs On-Demand at $2.40/hour - Reserve from just $1.90/hour. Reserve here

Reserve here

Deploy 8 to 16,384 NVIDIA H100 SXM GPUs on the AI Supercloud. Learn More

alert

We’ve been made aware of a fraudulent website impersonating Hyperstack at hyperstack.my.
This domain is not affiliated with Hyperstack or NexGen Cloud.

If you’ve been approached or interacted with this site, please contact our team immediately at support@hyperstack.cloud.

close
|

Updated on 23 Jan 2026

How to Deploy Open WebUI: A Quick Setup Guide

TABLE OF CONTENTS

NVIDIA H100 SXM On-Demand

Sign up/Login
summary
If you are looking for a simple way to run a ChatGPT-like interface on your own infrastructure without giving up performance or control, OpenWebUI is built for exactly that. OpenWebUI is an open-source, self-hosted AI web interface that lets you interact with LLMs using Ollama or OpenAI-compatible APIs, all while keeping your data private. When paired with GPU-powered cloud infrastructure, it delivers fast, responsive inference suitable for real development and production workloads.

This setup guide shows you how to deploy Open WebUI on Hyperstack so you can quickly get started using GPU-powered cloud infrastructure. Follow the guide below to launch a working setup in just a few minutes.

What is OpenWebUI?

Open WebUI is an open-source, feature-rich, and user-friendly self-hosted AI interface designed to run entirely offline. Built on universal standards, it supports Ollama and OpenAI-compatible APIs, making it easy to connect local or remote LLM backends. It provides a ChatGPT-like web experience while giving you full control over models, data and infrastructure.

Running Open WebUI on Hyperstack lets you:

  1. Use high-end on-demand GPUs like NVIDIA A100, NVIDIA H100 and more

  2. Deploy on pre-configured environments with one click

  3. Scale resources as your usage grows

  4. Pay only for the resources you use

What You’ll Need

Before you start, make sure you have:

  1. A Hyperstack account. Follow our documentation to create your account.

  2. Access to a GPU VM

  3. Basic familiarity with the command line

  4. An SSH client

Step 1: Launch a GPU VM on Hyperstack

1. Log in to the Hyperstack Console.

2. Create a new Virtual Machine with:

Screenshot 2026-01-22 170901

  • OS image: Latest Ubuntu with CUDA and Docker pre-installed. Select the "Ubuntu Server 24.04 LTS R570 CUDA 12.8 with Docker".

Screenshot 2026-01-22 171229

3. Launch the VM. Hyperstack will automatically select the region.

Once the VM is running, connect via SSH and note the public IP address.

Step 2: Deploy Open WebUI

Hyperstack’s Ubuntu images come with Docker, so no additional installation is needed. Now, run the following commands to deploy OpenWebUI connected to Ollama:

# 1) Create a Docker network
docker network create ollama-net

# 2) Start Ollama container
sudo docker run -d --gpus=all --network ollama-net \
  -p 11434:11434 \
  -v /home/ubuntu/ollama:/root/.ollama \
  --name ollama \
  --restart always \
  ollama/ollama:latest

# 3) Pull a model (example: qwen3:4b)
sudo docker exec -it ollama ollama pull qwen3:4b

# 4) Start Open WebUI container (connects to Ollama API)
sudo docker run -d --network ollama-net \
  -p 3000:8080 \
  -v open-webui:/app/backend/data \
  --name open-webui \
  --restart always \
  -e OLLAMA_BASE_URL=http://ollama:11434

Step 3: Access the Web Interface

1. Open your VM’s firewall settings:

2. Allow port 3000 for your IP address (or all IPs if testing, but less secure). For instructions, see here.

3. Open your browser and go to: http://[your-public-ip]:3000

For example: http://198.145.126.7:3000

4. Set up an admin account for Open WebUI and save your credentials.

Frame 1-1-1

Once done, you can:

  • Create users
  • Connect models
  • Start chatting with your AI setup

Step 4: Interact with Open WebUI

Depending on your setup, you can connect Open WebUI to:

  • A local model running on the same GPU

  • A model server inside your Hyperstack environment

  • An external inference endpoint

OpenWebUI will route prompts directly to your chosen model backend.

Common Issues and Quick Fixes

Here are quick fixes and tips to use if you encounter any issues with deploying Open WebUI:

WebUI not loading?

  • Confirm port 3000 is open in firewall settings

Slow responses?

  • Check GPU availability: nvidia-smi
  • Monitor GPU memory usage

Conclusion

Deploying Open WebUI on Hyperstack gives you a fast, flexible way to run a self-hosted AI chat interface on powerful GPU infrastructure. With just a few steps, you can go from instance launch to a fully working AI UI, ready to experiment, build and scale.

Get Started on Hyperstack

Spin up a GPU VM and deploy Open WebUI in minutes.

Run on Hyperstack →

FAQs

Is Open WebUI free to use?

Yes. Open WebUI is open-source and free to use. You only pay for the infrastructure (such as GPU VMs) that you run it on.

Is Open WebUI easy to install?

Yes. With Docker, Open WebUI can be deployed in minutes. The container-based setup makes it simple to reproduce, upgrade or tear down environments.

What is Open WebUI used for?

Open WebUI is commonly used for:

  • Running local LLMs with a ChatGPT-like interface

  • Building internal AI tools and demos

  • Testing and comparing models

  • Creating private, offline AI chat systems

Can I use Open WebUI with external APIs?

Yes. Open WebUI supports OpenAI-compatible APIs, allowing you to connect external inference endpoints alongside local models.

Why should I deploy Open WebUI on Hyperstack?

Deploying Open WebUI on Hyperstack GPU infrastructure allows you to combine a powerful UI with high-performance compute:

  • Use high-end on-demand GPUs like NVIDIA A100, NVIDIA H100 and more

  • Deploy on pre-configured environments with one click

  • Scale resources as your usage grows

  • Pay only for the resources you use

Subscribe to Hyperstack!

Enter your email to get updates to your inbox every week

Get Started

Ready to build the next big thing in AI?

Sign up now
Talk to an expert

Share On Social Media

12 Dec 2025

If you’re looking to run Devstral 2 efficiently on cloud GPUs, this tutorial shows you ...

11 Dec 2025

Ever spent hours training a model only to wonder if it actually gets better? If you’ve ...