<img alt="" src="https://secure.insightful-enterprise-intelligence.com/783141.png" style="display:none;">

NVIDIA H100 SXMs On-Demand at $3.00/hour - Reserve from just $2.10/hour. Reserve here

Deploy 8 to 16,384 NVIDIA H100 SXM GPUs on the AI Supercloud. Learn More

|

Published on 29 Aug 2024

Deploying and Using Flux.1 on Hyperstack: A Quick Start Guide

TABLE OF CONTENTS

updated

Updated: 10 Dec 2024

NVIDIA H100 GPUs On-Demand

Sign up/Login

Deploying advanced AI models like FLUX.1 on Hyperstack provides the perfect environment for running large-scale, high-performance applications. This guide will walk you through the process of deploying the FLUX.1 model, specifically focusing on the ''schnell' variant, which is optimised for speed. For those seeking higher-quality outputs, we'll also cover how to access the dev variant by modifying the deployment process.

Deployment Process

Now, let's walk through the step-by-step process of deploying FLUX.1 on Hyperstack.

Step 1: Accessing Hyperstack

  1. Visit the Hyperstack website and log in to your account.
  2. If you don't already have an account, you'll need to create one and set up your billing information. Check our documentation to get started with Hyperstack.
  3. Once logged in, you'll enter the Hyperstack dashboard, which provides an overview of your resources and deployments.

Step 2: Deploying a New Virtual Machine

Initiate Deployment

  1. Navigate to the "Virtual Machines" section and click "Deploy New Virtual Machine."
  2. Click it to start the deployment process.

screenshot-edited

Select Hardware Configuration

  1. In the hardware options, choose the H100-PCIe GPU configuration for optimal performance. This configuration ensures that your deployment will handle the FLUX.1 model efficiently, even for high-demand tasks.

image (11)

Choose the Operating System

  1. Select the "Server 22.04 LTS R535 CUDA 12.2".
  2. This image comes pre-installed with Ubuntu 22.04 LTS and NVIDIA drivers (R535) along with CUDA 12.2, providing an optimised environment for AI workloads.

Select a Keypair

  1. Select one of the keypairs in your account. Don't have a keypair yet? See our Getting Started tutorial for creating one.

Network Configuration

  1. Ensure you assign a Public IP to your Virtual machine.
  2. This allows you to access your VM from the internet, which is crucial for remote management and API access.

Enable SSH Access

  1. Make sure to enable an SSH connection.
  2. You'll need this to securely connect and manage your VM.

Add Firewall Rules

  1. Open port "7680" to allow incoming traffic on this port. This port will be used to interact with the FLUX.1 model via the Gradio demo.

Please note: This will open your port to the public internet, allowing anyone with the public IP address and port number to access the dashboard. In case you don't want this, please refer to the section below at "Interacting with Flux.1"

open port 7860

Configure Additional Settings

  1. Look for the "Configure Additional Settings" section and click on it.
  2. Here, you'll find a field for cloud-init scripts. Use the provided cloud-init script for deployment and ensure it is in bash syntax. This script is pre-configured to deploy the "schnell" model by default, which is optimised for speed. Click here to get the cloud-init script! 
  3. If you want to request access to the gated Dev model on HuggingFace (see screenshot below)
Here's the relevant section from the cloud-init script:
FLUX_MODEL_NAME="flux-schnell"
# FLUX_MODEL_NAME="flux-dev"
# HF_TOKEN="[insert HuggingFace token]"


Gated model agreement

Review and Deploy the Script

  1. Double-check all your settings.
  2. Paste the cloud-init script into the initialisation section when deploying your VM. The script will automatically install the necessary dependencies, clone the FLUX repository, and set up the environment.
  3. Click the "Deploy" button to launch your virtual machine.

Step 3: Setting Up the FLUX.1 Model

  1. After deploying your VM, the cloud-init script will begin its work. This process should take about 5-10 minutes.

Step 4: Accessing Your VM

Once the initialisation is complete, you can access your VM:

Locate SSH Details

  1. In the Hyperstack dashboard, find your VM's details.
  2. Look for the public IP address, which you will need to connect to your VM with SSH.

Connect via SSH

  1. Open a terminal on your local machine.
  2. Use the command ssh -i [path_to_ssh_key] [os_username]@[vm_ip_address] (e.g: ssh -i /users/username/downloads/keypair_hyperstack ubuntu@0.0.0.0.0)
  3. Replace username and ip_address with the details provided by Hyperstack.

Interacting with Flux.1 

Once the deployment is complete, access the Gradio demo interface by navigating to [public-ip]:7860 in your web browser.

Please note: This link will be accessible to anyone with the link. To restrict access, you can disable the Public IP and use SSH port forwarding instead.

  1. Disable public-ip
  2. Use SSH port forwarding together with your keypair, with this command:
    ssh -i [path_to_ssh_key] -L 7860:localhost:7860 [os_username]@[vm_ip_address] # e.g: ssh -i /users/username/downloads/keypair_hyperstack -L 7860:localhost:7860 ubuntu@0.0.0.0
  3. After running the above command, go to localhost:7860 in your browser to access the demo.

Step 5: Hibernating Your VM

When you're finished with your current workload, you can hibernate your VM to avoid incurring unnecessary costs:

  1. In the Hyperstack dashboard, locate your Virtual machine.
  2. Look for a "Hibernate" option.
  3. Click to hibernate the VM, which will stop billing for compute resources while preserving your setup.

To continue your work without repeating the setup process:

  1. Return to the Hyperstack dashboard and find your hibernated VM.
  2. Select the "Resume" or "Start" option.
  3. Wait a few moments for the VM to become active.
  4. Reconnect via SSH using the same credentials as before.


Explore our latest tutorials:

Subscribe to Hyperstack!

Enter your email to get updates to your inbox every week

Get Started

Ready to build the next big thing in AI?

Sign up now
Talk to an expert

Share On Social Media

Hyperstack - Tutorials link

29 Nov 2024

The Hyperstack LLM Inference Toolkit is an open-source tool designed to simplify the ...

Hyperstack - Tutorials link

25 Nov 2024

Adopting open source across every industry is no longer an “if”, but a “what, when and ...

Hyperstack - Tutorials link

20 Nov 2024

Mistral AI’s Pixtral Large is a cutting-edge 123B multimodal model that excels in image ...