<img alt="" src="https://secure.insightful-enterprise-intelligence.com/783141.png" style="display:none;">
NVIDIA HGX B100

Advance Generative AI with NVIDIA Blackwell 

Purpose-built to drive the next wave of generative AI innovation, the NVIDIA HGX B100 delivers unparalleled computing power to scale complex AI models. Experience the power of the future today with NVIDIA Blackwell B100.

nvidia-blackwell-b100-product-1

Unrivalled Performance in...

tick_check

Acceleration

Configuration of 8 GPUs, up to 112 petaFLOPs of AI performance

tick_check

Memory

Up to 1.5TB of fast memory  for speedups

tick_check

GPU Bandwidth

Up to 192 GB HBM3e massive memory bandwidth 

tick_check

Networking

5th-gen NVLink with 1.8TB/s of GPU-to-GPU interconnect and InfiniBand networking

Unrivalled Performance in...

Acceleration

Configuration of 8 GPUs, up to 112 petaFLOPs of AI performance

GPU Bandwidth

Up to 192 GB HBM3e massive memory bandwidth 

Memory

Up to 1.5TB of fast memory  for speedups

Networking

5th-gen NVLink with 1.8TB/s of GPU-to-GPU interconnect and InfiniBand networking

Boost LLM with NVIDIA HGX B100

Group 41022

Built with Blackwell Architecture

Introducing the Blackwell-architecture packed with 208 billion transistors and manufactured using TSMC's cutting-edge 4NP process, these GPUs redefine performance. Featuring two interconnected GPU dies forming a unified chip with a 10 TB/second link, the Blackwell GPUs are a true force to be reckoned with. The second-generation Transformer Engine and 4-bit floating-point AI inference support double the compute and model sizes.

Group 41024

Peak Security and Reliablity

The innovative RAS Engine ensures unparalleled reliability, availability, and serviceability, while its AI-based preventative maintenance capabilities maximise system uptime and reduce operating costs. Advanced confidential computing features safeguard your AI models and customer data without sacrificing performance.

Benefits of NVIDIA HGX B100

graphic-card-1

A New Class of AI Superchip

Custom-built 4NP TSMC, this GPU pushes the boundaries of performance. Its  an innovative design consisting of two GPU dies, each intricately connected by 10 TB/second chip-to-chip link. This seamless integration forms a unified GPU to handle the most demanding tasks

DGX-reference-architecture

LLM Innovation

With cutting-edge Blackwell Tensor Core technology, combined with TensorRT-LLM and Nemo Framework innovations, get unmatched acceleration for LLM and Mixture-of-Experts models

Vector

AI Capabilities

Second-generation Transformer Engine introduces groundbreaking innovations, including 8-bit floating-point (FP8) precision and support for new numerical formats. Multi-GPU scaling with the fifth-generation NVLink interconnect with 1.8TB/s of bidirectional GPU-to-GPU bandwidth

vector-2

Reliable AI

The RAS Engine incorporates AI-based preventative maintenance capabilities directly at the chip level for advanced diagnostics and forecasting of potential reliability issues before they occur. It boosts system uptime and maintain seamless operations for your large-scale AI deployments

Group 88

Data Security

Advanced confidential computing capabilitieS to secure AI models and customer data without compromising performance. New native interface encryption protocols for data-sensitive industries 

Group 41022-1

Data Acceleration

Accelerate database queries and experience the highest performance in data analytics and data science with a dedicated decompression engine. Supporting the latest formats, this engine allows you to tackle vast amounts of data 

AI Innovation

FP4 precision with 14 petaFLOPS delivers a massive leap in AI performance to lead in innovation. 

1

Ultra-fast NVLink 

The NVIDIA HGX B100 integrates powerful Blackwell B100 GPUs with ultra-fast NVLink interconnects for accelerated computing and generative AI workloads. 

2

Massive Memory

192 GB HBM3e of memory to handle the largest AI models and datasets for seamless scalability.

3

Technical Specifications

GPU: NVIDIA HGX B100

GPU Memory: 192 GB HBM3e

Power: 700W

FP4 Tensor Core
14 petaFLOPS
FP8/FP6 Tensor Core
7 petaFLOPS
INT8 Tensor Core
7 petaOPs
FP16/BF16 Tensor Core
3.5 petaFLOPS
TF32 Tensor Core
1.8 petaFLOPS
FP64 Tensor Core
30 teraFLOPS
GPU memory
Up to 192 GB HBM3e
Bandwidth
Up to 8 TB/s
Multi-Instance GPU (MIG)
7
Decompression Engine
Yes
Decoders
2x 7 NVDEC, 2x 7 NVJPEG
Power
Up to 700W
Interconnect
5th Generation NVLink: 1.8TB/s, PCIe Gen6: 256GB/s
FP4 Tensor Core
14 petaFLOPS
FP8/FP6 Tensor Core
7 petaFLOPS
INT8 Tensor Core
7 petaOPs
FP16/BF16 Tensor Core
3.5 petaFLOPS
TF32 Tensor Core
1.8 petaFLOPS
FP64 Tensor Core
30 teraFLOPS
GPU memory
Up to 192 GB HBM3e
Bandwidth
Up to 8 TB/s
Multi-Instance GPU (MIG)
7
Decompression Engine
Yes
Decoders
2x 7 NVDEC, 2x 7 NVJPEG
Power
Up to 700W
Interconnect
5th Generation NVLink: 1.8TB/s, PCIe Gen6: 256GB/s

New Wave of Gen AI with the NVIDIA HGX B100

New Wave of Gen AI

FAQ

Frequently asked questions about the NVIDIA HGX B100.

How much is the NVIDIA B100 memory?

The NVIDIA B100 GPU has up to 192 GB HBM3e memory.

What is NVIDIA B100 price?

The NVIDIA B100 will be available in Q4 2024, and is available on Hyperstack to reserve here. Our team will then contact you to discuss pricing.

Which applications can take advantage of the NVIDIA HGX B100?

The NVIDIA HGX B100 GPU is well-suited for generative AI and accelerated computing that can help in data processing, electronic design automation, computer-aided engineering and quantum computing.

What are NVIDIA B100 specifications?

The NVIDIA B100 specifications include a transformer engine, RAS engine, confidential computing capabilities, decompression engine and a new class of AI Superchip.

What is the Max thermal design power (TDP) of NVIDIA HGX B100?

The Max thermal design power (TDP) of NVIDIA HGX B100 is 700W.