xcloud Logo
Nvidia Hgx B200

NVIDIA HGX™ B200

High-performance server platform purpose-built for the Blackwell architecture.

The NVIDIA HGX B200 combines eight Blackwell GPUs interconnected via fifth-generation NVLink to provide unprecedented compute density. It is the most powerful single-node solution for generative AI and LLM inference today.

NVIDIA HGX™ B200

Architecture Highlights

15x Inference Perf

Overwhelming advantage in LLM inference compared to previous generations.

1.1TB Memory

Massive HBM3e capacity to support the largest models.

Application Scenarios

Optimized for variety of advanced computational workloads

Generative AI Training

Rapid iteration of GPT-scale and similar large pre-trained models.

Real-time LLM Inference

Serving millions of users with low-latency content generation.

Core Capabilities

2.5 PFLOPS FP8 Performance
144GB HBM3e Memory
1.8TB/s NVLink Bandwidth
Superior Energy Efficiency

Technical Specifications

GPU Quantity8x Blackwell B200
Compute (FP8)36 PFLOPS
Total Memory1152 GB

Need Professional Advice?

Our consulting team is ready to help accelerate your business.