NVIDIA lauches next-gen HGX with B300 GPUs
NVIDIA HGX B300 Server
The NVIDIA HGX B300 is the most advanced server platform engineered to power the next wave of AI and high-performance computing breakthroughs to date.
Based on the Blackwell Ultra architecture, this platform sets a new benchmark in scale, performance, and memory bandwidth for massive AI workloads, large language model (LLM) inference, and real-time simulation.
Why Choose HGX B300?
The HGX B300 is purpose-built for enterprises and research institutions needing massive compute capabilities. From LLMs with hundreds of billions of parameters to real-time AI inference across multimodal tasks, this platform offers:
- Double the memory and GPU interconnect bandwidth over HGX B200
- Massive parallelism for faster time-to-results in AI/ML workflows
- Energy efficiency optimized through 5th-gen NVLink and advanced thermal control
- Futureproof AI infrastructure, ready for GPT-5+ and beyond

Top applications

Scientific simulation & digital twins

AI/ML model training & LLM inference

Cybersecurity & real-time data analysis
HGX B300 vs HGX B200: comparison
Feature | HGX B300 | HGX B200 |
Form Factor | 8x NVIDIA Blackwell Ultra SXM | 8x NVIDIA Blackwell SXM |
FP4 Tensor Core | 144 PFLOPS | 105 PFLOPS | 144 PFLOPS | 72 PFLOPS |
FP8/FP6 Tensor Core | 72 PFLOPS | 72 PFLOPS 72 PFLOPS |
INT8 Tensor Core | 2 POPS | 72 PFLOPS |
FP16/BF16 Tensor Core | 36 PFLOPS | 36 PFLOPS |
TF32 Tensor Core | 18 PFLOPS | 18 PFLOPS |
FP32 | 600 TFLOPS | 600 TFLOPS |
FP64/FP64 Tensor Core | 10 TFLOPS | 296 TFLOPS |
Total Memory | Up to 2.3 TB | 1.4 TB |
NVLink | Fifth generation | Fifth generation |
NVIDIA NVSwitch™ | NVLink 5 Switch | NVLink 5 Switch |
NVSwitch GPU-to-GPU Bandwidth | 1.8 TB/s | 1.8 TB/s |
Total NVLink Bandwidth | 14.4 TB/s | 14.4 TB/s |
Networking Bandwidth | 1.6 TB/s | 0.8 TB/s |
Attention Performance | 2X | 1X |
HGX B300 vs HGX H100: comparison
AI Reasoning Inference
Courtesy of NVIDIA.
HGX B300 achieves up to 11x higher inference performance over the previous H100 generation for models such as Llama 3.1 405B.
AI Training

The second-generation Transformer Engine, featuring 8-bit floating point (FP8) and new precisions, enables a remarkable 4x faster training for large language models like Llama 3.1 405B.
Interested? Get in touch with our sales team! We have servers and configurations available now.
2CRSi is NVIDIA Elite Partner.