Supermicro Grows AI Optimized Product Portfolio with a New Generation of Systems and Rack Architectures Featuring New NVIDIA Blackwell Architecture Solutions
San Jose, Calif. at NVIDIA GTC 2024, -- March 18, 2024 – Supermicro, Inc. (NASDAQ: SMCI), a Total IT Solution Provider for AI, Cloud, Storage, and 5G/Edge, is announcing new AI systems for large-scale generative AI featuring NVIDIA's next-generation of data center products, including the latest NVIDIA GB200 Grace™ Blackwell Superchip, the NVIDIA B200 Tensor Core and B100 Tensor Core GPUs. Supermicro is enhancing its current NVIDIA HGX™ H100/H200 8-GPU systems to be drop-in ready for the NVIDIA HGX™ B100 8-GPU and enhanced to support the B200, resulting in a reduced time to delivery. Additionally, Supermicro will further strengthen its broad NVIDIA MGX™ systems lineup with new offerings featuring the NVIDIA GB200, including the NVIDIA GB200 NVL72, a complete rack level solution with 72 NVIDIA Blackwell GPUs. Supermicro is also adding new systems to its lineup, including the 4U NVIDIA HGX B200 8-GPU liquid-cooled system.
"Our focus on building block architecture and rack-scale Total IT for AI has enabled us to design next-generation systems for the enhanced requirements of NVIDIA Blackwell architecture-based GPUs, such as our new 4U liquid-cooled NVIDIA HGX B200 8-GPU based system, as well as our fully integrated direct-to-chip liquid cooled racks with NVIDIA GB200 NVL72," said Charles Liang, president and CEO of Supermicro. "These new products are built upon Supermicro and NVIDIA's proven HGX and MGX system architecture, optimizing for the new capabilities of NVIDIA Blackwell GPUs. Supermicro has the expertise to incorporate 1kW GPUs into a wide range of air-cooled and liquid-cooled systems, as well as the rack scale production capacity of 5,000 racks/month and anticipates being first-to-market in deploying full rack clusters featuring NVIDIA Blackwell GPUs."
Supermicro's direct-to-chip liquid cooling technology will allow for the increased thermal design power (TDP) of the latest GPUs and deliver the full potential of the NVIDIA Blackwell GPUs. Supermicro's HGX and MGX Systems with NVIDIA Blackwell are the building blocks for the future of AI infrastructure and will deliver groundbreaking performance for multi-trillion parameter AI training and real-time AI inference.
A wide range of GPU-optimized Supermicro systems will be ready for the NVIDIA Blackwell B200 and B100 Tensor Core GPU and validated for the latest NVIDIA AI Enterprise software, which adds support for NVIDIA NIM inference microservices. The Supermicro systems include:
- NVIDIA HGX B100 8-GPU and HGX B200 8-GPU systems
- 5U/4U PCIe GPU system with up to 10 GPUs
- SuperBlade® with up to 20 B100 GPUs for 8U enclosures and up to 10 B100 GPUs in 6U enclosures
- 2U Hyper with up to 3 B100 GPUs
- Supermicro 2U x86 MGX systems with up to 4 B100 GPUs
For training massive foundational AI models, Supermicro is prepared to be the first-to-market to release NVIDIA HGX B200 8-GPU and HGX B100 8-GPU systems. These systems feature 8 NVIDIA Blackwell GPUs connected via a high-speed fifth-generation NVIDIA® NVLink® interconnect at 1.8TB/s, doubling the previous generation performance, with 1.5TB total high-bandwidth memory and will deliver 3X faster training results for LLMs, such as the GPT-MoE-1.8T model, compared to the NVIDIA Hopper architecture generation. These systems feature advanced networking to scale to clusters, supporting both NVIDIA Quantum-2 InfiniBand and NVIDIA Spectrum-X Ethernet options with a 1:1 GPU-to-NIC ratio.
"Supermicro continues to bring to market an amazing range of accelerated computing platform servers that are tuned for AI training and inference that can address any need in the market today, said Kaustubh Sanghani, vice president of GPU product management at NVIDIA. "We work closely with Supermicro to bring the most optimized solutions to customers.”
For the most demanding LLM inference workloads, Supermicro is releasing several new MGX systems built with the NVIDIA GB200 Grace Blackwell Superchip, which combines an NVIDIA Grace CPU with two NVIDIA Blackwell GPUs. Supermicro's NVIDIA MGX with GB200 systems will deliver a vast leap in performance for AI inference with up to 30x speed-ups compared to the NVIDIA HGX H100. Supermicro and NVIDIA have developed a rack-scale solution with the NVIDIA GB200 NVL72, connecting 36 Grace CPUs and 72 Blackwell GPUs in a single rack. All 72 GPUs are interconnected with fifth-generation NVIDIA NVLink for GPU-to-GPU communication at 1.8TB/s. In addition, for inference workloads, Supermicro is announcing the ARS-221GL-NHIR, a 2U server based on the GH200 line of products, which will have two GH200 servers connected via a 900Gb/s high speed interconnect. Come to the Supermicro Booth at GTC to learn more.
Supermicro systems will also support the upcoming NVIDIA Quantum-X800 InfiniBand platform, consisting of the NVIDIA Quantum-X800 QM3400 switch and the SuperNIC800, and the NVIDIA Spectrum-X800 Ethernet platform, consisting of the NVIDIA Spectrum-X800 SN5600 switch and the SuperNIC800. Optimized for the NVIDIA Blackwell architecture, the NVIDIA Quantum-X800, and Spectrum-X800 will deliver the highest level of networking performance for AI infrastructures.
For more information on Supermicro NVIDIA solutions, visit https://www.supermicro.com/en/accelerators/nvidia
Supermicro's upcoming systems lineup featuring NVIDIA B200 and GB200 consists of:
- The Supermicro's NVIDIA HGX B200 8-GPU air-cooled and liquid-cooled systems are for the highest generative AI training performance. This system features 8 NVIDIA Blackwell GPUs connected via fifth generation NVLink with a pool of 1.5TB high-bandwidth memory (up to 60TB/s) to speed up AI training workloads.
- Supermicro's best-selling AI Training System, the 4U/8U system with NVIDIA HGX H100/H200 8-GPU, will support NVIDIA's upcoming HGX B100 8-GPU.
- A Supermicro Rack-Level Solution featuring GB200 Superchip systems as server nodes with 2 Grace CPUs and 4 NVIDIA Blackwell GPUs per node. Supermicro's direct-to-chip liquid-cooling maximizes density with 72 GB200 192GB GPUs (1200W TDP per GPU), all in a single 44U ORV3 rack.
Supermicro at GTC 2024
Supermicro will demonstrate a complete portfolio of GPU systems for AI at NVIDIA's GTC 2024 event from March 18-21 at the San Jose Convention Center. Visit Supermicro at booth #1016 to see solutions built for a wide range of AI applications, including training generative AI models, AI inference, and edge AI. Supermicro will also showcase two rack-level solutions, including a concept rack with systems featuring the upcoming NVIDIA GB200 with 72 liquid-cooled GPUs interconnected with fifth-generation NVLink.
Supermicro solutions that will be on display at GTC 2024 include:
- Supermicro liquid-cooled AI training rack featuring 8 4U 8-GPU systems with NVIDIA HGX H200 8-GPUs
- Supermicro concept ORV3 rack with liquid-cooled MGX system nodes, hosting a total of 72 NVIDIA GB200 Superchips connected via fifth-generation NVLink
- Supermicro MGX systems, including the 1U Liquid-Cooled NVIDIA GH200 Grace Hopper Superchip system
- Supermicro short-depth Hyper-E system for delivering GPU computing at the edge
- Supermicro Petascale all-flash storage system for high-performance AI data pipelines