Super Micro‘s new 4U server supports eight NVIDIA HGX A100 GPUs to power applications from the Edge to the cloud.
“Supermicro has introduced a new 4U system, with a NVIDIA HGX A100 8-GPU baseboard (formerly named Delta), that delivers 6x AI training performance and 7x inference workload capacity when compared to current systems,” said Charles Liang, CEO and president of Supermicro.
“Also, the recently announced NVIDIA HGX A100 4-GPU board (formerly named Redstone) is showing wide market acceptance, and we are excited by the ongoing global customer engagement. These new Supermicro systems significantly boost overall performance for accelerated workloads required for rapidly changing markets, including HPC, data analytics, deep learning training, and inference.”
Featuring Supermicro’s optimised thermal design, including custom heatsinks and optional liquid cooling, the latest high-density 2U and 4U servers pack NVIDIA HGX A100 4-GPU 8-GPU baseboards, along with a new 4U server supporting eight NVIDIA A100 PCI-E GPUs. The I/O Module (AIOM) further enhances networking communication with high flexibility. It can be coupled with high-speed, low latency PCI-E 4.0 storage and networking devices that support NVIDIA GPUDirect RDMA and GPUDirect Storage with NVME over Fabrics (NVMe-oF) on NVIDIA Mellanox InfiniBand that feeds the expandable multi-GPU system with a continuous stream of data flow without bottlenecks. In addition, Supermicro’s Titanium Level power supplies add good green credentials to the system with a high efficiency rating of 96%, while allowing redundant support for the GPUs.
“Supermicro systems powered by the NVIDIA A100 can quickly scale to thousands of GPUs or, using new multi-instance GPU technology, each A100 GPU can be partitioned into seven isolated GPU instances to run different jobs,” said Paresh Kharya, senior director of product management and marketing at NVIDIA. “NVIDIA A100 Tensor Core GPUs with TensorFloat 32 provides up to 20 times more compute capacity compared to the previous generation without requiring any code changes.”
Advanced 2U design for HGX A100 4-GPU
This 2U system features the NVIDIA HGX A100 4-GPU baseboard with Supermicro’s advanced thermal heatsink design to maintain the optimum system temperature under full load, all in a compact form factor. The system enables high GPU peer-to-peer communication via NVIDIA NVLink, up to 8TB of DDR4 3200Mhz system memory, five PCI-E 4.0 I/O slots supporting GPUDirect RDMA as well as allowing four hot-swappable NVMe with GPUDirect Storage capability.
Dense, versatile 4U design with HGX A100 8-GPU
The new 4U GPU system features the NVIDIA HGX A100 8-GPU baseboard, up to six NVMe U.2 and two NVMe M.2, ten PCI-E 4.0 x16 slots, with Supermicro’s AIOM support enhancing the 8-GPU communication and data flow between systems through the latest technology stacks such as GPUDirect RDMA, GPUDirect Storage, and NVMe-oF on InfiniBand. The system uses NVIDIA NVLink and NVSwitch technology. It is ideal for large-scale deep learning training, neural network model applications for research or national laboratories, supercomputing clusters, and HPC cloud services.
Hyper-dense, energy-efficient 8U SuperBlade with 20 A100 PCI-E GPUs
The density GPU blade server can support up to 20 nodes and 40 GPUs with two single-width GPUs per node, or one NVIDIA Tensor Core A100 PCI-E GPU per node in Supermicro’s 8U SuperBlade enclosure. The 20 NVIDIA A100 GPUs in 8U elevates the density of computing power in a smaller footprint allowing customers to save on TCO. To support the GPU-optimised configuration and sustain the top performance and throughput needed for demanding AI applications, the SuperBlade provides a 100% non-blocking HDR 200Gb/s InfiniBand networking infrastructure to accelerate deep learning and enable real-time analysis and decision making.
Read the latest edition of PCR’s monthly magazine below: