Supermicro NVIDIA MGX™ Systems

Ultimate building blocks for accelerating data centers

Accelerating the market launch of generative AI and beyond

Build customized AI solutions with Supermicro NVIDIA MGX™ systems equipped with the latest NVIDIA GH200 Grace Hopper™ superchip and NVIDIA Grace™ CPU superchip. The new modular architecture is designed to standardize AI infrastructure and accelerated computing in compact 1U and 2U form factors while providing ultimate flexibility and expandability for current and future GPUs, DPUs and CPUs.

Supermicro's advanced liquid cooling technology enables high-density, energy-efficient configurations, such as a 1U 2-node system with two NVIDIA GH200 Grace Hopper superchips, each consisting of a single NVIDIA H100 Tensor Core GPU and an NVIDIA Grace CPU, integrated with a high-speed interconnect. Supermicro ships thousands of rack-scale AI clusters every month from facilities around the world, ensuring plug-and-play compatibility.

Notification of availability

Technical specifications

Supermicro's 1U NVIDIA MGX systems feature up to 2 NVIDIA GH200 Grace Hopper superchips with 2 NVIDIA H100 GPUs and 2 NVIDIA Grace CPUs, each with 480GB of LPDDR5X memory for the CPU and 96GB of HBM3 or 144GB of HBM3e memory for the GPU.the high-bandwidth, low-latency, memory-coherent NVIDIA C2C connects CPU, GPU and memory at 900 GB/s - seven times faster than PCIe 5.0. The modular architecture offers multiple PCIe 5.0 x16 FHFL slots to accommodate DPUs for cloud and data management as well as expandability for additional GPUs, networking and storage.

With the 1U-2 node design featuring 2 NVIDIA GH200 Grace Hopper superchips, Supermicro's proven direct-to-chip liquid cooling solutions can reduce operating costs by more than 40% while increasing compute density and simplifying rack-scale deployment for Large Language Model (LLM). Cluster and HPC applications.

The 2U Supermicro NVIDIA MGX platform supports both NVIDIA Grace and x86 CPUs with up to 4 full-size data center GPUs, such as NVIDIA H100 PCIe, H100 NVL or L40S.it also offers three additional PCIe 5.0 x16 slots for I/O connectivity and eight hot-swap EDSFF storage bays.

Supermicro offers NVIDIA networking to secure and accelerate AI workloads on the MGX platform.this includes a combination of NVIDIA BlueField-3 DPUs, which provide 2x 200 Gbps connectivity to accelerate user-to-cloud and data storage access, and ConnectX-7 adapters, which provide up to 400 Gbps InfiniBand or Ethernet connectivity between GPU servers.

Developers can quickly leverage these new systems and NVIDIA software offerings for any industry workload.these offerings include NVIDIA AI Enterprise, enterprise-class software that powers the NVIDIA AI platform and optimizes the development and deployment of production-ready generative AI, computer vision, speech AI and more. In addition, the NVIDIA HPC software development kit provides the essential tools needed to advance scientific computing.
Every aspect of Supermicro NVIDIA MGX systems is designed to increase efficiency, from intelligent thermal design to component selection.nVIDIA Grace superchip CPUs have 144 cores and deliver up to twice the performance per watt compared to today's industry-standard x86 CPUs. Specific Supermicro NVIDIA MGX systems can be configured with two nodes in 1U, for a total of 288 cores on two Grace CPU superchips to deliver breakthrough compute densities and energy efficiency in hyperscale and edge data centers.

1U Grace Hopper MGX Systems

CPU+GPU Coherent Memory System for AI and HPC Applications
  • Up to 2 NVIDIA GH200 Grace Hopper Superchips featuring 72-core ARM CPU and H100 Tensor Core GPU tightly coupled with coherent memory
  • Up to 96GB HBM3 and 480GB LPDDR5X integrated memory per Grace Hopper Superchip
  • NVLink® Chip-2-Chip (C2C) high-bandwidth and low-latency interconnect
  • Up to 3 PCIe 5.0 x16 slots supporting NVIDIA BlueField®-3, NVIDIA ConnectX®-7 or additional GPUs
  • 8 hot-swap E1.S and 2 M.2 slots
  • Air Cooling and liquid cooling options
Download Datasheet

1U Grace Hopper MGX system configurations at a glance

Build new solutions for accelerated infrastructures that enable scientists and engineers to focus on solving the world's most important problems with larger data sets, more complex models, and new generative AI workloads. In the same 1U chassis, Supermicro's dual NVIDIA GH200 Grace Hopper superchip systems deliver the highest performance for any application on the CUDA platform with significant speedups for AI workloads with high memory requirements. In addition to hosting up to 2 onboard H100 GPUs in a 1U form factor, the modular bays enable full-size PCIe expansion for current and future accelerated compute components, high-speed scale-out and clustering.
SKU
Form Factor
​​​​​​​
CPU
​​​​​​​
GPU
​​​​​​​
Memory
​​​​​​​
Drives
Networking
​​​​​​​
Interconnect
​​​​​​​
Cooling
Power
​​​​​​​
ARS-111GL-NHR
1U system with single NVIDIA Grace Hopper Superchip (air-cooled)
72-core Grace Arm Neoverse V2 CPU + H100 Tensor Core GPU in a single chip
NVIDIA H100 Tensor Core GPU with 96GB of HBM3 or 144GB of HBM3e (coming soon)
Up to 480GB of integrated LPDDR5X with ECC
(Up to 480GB + 144GB of fast-access memory)
8x Hot-swap E1.S drives and 2x M.2 NVMe drives
3x PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or ConnectX-7
NVLink-C2C with 900GB/s for CPU-GPU interconnect
Air-cooling
2x 2000W Redundant Titanium Level power supplies
ARS-111GL-NHR-LCC
1U system with single NVIDIA Grace Hopper Superchip (liquid-cooled)
72-core Grace Arm Neoverse V2 CPU + H100 Tensor Core GPU in a single chip
NVIDIA H100 Tensor Core GPU with 96GB of HBM3 or 144GB of HBM3e (coming soon)
Up to 480GB of integrated LPDDR5X memory with ECC (Up to 480GB + 144GB of fast-access memory)
8x Hot-swap E1.S drives and 2x M.2 NVMe drives
3x PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or ConnectX-7
NVLink-C2C with 900GB/s for CPU-GPU interconnect
Liquid-cooling
2x 2000W Redundant Titanium Level power supplies
ARS-111GL-DNHR-LCC
1U 2-node system with NVIDIA Grace Hopper Superchip per node (liquid-cooled)
2x 72-core Grace Arm Neoverse V2 CPU + H100 Tensor Core GPU in a single chip (1 per node)
NVIDIA H100 Tensor Core GPU with 96GB of HBM3 or 144GB of HBM3e per node (coming soon)
Up to 480GB of LPDDR5X per node (Up to 480GB + 144GB of fast-access memory per node)
8x Hot-swap E1.S drives and 2x M.2 NVMe drives
2x PCIe 5.0 x16 slots per node supporting NVIDIA BlueField-3 or ConnectX-7
NVLink-C2C with 900GB/s for CPU-GPU interconnect
Liquid-cooling
2x 2700W Redundant Titanium Level power supplies

Grace and x86 MGX Systems

Modular Building Block Platform Supporting today’s and Future GPUs, CPUs, and DPUs
  • NVIDIA Grace™ CPU Superchip (144-core) or 4th Gen Intel® Xeon® Scalable processor
  • Up to 480GB integrated LPDDR5X DRAM memory with ECC and up to 1TB/s of bandwidth (with Grace CPU Superchip) or up to 2TB 4800MT/s ECC DDR5 DRAM (with Intel CPU)
  • Up to 4 NVIDIA double-width PCIe GPUs, including H100 PCIe, H100 NVL, L40S, and more
  • Up to 8 Hot-Swap E1.S and 2 M.2 NVMe slots
  • Up to 3 PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or NVIDIA ConnectX-7
Download Datasheet

Grace and x86 MGX system configurations at a glance

Supermicro NVIDIA MGX™ 1U/2U systems with Grace™ CPU Superchip and x86 CPUs are fully optimized to support up to 4 GPUs over PCIe without compromising I/O networking or heat dissipation. The ultimate building block architecture allows you to optimize these systems for a variety of accelerated workloads and domains, including AI training and inference, HPC, data analytics, visualization/Omniverse™ and hyperscale cloud applications.
SKU
Form Factor
​​​​​​​
CPU
​​​​​​​
GPU
​​​​​​​
Memory
​​​​​​​
Drives
​​​​​​​
Networking
​​​​​​​
Interconnect
​​​​​​​
Cooling
Power
​​​​​​​
ARS-121L-DNR
1U 2-node system with NVIDIA Grace CPU Superchip per node
144-core Grace Arm Neoverse V2 CPU in a single chip per node (total of 288 cores in one system)
Please contact our sales for possible configurations
Up to 480GB of integrated LPDDR5X memory with ECC and up to 1TB/s of bandwidth per node
Up to 4x hot-swap E1.S drives and 2x M.2 NVMe drives per node
2x PCIe 5.0 x16 slots per node supporting NVIDIA BlueField-3 or ConnectX-7 (e.g., 1 GPU and 1 BlueField-3)
NVLink™-C2C with 900GB/s for CPU-CPU interconnect (within node)
Air-cooling
2x 2700W Redundant Titanium Level power supplies
ARS-221GL-NR
2U GPU system with single NVIDIA Grace CPU Superchip
144-core Grace Arm Neoverse V2 CPU in a single chip
Up to 4 double-width GPUs including NVIDIA H100 PCIe, H100 NVL, L40S
Up to 480GB of integrated LPDDR5X memory with ECC and up to 1TB/s of bandwidth per node
Up to 8x hot-swap E1.S drives and 2x M.2 NVMe drives
3x PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or ConnectX-7 (in addition to 4x PCIe 5.0 x16 slots for GPUs)
NVLink Bridge GPU-GPU interconnect supported (e.g., H100 NVL)
Air-cooling
3x 2000W Redundant Titanium Level power supplies
SYS-221GE-NR
2U GPU system with dual x86 CPUs
​​​​​​​
4th Gen Intel Xeon Scalable Processors (Up to 56-core per socket)
Up to 4 double-width GPUs including NVIDIA H100 PCIe, H100 NVL, L40S
Up to 2TB, 32x DIMM slots, ECC DDR5-4800
​​​​​​​
Up to 8x hot-swap E1.S drives and 2x M.2 NVMe drives
3x PCIe 5.0 x16 slots supporting NVIDIA BlueField-3 or ConnectX-7 (in addition to 4x PCIe 5.0 x16 slots for GPUs)
NVLink™ Bridge GPU-GPU interconnect supported (e.g., H100 NVL)
Air-cooling
3x 2000W Redundant Titanium Level power supplies