SETTING THE STANDARD FOR KI INFRASTRUCTURE

Enterprises undergoing AI transformation need an AI infrastructure platform that improves upon traditional approaches that in the past included slow compute architectures and were separated by analytics, training and inference workloads. The preceding approach increased complexity, drove up costs and limited scalability.

The DGX H100 is an AI powerhouse equipped with the groundbreaking NVIDIA H100 Tensor Core GPU.​​​​​​​

It revolutionizes the enterprise data center with infrastructure that unifies AI and data science workloads.

THE PORTFOLIO OF WORLD-LEADING PURPOSE-BUILT KI SYSTEMS

Nvidia DGX Station Image

AI Workgroup System

Server-class workstations
are ideal for experimentation
and team development.
no data center required.
Nvidia DGX Station Image

AI training, inference, and analytics

A range of server solutions to help you tackle the most complex AI challenges.
Nvidia DGX POD image

Scaled AI infrastructure solution

Industry-standard infrastructure designs for AI companies
REQUEST A QUOTEMORE INFORMATION
Nvidia DGX Superpod Image

Turnkey AI infrastructure

Industry-leading full-cycle infrastructure - the fastest path to AI innovation at scale.
REQUEST A QUOTEMORE INFORMATION

ADVANTAGES

Optimirte Zusatzsoftware Image

With optimized AI software, including an AI-optimized base operating system, you can simplify deployment and be more productive in hours instead of months.

MORE INFOS
Ki Bereich Image

DGX is the core building block of a number ofTOP500 supercomputers and has been adopted by many leading companies.

MORE INFORMATION
Ki Cluster Image

The architecture is proven for multi-node scalability and was developed with industry leaders in storage, compute and networking.

MORE INFORMATION
Ki Expericence Image

NVIDIA DGXperts is a global team of over 14,000 AI-fluent professionals AI professionals who have acquired a wealth of experience over the last decade to help you maximize the value of your DGX investment.

MORE INFORMATION

Game-changing performance: NVIDIA A100 Tensor Core GPU

Game-changing performance based on the NVIDIA A100 GPU, delivering the world's first 5 PFLOPS
AI system that can effortlessly run analysis, training, and inference workloads simultaneously.
The NVIDIA DGX A100 features eight NVIDIA A100 Tensor Core GPUs, providing users with unmatched acceleration, and is fully optimized for NVIDIA CUDA-X™ software and the end-to-end NVIDIA Data Center Solution Stack.

NVIDIA A100 GPUs offer a new precision, TF32, that works like FP32 to deliver a 20X increase in FLOPS performance for AI compared to the previous generation - and best of all, no code changes are required to achieve this speed increase. And when using NVIDIA's automatic mixed-precision, the A100 provides an additional 2x performance boost with just one additional line of code at FP16 precision. The A100 GPU also features class-leading memory bandwidth of 1.6 terabytes per second (TB/s), an increase of more than 70% over the last generation. In addition, the A100 GPU has significantly more on-chip memory, including a 40 MB Level 2 cache that is nearly 7 times larger than the previous generation, maximizing compute performance. The DGX A100 also introduces next-generation NVIDIA NVLink™, which doubles direct GPU-to-GPU bandwidth to 600 gigabytes per second (GB/s), nearly 10X higher than PCIe Gen 4, and a new NVIDIA NVSwitch that is 2X faster than the last generation.

this unprecedented performance delivers the fastest time-to-solution for training, inference, and analytics workloads, enabling users to address challenges that were not possible or practical before.
Vergleichsimage

Unmatched Flexibility: New Multi Instance GPU (MIG) Innovation

Unmatched flexibility with Multi-Instance GPU (MIG) innovation that enables 7x inference performance per GPU and the ability to allocate resources that are right-sized for specific workloads.
MIG partitions a single NVIDIA A100 GPU into up to seven independent GPU instances. These run concurrently, each with its own memory, cache, and streaming multiprocessors. This allows the A100 GPU to provide guaranteed quality-of-service (QoS) at up to 7x higher utilization compared to previous GPUs.
Nvidia Flexibility Image
MIG partitions a single NVIDIA A100 GPU into up to seven independent GPU instances. These run concurrently, each with its own memory, cache, and streaming multiprocessors. This allows the A100 GPU to provide guaranteed quality-of-service (QoS) at up to 7x higher utilization compared to previous GPUs.
MIG Compartmentalizes Image
With the DGX A100, you can use up to 56 MIG slices to solve problems with inflexible infrastructures and accurately allocate compute power to each workload. You no longer have to struggle to divide time on a box among multiple competing projects. With MIG on DGX A100, you have enough compute power to support your entire data science team.
Bert-Large Inference Image

Low total cost of ownership: Universal AI platform

Unparalleled TCO/ROI metrics with all the performance of a modern AI data center at 1/10 the cost, 1/25 the space, and 1/20 the performance
KI Rechenzentrum Image
Today's AI data center
  • 25 Rack for training & inference
  • 630 kW
  • $11M
DGX A100 Rechenzentrum Image
DGX A100 Data Center
  • 1 rack (5 x DGX A100s)
  • 28 kW
  • $1M
Traditional AI infrastructures typically consist of three separate specialized clusters: training (GPU-based), inference (often CPU-based), and analysis (CPU-based). These inflexible infrastructure silos were never designed for the pace of AI. Most data centers dealing with AI workloads will likely find that these resources are either over- or under-utilized at any given time. The DGX A100 data center with MIG provides you with a single system that can flexibly adapt to your workload requirements.
Traditional Infrastructure Image
DGX A100 Infrastructure Image
In many data centers, the demand for computing resources rises and falls, resulting in servers that are mostly underutilized. IT ends up having to buy excess capacity to protect against occasional spikes. With the DGX A100, you can now right-size resources for each job and increase utilization, which lowers TCO.

With DGX A100 data centers, you can easily adapt to changing business needs by deploying a single elastic infrastructure that is much more efficient.
DGX A100 Comparision Image

DGX Station a100: The workgroup appliance for the age of AI

The NVIDIA DGX Station A100 brings AI supercomputing to data science teams, providing data center technology without a data center or additional IT infrastructure. It is the world's only workstation-style system with four fully interconnected NVIDIA A100 Tensor Core GPUs, the same GPUs as the NVIDIA DGX A100 server, interconnected via third-generation, high-bandwidth NVIDIA NVLink. It also uses a top-of-the-line server-grade CPU, super-fast NVMe storage, and state-of-the-art PCIe Gen4 buses.
DGX STATION A100 PERFORMANCE
DGX A100 Image
A KI DEVICE YOU CAN PLACE ANYWHERE
The DGX Station A100 is suitable for use in a standard office environment without special power or cooling and simply plugs into any standard power outlet. It also includes the same Baseboard Management Controller (BMC) as the NVIDIA DGX A100, so system administrators can perform all necessary tasks via a remote connection. And its innovative cooling-based design means it feels cool to the touch.
SUPERCOMPUTING FOR DATA SCIENCE TEAMS
DGX Station A100 can run training, inference, and analytics workloads in parallel, and with MIG can provision up to 28 separate GPU devices for individual users and jobs, so activity is limited and does not impact overall system performance
SUPPORTED BY THE DGX SOFTWARE STACK
The DGX Station A100 shares the same fully optimized NVIDIA DGX™ software stack as all DGX systems, providing maximum performance and full interoperability with DGX-based infrastructure.
Data Science Image

Why nvidia dgx ?

Nearly every company recognizes the importance of AI for true business transformation. In a recent study, 84% of executives surveyed fear they will not achieve their growth goals if they do not scale AI. However, nearly 76% also said they are having trouble scaling AI in their business. Many companies are hindered by the complexity and cost of deploying the right infrastructure. For most organizations, one of the biggest AI challenges is finding AI infrastructure experts and implementing IT-proven platforms that deliver predictable and scalable performance. As a result, more and more enterprises are choosing NVIDIA's highly optimized DGX systems to power their infrastructures and enable their on-prem AI initiatives.
Nvidia Ai Know-How Image
Access to NVIDIA AI know-how
NVIDIA DGX systems come with direct access to NVIDIA AI experts trained on our in-house infrastructure, NVIDIA SATURNV. They have the most extensive track record of field-proven implementations, so customers get the fastest time to resolution.
vollständing Optimiert Image
Fully optimized, field-tested
NVIDIA DGX systems are AI supercomputers with a full-stack solution that integrates innovation and optimization across hardware and software. Thousands of customers have deployed DGX systems to date, including nine of the top 10 government institutions and eight of the top 10 universities in the US.
KI image
Supported by the largest test site for AI
NVIDIA DGX SATURNV is the massive AI infrastructure where NVIDIA's most important work is done. With thousands of DGX AI systems, NVIDIA is constantly finding opportunities for improvements and enhancements that are then rolled out to DGX customers.
Vertrauenswürdige Infrastruktur-Lösungen Image
Trusted infrastructure solutions
Only NVIDIA DGX offers the most comprehensive portfolio of infrastructure solutions built on the most trusted names in data center storage and networking, allowing customers to scale as their business grows.
Einsatz von KI Image
The easiest and fastest way to deploy AI
NVIDIA and its partners offer a full range of options, from DGX-as-a-Service to DGX colocation solutions to full-fledged DGX PODs deployed on-premise to help organizations deploy AI more easily, quickly and cost-effectively.