Due to the widening chip crisis and the resulting, significant price increases of the major IT manufacturers, online price calculations are currently not possible. We therefore point out that price inquiries via our website may differ from the final offer!
These cookies are necessary for the basic functions of the shop.
"Allow all cookies" cookie
"Decline all cookies" cookie
Matches with only "firstvisit"
These cookies are used to make the shopping experience even more appealing, for example for the recognition of the visitor.
Statistics & Tracking
Track device being used
NVIDIA DGX™ A100
First AI system built on new NVIDIA A100 GPU
8x NVIDIA A100 Tensor Core GPUs, which deliver unmatched acceleration
Tensor Float 32 (TF32) for 20x higher FLOPS
8x NVIDIA A100 80 GB GPUs
Available with up to 640 gigabytes (GB) of total GPU memory
Every business needs to transform itself using artificial intelligence (AI), not just to survive, but to thrive in challenging times. However, the enterprise needs an AI infrastructure platform that improves on traditional approaches that in the past involved slow computing architectures separated by analytics, training and inference workloads. The old approach introduced complexity, drove up costs, limited scalability and was not ready for modern AI. Enterprises, developers, data scientists and researchers need a new platform that unifies all AI workloads, simplifies infrastructure and accelerates ROI.
Editable editable, click me for edit, editable, click me for edit, editable, click me for edit ...
This could be a description of a product, maybe you want to have that but it will be shortend to a certain amount of charachters ...
THE UNIVERSAL SYSTEM FOR ALL KI WORKLOADS
NVIDIA DGX™ A100 is the universal system for all AI workloads - from analytics to training to inference. The DGX A100 sets new standards for compute density, packing 5 petaFLOPS of AI performance into a 6U form factor and replacing legacy compute infrastructure with a single, unified system. The DGX A100 also offers the unprecedented ability to allocate compute power in a fine-grained manner by leveraging the Multi-Instance GPU (MIG) feature of the NVIDIA A100 Tensor Core GPU, which allows administrators to allocate resources that are right-sized for specific workloads.
The DGX A100 is available with up to 640 gigabytes (GB) of total GPU memory, which increases performance by up to three times for large training jobs and doubles the size of MIG instances. This allows the DGX A100 to handle the largest and most complex jobs as well as the simplest and smallest. The DGX A100 runs the DGX software stack and optimised software from NGC. The combination of dense compute performance and complete workload flexibility makes the DGX A100 an ideal choice for both single-node deployments and large-scale Slurm and Kubernetes clusters deployed with NVIDIA DeepOps.
IT DEPLOYMENT SERVICES AVAILABLE
Want to shorten time to insights and accelerate ROI from AI? Let our professional IT team accelerate, deploy and integrate the world's first 5 petaFLOPS AI system, NVIDIA® DGX™ A100, seamlessly and non-disruptively into your infrastructure with 24/7 support.
Get the results and outcomes you need:
Site analysis, readiness, pre-testing and staging.
Access to dedicated engineers, solution architects and support technicians.
Deployment planning, scheduling and project management.
Shipping, logistics management and inventory provisioning.
On-site installation, on-site or remote software configuration.
Post-deployment check-up, support, ticketing and maintenance.
Lifecycle management including design, upgrades, recovery, repair and disposal.
Rack and stack and integration services and multi-site deployment.
Tailored break-fix and managed service contracts.
NVIDIA A100 Tensor Core GPU
8X NVIDIA A100 GPUS WITH UP TO 640 GB TOTAL GPU MEMORY 12 NVLinks/GPU, 600 GB/s GPU-to-GPU Bi-directonal Bandwidth
6X NVIDIA NVSWITCHES 4.8 TB/s Bi-directional Bandwidth, 2X More than Previous Generation NVSwitch
Multi-Instanz-GPU (MIG) The eight A100 GPUs in the DGX A100 can be configured into up to 56 GPU instances, each fully isolated with its own memory, cache and high-bandwidth processing cores.
NVLink & NVSwitch Third-generation NVIDIA® NVLink™ doubles direct GPU-to-GPU bandwidth to 600 GB/s, which is nearly 10 times faster than PCIe Gen4 and next-generation NVIDIA NVSwitch™, which is 2 times faster than the previous generation.
Infiniband The new Mellanox ConnectX-6 VPI HDR InfiniBand/Ethernet adapters run at 200 gigabits per second (Gb/s) to create a high-speed fabric for large AI workloads.
Optimised software stack Integrated DGX software stack, including an AI-tuned base operating system, all required system software and GPU-accelerated applications, pre-trained models and more.
Integrated safety The most robust security for AI implementations with a multi-layered approach spanning the Baseboard Management Controller (BMC), CPU board, GPU board, self-encrypted drives and secure boot.
ESSENTIAL BUILDING BLOCK OF THE KI DATA CENTRE
THE UNIVERSAL SYSTEM FOR EVERY KI WORKLOAD
A universal building block to run any workload at any time - from analytics to AI training to inferencing. from analytics to AI training to inferencing.
INTEGRATED ACCESS TO KI EXPERTISE
Accelerate AI transformation with NVIDIA DGXperts, to maximise the value of your DGX investment.
GAME-CHANGING PERFORMANCE FOR INNOVATORS
Provides unprecedented acceleration with eight A100 GPUs and is fully optimised for NVIDIA CUDA-X™ software and the end-to-end NVIDIA data centre solution stack.
UNRIVALLED DATA CENTRE SCALABILITY
Build a world-class AI infrastructure that is scalable to keep pace with demand.