NVIDIA AI Enterprise is an end-to-end, cloud-native suite of AI and data analytics software optimized, certified, and supported by NVIDIA to run on VMware vSphere with NVIDIA-Certified Systems™. it includes key NVIDIA technologies for rapidly deploying, managing and scaling AI workloads in the modern hybrid cloud.

Optimized for AI

Proven open source frameworks and tools optimize AI development and deployment. Companies can focus on creating value through AI.

Certified for execution in any place

NVIDIA AI Enterprise is certified to run on GPU-accelerated mainstream systems, CPU-only servers, or in the public cloud (NVIDIA-Certified Systems™). With NVIDIA AI Enterprise, AI projects can be ported throughout today's increasingly hybrid data center.

VMware + Red Hat + NVIDIA AI CAPABLE PLATFORM

NVIDIA AI Enterprise Schichtmodell

NVIDIA EGX platform

One architecture. For every enterprise workload.
discover the platform that unifies the data center and brings accelerated computing to every enterprise.
LEARN MORE

ADVANTAGES

OPTIMIZED FOR PERFORMANCE

Achieve near bare-metal performance across multiple nodes to enable large, complex training and machine learning workloads.

CERTIFIED FOR RED HAT AND VMWARE VSPHERE

Reduce deployment risk with a full suite of NVIDIA AI software certified for the VMware and Red Hat data center.

NVIDIA ENTERPRISE SUPPORT

Ensure mission-critical AI projects stay on track with enterprise-level support from NVIDIA.

FLEXIBILITY THROUGH RED HAT OPENSHIFT

This joint solution from NVIDIA and Red Hats gives data scientists the flexibility to use ML tools in containers to quickly create, scale, replicate and share their ML modeling results.

KI READY FOR ANY WORKLOAD

With NVIDIA RAPIDS™, organizations achieve up to 70x faster performance and up to 20x better cost efficiency than comparable CPU-only configurations. This allows you to deploy AI on existing infrastructure with the support of NVIDIA AI experts.

SUPPORT FOR NEW NVIDIA HARDWARE

Support for new NVIDIA hardware, including the NVIDIA A100X and A30X converged accelerators, which enable faster, more efficient, and more secure AI systems, and the NVIDIA A2 GPU for space-constrained environments.

NEW UPDATED NVIDIA AI CONTAINERS

NVIDIA TAO Toolkit and the updated Triton Inference Server with FIL, further streamline AI development and deployment. The NVIDIA TAO Toolkit accelerates AI development by 10x without requiring AI expertise, and the updated NVIDIA Triton Inference Server now supports a FIL (Forest Inference Library) backend that provides the best inference performance for both neural networks and tree-based models on GPUs, enabling simplified deployment of large tree models on GPUs with low latency and high accuracy. NVIDIA® TensorRT™-based applications are up to 40 times faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks.

DEPLOYMENT OF KI IN THE CLOUD

Support for NVIDIA AI Enterprise based Virtual Machine Images (VMIs) for use in Cloud Service Provider (CSP) infrastructure. Customers who have purchased NVIDIA AI Enterprise software can now deploy to specific NVIDIA GPU-accelerated cloud instances in AWS, Azure, or GCP instances with full support from NVIDIA.
Terminology
NOTICE
MEANING
Support Services
Includes technical support, upgrade and maintenance.
Unlimited license
A non-expiring, permanent software license that can be used in perpetuity without the need to renew. Support Services are required and are available in three- or five-year increments. One-year support services are available for renewals only.
Subscription
A software license that is active for a specified period of time defined by the terms of the subscription. A subscription includes support services for the duration of the subscription term.
License server
An application that manages licensing and is installed on a physical or virtual server.
GPU
Graphics processing unit
CPU
The central processing unit (CPU) of a computer is the part of the computer that retrieves and executes instructions.
CPU socket licensing
(1) For on-premise deployments, the number of physical processors in the computing environment on which NVIDIA AI Enterprise is installed, or (2) in a cloud computing environment, the number of virtual CPUs for the computing instance on which NVIDIA AI Enterprise is running. NVIDIA requires one license per CPU socket.
NVIDIA Virtual GPU Software Licensed Products
PRODUCT
DESCRIPTION
NVIDIA Virtual Applications (vApps)l
For organizations using Citrix virtual apps and desktops
,RDSH or other app streaming or session-based solutions.
Designed for PC-level applications and server-based desktops.

NVIDIA Virtual PC (vPC)
For users who want a virtual desktop, but a great PC Windows applications
, browsers and high-definition video


.

NVIDIA RTX™ Virtual Workstation (vWS)
For users who want to be able to
use
professional remote graphics applications
with full performance on any device,
anywhere

NVIDIA Virtual Compute Server (vCS)
For compute-intensive server workloads, such as artificial intelligence
(AI), deep learning, or high-performance computing
(HPC)





Supported NVIDIA GPUs optimized for compute loads
NVIDIA HGX A100
NVIDIA A100
NVIDIA A30
Recommended use case
Optimized calculations
Optimized calculations
Optimized calculations
Number of GPUs
4 NVIDIA A100/ 8 NVIDIA A100
1 NVIDIA A100
1 NVIDIA A100
FP32 cores / GPU
6,912
6,912
3584
Tensor cores / GPURT cores
432
432
224
RT Cores
-
-
-
Total memory size / GPU
40 GB HBM2/80GB HBM2
40 GB HBM2/80GB HBM2
24 GB HBM2
MIG Instances/GPU
7
7
4
Max GPU Power / GPU
400 W
250W/300W
165 W
Form factor
4x SXM4 GPUs/8x SXM4
PCIe 4.0 Dual-Slot FHFL
PCIe 4.0 Dual-Slot FHFL
Card dimensions
-
10.5" × 4.4"
10.5" × 4.4"
Cooling solution
Passive
Passive
Passive
Our products
HGX A100 - SERVER SYSTEMSA100 - SERVER SYSTEMSNVIDIA A30 GRAPHICS CARD
Supported NVIDIA GPUs optimized for mixed workloads
NVIDIA HGX 100
NVIDIA A10
Recommended use case
Optimized calculations
NVIDIA vWS - Performance Optimized (midrange) vCS - Compute Optimized
Number of GPUs
4 NVIDIA A100/ 8 NVIDIA A100
1 NVIDIA A10
FP32 cores / GPU
6,912
9,216
Tensor cores / GPURT cores
432
288
RT Cores
-
72
Total memory size / GPU
40 GB HBM2/80GB HBM2
24 GB GDDR6
Max GPU Power / GPU
400 W
150 W
Form factor
4x SXM4 GPUs/8x SXM4
PCIe 4.0 Single Slot FHFL
Card dimensions
-
10.5" × 4.4"
Cooling solution
Passive
Passive
Our products
HGX A100 -
SERVER SYSTEMS
NVIDIA A10 GRAPHICS CARDS
General information on procurement
ELIGIBILITY
NVIDIA vGPU PRODUCTION SUMS
Maintenance
Access to all maintenance releases, bug fixes, and security patches
for flexible upgrades in accordance with the NVIDIA Virtual
GPU Software Lifecycle Policy.

Upgrades
Access to all new major versions with feature enhancementsAdditions
and new hardware support
Long-term maintenance of branch offices
Available for up to 3 years from general availability per theNVIDIA
Virtual GPU Software Lifecycle Policy
Direct support
Direct access to NVIDIA support for timely resolution of
customer-specific issues
Support availability
Customer support available during normal business hours Cases24
× 7 accepted



Access to the knowledgebase

Web support
Email support
Telephone support