Uvation MarketplaceMarketplace
  • loading

    Help is here whenever you need it.

    Sourcing and Sales

    We leverage our trusted supplier network to source the electronic components you need, exactly when you need them.

    Product Lifecycle

    Protecting your supply chain from disruptions with expert sourcing and management, ensuring continuity through end-of-life and obsolescence.

    Self Service Ordering

    Empowering you with seamless self-service solutions, anytime, anywhere.

    Rewards Incentive

    Earn more with our rewarding incentive program—your path to greater rewards starts here.

    Financing & Leasing

    Discover flexible financing and leasing solutions designed to align with your budget and growth goals, making your investments easier and more manageable.

    Product Information

    An Order-of-Magnitude Leap for Accelerated Computing

    The Most Powerful End-to-End AI and HPC Data Center Platform

    A100 is part of the complete NVIDIA data center solution that incorporates building blocks across hardware, networking, software, libraries, and optimized AI models and applications from NGC. Representing the most powerful end-to-end AI and HPC platform for data centers, it allows researchers to deliver real-world results and deploy solutions into production at scale.

     
     

    Deep Learning Training

     
     

    Up to 3X Higher AI Training on Largest Models

    DLRM Training

     
    Up to 3X Higher AI Training on Largest Models
     
     
     
     
     

    ​AI models are exploding in complexity as they take on next-level challenges such as conversational AI. Training them requires massive compute power and scalability.

    NVIDIA A100 Tensor Cores with Tensor Float (TF32) provide up to 20X higher performance over the NVIDIA Volta with zero code changes and an additional 2X boost with automatic mixed precision and FP16. When combined with NVIDIA® NVLink®, NVIDIA NVSwitch, PCI Gen4, NVIDIA® InfiniBand®, and the NVIDIA Magnum IO SDK, it’s possible to scale to thousands of A100 GPUs.

    A training workload like BERT can be solved at scale in under a minute by 2,048 A100 GPUs, a world record for time to solution.

    For the largest models with massive data tables like deep learning recommendation models (DLRM), A100 80GB reaches up to 1.3 TB of unified memory per node and delivers up to a 3X throughput increase over A100 40GB.

    NVIDIA’s leadership in MLPerf, setting multiple performance records in the industry-wide benchmark for AI training.

    Learn More About A100 for Training
     

    Deep Learning Inference

    A100 introduces groundbreaking features to optimize inference workloads. It accelerates a full range of precision, from FP32 to INT4. Multi-Instance GPU (MIG) technology lets multiple networks operate simultaneously on a single A100 for optimal utilization of compute resources. And structural sparsity support delivers up to 2X more performance on top of A100’s other inference performance gains.

    On state-of-the-art conversational AI models like BERT, A100 accelerates inference throughput up to 249X over CPUs.

    On the most complex models that are batch-size constrained like RNN-T for automatic speech recognition, A100 80GB’s increased memory capacity doubles the size of each MIG and delivers up to 1.25X higher throughput over A100 40GB.

    NVIDIA’s market-leading performance was demonstrated in MLPerf Inference. A100 brings 20X more performance to further extend that leadership.

    Learn More About A100 for Inference
     

    Up to 249X Higher AI Inference Performance
    Over CPUs

    BERT-LARGE Inference

     
    Up to 249X Higher AI Inference Performance  Over CPUs
     
     
     
     

    Up to 1.25X Higher AI Inference Performance
    Over A100 40GB

    RNN-T Inference: Single Stream

     
    Up to 1.25X Higher AI Inference Performance  Over A100 40GB
     
     
     
     
     

    High-Performance Computing

    To unlock next-generation discoveries, scientists look to simulations to better understand the world around us.

    NVIDIA A100 introduces double precision Tensor Cores  to deliver the biggest leap in HPC performance since the introduction of GPUs. Combined with 80GB of the fastest GPU memory, researchers can reduce a 10-hour, double-precision simulation to under four hours on A100. HPC applications can also leverage TF32 to achieve up to 11X higher throughput for single-precision, dense matrix-multiply operations.

    For the HPC applications with the largest datasets, A100 80GB’s additional memory delivers up to a 2X throughput increase with Quantum Espresso, a materials simulation. This massive memory and unprecedented memory bandwidth makes the A100 80GB the ideal platform for next-generation workloads.

    Learn More About A100 for Hpc
     

    11X More HPC Performance in Four Years

    Top HPC Apps​

     
    11X More HPC Performance  in Four Years
     
     
     
     
     

    Up to 1.8X Higher Performance for HPC Applications

    Quantum Espresso​

     
    Up to 1.8X Higher Performance  for HPC Applications
     
     
     

    High-Performance Data Analytics

     
     

    2X Faster than A100 40GB on Big Data Analytics Benchmark

     
    2X Faster than A100 40GB on Big Data Analytics Benchmark
     
     
     
     

    Data scientists need to be able to analyze, visualize, and turn massive datasets into insights. But scale-out solutions are often bogged down by datasets scattered across multiple servers.

    Accelerated servers with A100 provide the needed compute power—along with massive memory, over 2 TB/sec of memory bandwidth, and scalability with NVIDIA® NVLink® and NVSwitch, —to tackle these workloads. Combined with InfiniBand, NVIDIA Magnum IO and the RAPIDS suite of open-source libraries, including the RAPIDS Accelerator for Apache Spark for GPU-accelerated data analytics, the NVIDIA data center platform accelerates these huge workloads at unprecedented levels of performance and efficiency.

    On a big data analytics benchmark, A100 80GB delivered insights with a 2X increase over A100 40GB, making it ideally suited for emerging workloads with exploding dataset sizes.

    Learn More About Data Analytics
     
     

    Enterprise-Ready Utilization

     
     

    7X Higher Inference Throughput with Multi-Instance GPU (MIG)

    BERT Large Inference

     
    7X Higher Inference Throughput with Multi-Instance GPU (MIG)
     
     
     
     
     

    Enterprise-Ready Utilization

    A100 with MIG maximizes the utilization of GPU-accelerated infrastructure. With MIG, an A100 GPU can be partitioned into as many as seven independent instances, giving multiple users access to GPU acceleration. With A100 40GB, each MIG instance can be allocated up to 5GB, and with A100 80GB’s increased memory capacity, that size is doubled to 10GB.

    MIG works with Kubernetes, containers, and hypervisor-based server virtualization. MIG lets infrastructure managers offer a right-sized GPU with guaranteed quality of service (QoS) for every job, extending the reach of accelerated computing resources to every user.

     
     

    Subscribe and Save on Your Next Purchase

    Subscribe to Uvation’s newsletter and get the special discount code for your first purchase. Save up to 10% on your first purchase.

    uvation