A Comparative Analysis of NVIDIA A100 Vs. H100 Vs. L40S Vs. H200

A Comparative Analysis of NVIDIA A100 Vs. H100 Vs. L40S Vs. H200

NVIDIA recently announced the 2024 release of the NVIDIA HGX™ H200 GPU—a new, supercharged addition to its leading AI computing platform. Gcore is excited about the announcement of the H200 GPU because we use the A100 and H100 GPUs to power up our AI GPU cloud infrastructure and look forward to adding the L40S GPUs to our AI GPU configurations in Q1-2024. So we consider this the right time to share a comparative analysis of the NVIDIA GPUs: the current generation A100 and H100, the new-generation L40S, and the forthcoming H200.

Comparison of A100 Vs. H100 Vs. L40S Vs. H200

The NVIDIA A100, H100, L40S, and H200 represent some of the most advanced and powerful GPUs in the company’s lineup. They’re designed specifically for professional, enterprise, and data center applications, and they feature architectures and technologies optimized for computational tasks, AI, and data processing. Let’s see how they stack up against each other on key technical specifications.

ArchitectureAmpereHopperAda LovelaceHopper
Release Year2020202220232024
FP649.7 TFLOPS34 TFLOPSData not available34 TFLOPS
FP64 Tensor Core19.5 TFLOPS67 TFLOPSData not available67 TFLOPS
TF32 Tensor Core312 TFLOPS989 TFLOPS183 | 366* TFLOPS989 TFLOPS*
BFLOAT16 Tensor Core624 TFLOPS1,979 TFLOPS362.05 | 733* TFLOPS1,979 TFLOPS*
FP16 Tensor Core624 TFLOPS1,979 TFLOPS362.05 | 733* TFLOPS1,979 TFLOPS*
FP8 Tensor CoreNot applicable3,958 TFLOPS733 | 1,466* TFLOPS3,958 TFLOPS*
INT8 Tensor Core1248 TOPS3,958 TOPS733 | 1,466* TFLOPS3,958 TFLOPS*
INT4 Tensor CoreData not availableData not available733 | 1,466* TFLOPSData not available
GPU Memory80 GB HBM2e80 GB48GB GDDR6 with ECC141GB HBM3e
GPU Memory Bandwidth2,039 Gbps3.35 Tbps864 Gbps4.8 Tbps
DecodersNot applicable7 NVDEC 7 JPEGNot applicable7 NVDEC 7 JPEG
Max Thermal Design Power (TDP)400WUp to 700W (configurable)350WUp to 700W (configurable)
Multi-Instance GPUsUp to 7 MIGs @ 10 GBUp to 7 MIGs @ 10 GB eachNoUp to 7 MIGs @16.5 GB each
Form FactorSXMSXM4.4“ (H) x 10.5” (L), dual slotSXM**
InterconnectNVLink: 600 GB/s PCIe Gen4: 64 GB/sNVLink: 900GB/s PCIe Gen5: 128GB/sPCIe Gen4 x16: 64GB/s bidirectionalNVIDIA NVLink®: 900GB/s PCIe Gen5: 128GB/s
Server OptionsNVIDIA HGX™ A100-Partner and NVIDIA-Certified Systems with 4,8, or 16 GPUs NVIDIA DGX™ A100 with 8 GPUsNVIDIA HGX H100 Partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs NVIDIA DGX H100 with 8 GPUsData not availableNVIDIA HGX™ H200 partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs
NVIDIA AI EnterpriseIncludedAdd-onData not availableAdd-on
CUDA® Cores6,91216,89618,176Data not available

* With sparsity.
** Preliminary specification. May be subject to change.
Source: https://resources.nvidia.com/l/en-us-gpu

Based on the above comparison, we’re expecting the H200 to outperform the previous and current generation of NVIDIA data center GPUs across use cases. The current generation—the H100—is a close match to the H200, with near identical multi-precision computing performance. So, while H200s will offer improvements, H100s will remain a top option. As for the A100, it’s the least-performant GPU when compared to its successors, while still offering solid performance for certain tasks.

The L40S differs from the A100 and H100 because it includes Third-Gen RT cores (142) with 212 TFLOPS of RT core performance and 568 Fourth-Gen Tensor Cores. However, we don’t have sufficient information about these parameters for H200 yet, so it remains to be seen exactly how the L40S and H200 will stack up.


Let’s check out each GPU in turn to discover more about its features, performance, and the use cases where it shines.


The NVIDIA A100 GPU was the first GPU to feature the Ampere architecture back in 2020. Prior to the release of H100 in 2022, the A100 was a leading GPU platform. It offered a substantial leap in performance compared to its predecessors thanks to improved Tensor cores for AI, increased CUDA core count for parallel processing, enhanced memory, and the fastest-ever memory bandwidth at 2 Tbps. It supports multi-instance GPU (MIG) that allows a single A100 GPU to be partitioned into smaller, independent GPUs to maximize resource allocation and efficiency in cloud and data center environments.

Despite being surpassed in performance by newer models, the A100 GPU remains a great choice for training complex neural networks as part of deep learning and AI learning tasks because of its powerful Tensor Cores and high computational throughput. It also shines at AI inference tasks such as speech recognition, image classification, recommendation systems, data analytics and big data processing, scientific computing and simulations, and high-performance computing (HPC) tasks including genome sequencing and drug discovery.


The NVIDIA H100 GPU can handle the most demanding AI workloads and large-scale data processing tasks. H100 includes next-generation Tensor Cores, which dramatically enhance AI training and inference speeds. It also supports double precision (FP64), single precision (FP32), half precision (FP16), and integer (INT8) compute tasks.

The H100 offers a substantial performance boost over the A100, including the following benefits:

  • Six times faster—capable of four petaflops for FP8
  • 50% memory increase—uses HBM3 high-bandwidth memory up to 3 Tbps with external connectivity nearly reaching 5 Tbps
  • Up to six times faster model transformer training thanks to its new Transformer Engine

While H100 covers similar use cases and performance features as the A100, the H100 GPU can handle massive AI models, including those using transformer architectures and more complex scientific simulations. The H100 GPU is also a superior choice for real-time and responsive AI applications, like advanced conversational AI and real-time translations.


The L40S is one of NVIDIA’s most powerful GPUs, released in Q4 of 2023 (and joining Gcore’s infrastructure right away.) It’s designed to handle the next generation of data center workloads: generative AI, large language model (LLM) inference and training, 3D graphics, rendering, video, and scientific simulations.

NVIDIA L40S delivers up to 5x higher inference performance and up to 2x real-time ray-tracing (RT) performance compared to previous-generation GPUs, like the A100 and H100. The 48GB of GDDR6 memory with ECC (Error Correcting Code) plays a crucial role in maintaining data integrity in high-performance computing environments. It also comes equipped with over 18,000 CUDA cores—the parallel processors that are key to handling complex computational tasks.


The NVIDIA H200 is the latest in NVIDIA’s lineup of GPUs, scheduled to be shipped during Q2-2024. It’s the first GPU to offer 141 GB of HBM3e memory at 4.8 Tbps—nearly double the capacity of the NVIDIA H100 Tensor Core GPU with 1.4x more memory bandwidth. The latter is relevant in high-performance computing and results in up to 110x faster time-to-results compared to CPUs. Inference speed is double that of H100 GPUs when handling Llama2 70B inference.

The H200 is set to play a critical role in Artificial Intelligence of Things (AIoT) for edge computing and IoT applications. You can also expect the highest available GPU performance from H200s across application workloads, including LLP training and inference for the largest models beyond 175 billion parameters, and in generative AI and HPC applications.


Based on the initial specifications and preliminary performance benchmarks, the NVIDIA HGX™ H200 seems a significant step forward from A100 and H100 GPUs in terms of overall performance, energy savings, and TCO (total cost of ownership). We hope this comparative guide will help you to choose the right NVIDIA data center GPU as the ideal solution for solving your business problems in deep learning and AI, HPC, graphics, or virtualization in the data center or at the edge.

Gcore offers various AI GPU configurations for bare metal servers and virtual machines based on A100 and H100 GPUs. In addition, our Managed Kubernetes platform allows you to use bare metal servers and virtual machines with A100 and H100 GPUs as worker nodes. We’ll soon add more AI GPU configurations based on the latest L40S GPUs—stay tuned for updates!


Subscribe to our newsletter

Stay informed about the latest updates, news, and insights.