The NVIDIA H100 is a high-end data-centre GPU accelerator designed to meet the demands of large-scale artificial intelligence (AI), high performance computing (HPC), and inference workloads. Built on NVIDIA’s Hopper™ architecture, it delivers exceptional performance, scalability and security for modern data-centre applications.
 The H100 is engineered to accelerate everything from training large language models to running high-throughput inference workloads and complex scientific simulations.
 It supports both the SXM (high-density server) and PCIe (traditional server slot) form-factors, making it flexible for various deployment scenarios.
Here are some of the key technical specifications for the NVIDIA H100 (values vary depending on variant: SXM vs PCIe):
| Parameter | Value (approximate) | 
| On-board GPU memory | 80 GB (common) | 
| Memory bandwidth | Up to ~3.35 TB/s (SXM variant) ~2 TB/s (PCIe variant) | 
| FP64 peak performance | ~26 TFLOPS (PCIe) | 
| FP64 Tensor Core performance | ~51 TFLOPS (PCIe) | 
| INT8 / FP8 (inference) | Up to ~3 000 TOPS (on PCIe with sparsity) | 
| Maximum thermal design power (TDP) | Up to ~700 W (SXM variant) ~300-350 W (PCIe variant) | 
| Form factor / Interconnect | Variants include SXM and PCIe; NVLink bandwidth up to ~900 GB/s in top configurations. | 
| Multi-GPU scaling | Supports GPU-to-GPU interconnect enabling large clusters (e.g., up to 256 GPUs with NVLink switch system) for exascale workloads. |