The NVIDIA H100 80GB Tensor Core GPU is a flagship data-center accelerator purpose-built for Generative AI, LLM training/inference, HPC, and large-scale data analytics. Built on the NVIDIA Hopper™ architecture, H100 introduces the Transformer Engine with FP8 precision to dramatically increase throughput for transformer models, while its high-bandwidth memory and advanced interconnect options help scale performance from a single server to large GPU clusters.
Key Highlights
- 80GB High-Bandwidth Memory (HBM): Designed to keep large models and datasets resident on-GPU for faster training/inference.
- Transformer Engine + FP8: Dedicated acceleration for transformer workloads; NVIDIA highlights major speedups vs prior generation for large language models.
- 4th-Gen Tensor Cores (Hopper): High performance across AI precisions (FP8/FP16/BF16/INT8) for training and inference.
- MIG Support (Multi-Instance GPU): Partition a single GPU into up to 7 isolated instances (platform dependent) to improve utilization in shared environments.
- Enterprise & Exascale Scaling: Supports large multi-GPU scaling (e.g., NVLink/NVLink Switch System depending on form factor) for massive AI/HPC deployments.







