The H200 is the first GPU to offer HBM3e memory, providing 141GB of memory and 4.8 terabytes per second bandwidth. It will be available in several form factors, including Nvidia HGX H200 server boards and the Nvidia GH200 Grace Hopper Superchip. Amazon Web Services, Google Cloud, Microsoft Azure, and Oracle Cloud Infrastructure will be the first to deploy H200-based instances starting next year. However, Nvidia continues to face challenges with US government export restrictions limiting sales to China.
Key takeaways:
- Nvidia has announced the HGX H200 Tensor Core GPU, a powerful AI chip that could potentially accelerate AI applications and improve response times for existing AI models.
- The H200 is the first GPU to offer HBM3e memory, providing 141GB of memory and 4.8 terabytes per second bandwidth, which is 2.4 times the memory bandwidth of the Nvidia A100 released in 2020.
- Amazon Web Services, Google Cloud, Microsoft Azure, and Oracle Cloud Infrastructure will be the first to deploy H200-based instances starting next year, with availability from global system manufacturers and cloud service providers starting in Q2 2024.
- Nvidia is facing export restrictions from the US government, limiting sales to China. In response, Nvidia has introduced three new scaled-back AI chips for the Chinese market, which accounts for a quarter of Nvidia's data center chip revenue.