NVIDIA next-gen H100 Hopper GPU supports 6 stacks of high-bandwidth-memory

At the upcoming GTC 2022, NVIDIA CEO Mr. Jensen Huang will have a keynote speech at 8:00 am PST on March 22, which is likely to bring a new generation of Hopper architecture GPUs. In the past two days, a lot of details about the Hopper architecture GPU have been leaked.

According to VideoCardz, the leaked pictures show that the H100 Hopper GPU supports up to 6 high-bandwidth memory stacks and is still a monolithic structure, but its specific specifications are still uncertain. Perhaps as rumored, the GH202 will use a multi-chip module design (MCM) and advanced CoWoS packaging. Nvidia will launch a number of GH100-based products, including an SXM based H100 card for DGX mainboard, a DGX H100 station, and even a DGX H100 SuperPod.


Some netizens have sorted out the information stolen by hackers and the information that has been circulated before and summarized the situation of the H100 Hopper GPU. It is said that the GH100 is manufactured using TSMC’s 5nm process and will have 48MB of L2 cache, which is an improvement over the 40MB of the Ampere architecture GA100 and three times the AMD Instinct MI250 (16MB). However, compared with the 96MB of the flagship chip AD102 of the Ada (Lovelace) architecture, the L2 cache of the GH100 is half less.

It is rumored that the GH100 is equipped with 8 sets of GPCs, each GPC is equipped with 9 sets of TPCs, and each TPC has two sets of SMs. If the CUDA core ratio of each set of SMs does not change, it means a total of 144 sets of SMs and 18432 CUDAs core. And only 1 of the 8 groups of GPCs has a 3D engine, and the other 7 groups will not be equipped. Although the GH100 is equipped with 144 sets of SMs, it should not be all enabled, and Nvidia is expected to disable 15% to 20% of the SMs.