Skip to main content

Nvidia reveals first-ever CPU and Hopper GPU at GTC 2022

Nvidia CEO Jensen Huang kicked off the company’s Graphics Technology Conference (GTC) with a keynote speech full of announcements. The key reveals include Nvidia’s first-ever discrete CPU, named Grace, as well as its next-generation Hopper architecture, which will arrive later in 2022.

The Grace CPU Superchip is Nvidia’s first discrete CPU ever, but it won’t be at the heart of your next gaming PC. Nvidia announced the Grace CPU in 2021, but this Superchip, as Nvidia calls it, is something new. It puts together two Grace CPUs, similar to Apple’s M1 Ultra, connected through Nvidia’s NVLink technology.

Related Videos
A rendering of Nvidia's Grace Superchip.

Unlike the M1 Ultra, however, the Grace Superchip isn’t built for general performance. The 144-core GPU is built for A.I., data science, and applications with high memory requirements. The CPU still uses ARM cores, despite Nvidia’s abandoned $40 billion bid to purchase the company.

In addition to the Grace Superchip, Nvidia showed off its next generation Hopper architecture. This isn’t the architecture powering the RTX 4080, according to speculation. Instead, it’s built for Nvidia’s data center accelerators. Nvidia is debuting the architecture in the H100 GPU, which will replace Nvidia’s previous A100.

Nvidia calls the H100 the “world’s most advanced chip.” It’s built using chipmaker TSMC’s N4 manufacturing process, packing in a staggering 80 billion transistors. As if that wasn’t enough, it’s also the first GPU to support PCIe 5.0 and HBM3 memory. Nvidia says just 20 H100 GPUs can “sustain the equivalent of the entire world’s internet traffic,” showing the power of PCIe 5.0 and HBM3.

Customers will be able to access the GPU through Nvidia’s fourth-generation DGX servers, which combine eight H100 GPUs and 640GB of HBM3 memory. These machines, according to Nvidia, provide 32 petaFLOPs of A.I. performance, which is six times as much as last-gen’s A100.

If the DGX doesn’t offer enough power, Nvidia is also offering its DGX H100 SuperPod. This builds on Nvidia renting out its SuperPod accelerators last year, allowing those without the budget for massive data centers to harness the power of A.I. This machine combines 32 DGX H100 systems, delivering a massive 20TB of HBM3 memory and 1 exoFLOP of A.I. performance.

Nvidia Hopper GPU family.

Nvidia is debuting the new architecture with its own EOS supercomputer, which includes 18 DGX H100 SuperPods for a total of 4,608 H100 GPUs. Enabling this system is Nvidia’s fourth generation of NVLink, which provides a high bandwidth interconnect between massive clusters of GPUs.

As the number of GPUs scales up, Nvidia showed that the last-gen A100 would flatline. Hopper and fourth-gen NVLink don’t have that problem, according to the company. As the number of GPUs scales into the thousands, Nvidia says H100-based systems can provide up to nine times faster A.I. training than A100-based systems.

This next-gen architecture provides “game-changing performance benefits,” according to Nvidia. Although exciting for the world of A.I. and high-performance computing, we’re still eagerly awaiting announcements around Nvidia’s next-gen RTX 4080, which is rumored to launch later this year.

Editors' Recommendations

The popularity of ChatGPT may give Nvidia an unexpected boost
Nvidia's A100 data center GPU.

The constant buzz around OpenAI's ChatGPT refuses to wane. With Microsoft now using the same technology to power its brand-new Bing Chat, it's safe to say that ChatGPT may continue this upward trend for quite some time. That's good news for OpenAI and Microsoft, but they're not the only two companies to benefit.

According to a new report, the sales of Nvidia's data center graphics cards may be about to skyrocket. With the commercialization of ChatGPT, OpenAI might need as many as 10,000 new GPUs to support the growing model -- and Nvidia appears to be the most likely supplier.

Read more
No, ChatGPT isn’t going to cause another GPU shortage
Hopper H100 graphics card.

ChatGPT is exploding, and the backbone of its AI model relies on Nvidia graphics cards. One analyst said around 10,000 Nvidia GPUs were used to train ChatGPT, and as the service continues to expand, so does the need for GPUs. Anyone who lived through the rise of crypto in 2021 can smell a GPU shortage on the horizon.

I've seen a few reporters build that exact connection, but it's misguided. The days of crypto-driven-type GPU shortages are behind us. Although we'll likely see a surge in demand for graphics cards as AI continues to boom, that demand isn't directed toward the best graphics cards installed in gaming rigs.
Why Nvidia GPUs are built for AI

Read more
Here’s why ChatGPT might be ‘at capacity’ for you still
ChatGPT and OpenAI logos.

AI-powered ChatGPT has recently been frustrating a sizable number of potential new users due to its own popularity, resulting in a very common "at capacity" notice that many people are facing.

Unfortunately, right now, you have to wait it out or come back at a time when less people are using it. You could also try out one of the best ChatGPT alternatives, including Microsoft's recently announced ChatGPT integration in Bing.

Read more