MITB Banner

Jensen Huang Personally Delivers First NVIDIA DGX H200 to OpenAI 

H200 succeeding the highly sought-after H100, boasts 1.4 times more memory bandwidth and 1.8 times more memory capacity.

Share

Listen to this story

NVIDIA CEO Jensen Huang personally hand-delivered the first NVIDIA DGX H200 to OpenAI. In a post shared by Greg Brockman, president of OpenAI, Huang is seen posing with Brockman,and chief Sam Altman, with the DGX H200 in the middle.

“First @NVIDIA DGX H200 in the world, hand-delivered to OpenAI and dedicated by Jensen “to advance AI, computing, and humanity”, wrote Brockman on X. 

Huang delivering the first GPU to OpenAI appears to have become a new tradition. Back in 2016, Huang donated the first DGX-1 AI supercomputer to OpenAI, in support of democratizing AI technology. At that time, Elon Musk was the one who received it.

The new GPU could be a much-needed addition to OpenAI’s arsenal as the organization is currently working on GPT-5 and plans to make Sora publicly available this year.

NVIDIA introduced DGX H200 last year. The upgraded GPU, succeeding the highly sought-after H100, boasts 1.4 times more memory bandwidth and 1.8 times more memory capacity. These enhancements significantly enhance its capability to manage demanding generative AI tasks.

Moreover, the H200 has a faster memory specification known as HBM3e, elevating its memory bandwidth to 4.8 terabytes per second from the H100’s 3.35 terabytes per second. Its total memory capacity also rises to 141GB, up from the 80GB of its predecessor.

“To create intelligence with generative AI and HPC applications, vast amounts of data must be efficiently processed at high speed using large, fast GPU memory,” said Ian Buck, vice president of hyperscale and HPC at NVIDIA. “With NVIDIA H200, the industry’s leading end-to-end AI supercomputing platform just got faster to solve some of the world’s most important challenges.”

NVIDIA also launched a new AI supercomputer with H200. The NVIDIA DGX H200 utilises NVLink interconnect technology alongside the NVLink Switch System, merging 256 H200 superchips into a single GPU unit. 

The setup achieves an impressive 1 exaflop of performance and offers 144 terabytes of shared memory, marking a significant leap from the previous generation NVIDIA DGX A100 introduced in 2020, which had considerably less memory.

Share
Picture of Siddharth Jindal

Siddharth Jindal

Siddharth is a media graduate who loves to explore tech through journalism and putting forward ideas worth pondering about in the era of artificial intelligence.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.