MITB Banner

NVIDIA Introduces Very Big GPU, BLACKWELL

Compared to Hopper, the B200 promises 2.5x faster performance per GPU, and up to 30x faster performance for large language models.

Share

Listen to this story

In a blockbuster announcement at its annual GTC conference, NVIDIA CEO Jensen Huang unveiled the company’s next-generation Blackwell GPU architecture, promising massive performance gains to fuel the AI revolution. The GPU is named after David Harold Blackwell, an eminent American statistician and mathematician who has made significant contributions to various fields, including game theory, probability theory, information theory, and statistics.

The highlight is the flagship B200 GPU, a behemoth packing 208 billion transistors across two cutting-edge chiplets connected by a blazing 10TB/s link. Huang proclaimed it “the world’s most advanced GPU in production.”

Blackwell introduces several groundbreaking technologies. It features a second-gen Transformer Engine to double AI model sizes with new 4-bit precision. The 5th-gen NVLink interconnect enables up to 576 GPUs to work seamlessly on trillion-parameter models. An AI reliability engine maximises supercomputer uptime for weeks-long training runs.

“When we were told Blackwell’s ambitions were beyond the limits of physics, the engineers said ‘so what?'” Huang said, showcasing a liquid-cooled B200 prototype. “This is what happened.”

The Blackwell GPU represents a significant advancement in AI inference performance compared to the previous Hopper generation.

Compared to its predecessor Hopper, the B200 promises 2.5x faster FP8 AI performance per GPU, double FP16 throughput with new FP6 format support, and up to 30x faster performance for large language models.

This dramatic improvement is made possible by Blackwell’s architecture, which is specifically designed to efficiently manage the immense computational and memory demands of trillion-parameter AI models.

Major tech giants like Amazon, Google, Microsoft, and Tesla have already committed to adopting Blackwell.

Huang said training a GPT model with 1.8 trillion parameters [GPT-4] typically takes three to five months using 25,000 amperes. HOPPER architecture would require around 8,000 GPUs, consuming 15 megawatts of power and taking about 90 days (three months). In contrast, BLACKWELL would need just 2,000 GPUs and significantly less power—only four megawatts—for the same duration.

He said that NVIDIA aims to reduce computing costs and energy consumption, thereby facilitating the scaling up of computations necessary for training next-generation models.

To showcase Blackwell’s scale, NVIDIA unveiled the DGX SuperPOD, a next-gen AI supercomputer with up to 576 Blackwell GPUs and 11.5 exaflops of AI compute. Each DGX GB200 system packs 36 Blackwell GPUs coherently linked to Arm-based Grace CPUs.

While focused on AI and data centres initially, Blackwell’s innovations are expected to benefit gaming GPUs too. 

With Blackwell raising the bar, NVIDIA has clearly doubled down on its lead in the white-hot AI acceleration market. The race for the next AI breakthrough is on.

Share
Picture of Shyam Nandan Upadhyay

Shyam Nandan Upadhyay

Shyam is a tech journalist with expertise in policy and politics, and exhibits a fervent interest in scrutinising the convergence of AI and analytics in society. In his leisure time, he indulges in anime binges and mountain hikes.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.