MITB Banner

Roundup Of AI & Supercomputing Products Unveiled At NVIDIA GTC 2021

Highlights of NVIDIA CEO Jensen Huang’s keynote.

Share

“Softwares will be written by software running on AI computers” 

Jensen Huang, CEO, NVIDIA

Dubbed as the first kitchen keynote, last year NVIDIA CEO announced DLSS, which revolutionised RTX, RTX servers, NVIDIA Jarvis, A100 and DGX A100. This year too, Jensen Huang’s keynote speech acted as the symbolic ribbon cutting to one of the most anticipated events in the tech world.

The keynote speech highlighted the following developments:

NVIDIA opens gates to its omniverse

Image credits: NVIDIA

Built on top of RTX, Omniverse is a platform to connect multiple 3D worlds into a shared virtual world. It takes its inspiration from the term “metaverse”. “Pieces of the early-metaverse vision are already here in massive online social games like Fortnite or user-created virtual worlds like Minecraft,” said Huang.

“Omniverse is a platform built from the ground up to be physically based. It is fully path traced. Physics is simulated with NVIDIA PhysX, materials are simulated with NVIDIA MDL, and Omniverse is fully integrated with NVIDIA AI,” he added. 

The platform is cloud-native, multi-GPU scalable and runs on any RTX platform and can be streamed remotely on any device.   

Foray into data centre markets

The biggest takeaway from the keynote is the launch of NVIDIA’s first data centre CPU, Grace. Each CPU delivers 300 SPECint with a total of over 2,400 SPECint rate CPU performance from an eight-GPU DGX. 

The chipmaker has also announced the first-ever DPU made for AI and accelerated computing—Bluefield 3, which allows organisations to create applications with industry-leading performance and data centre security.

According to NVIDIA, Bluefield-3 can achieve 400 GB/s and has 10x processing capability of Bluefield-2. It is optimised for multi-latent, cloud-native environments software-defined, hardware-accelerated networking, storage, security and management services. It delivers equivalent data centre services of up to 300 CPU cores. 

Huang also announced DGX station, which has four 80GB A100 GPUs having more memory and bandwidth than the original DGX station. Meant for AI research, it also has refrigerated liquid cooling for its Epyc CPU and four A100 GPUs.

The station runs at a quiet 37 dB while only utilising up to 1,500 W of power. It delivers up to 2.5 petaFLOPS of floating-point performance. One of the most impressive features is that it transfers 8TB per second.

Talking about the performance, Huang said that the Megatron could linearly scale training up to 1 trillion parameters on the DGX SuperPOD with advanced optimisations and parallelisation algorithms. 

MLOps and Enterprise AI

NVIDIA EGX enterprise platform: The platform enables both existing and modern, data-intensive  applications to be accelerated and secure on the same infrastructure. It delivers the power of accelerated computing from the data centre to edge with a range of optimised hardware, an easy to deploy application and management software and a vast ecosystem of partners who offer EGX in their products.

Huang also gave a glimpse of frameworks that encapsulate the entire workflow to customise AI models. It applies transfer learning to your data to fine-tune models. “No one has all the data – sometimes it’s rare, sometimes they are trade secrets. No model can be trained for every skill. And the specialised ones are the most valuable,” said Huang as he introduced Nvidia TAO. 

TAO has a federated learning system allowing multiple users to train a shared model while having data privacy. Using TensorRT, it optimises the model for the target GPU system.

A Quantum cue

NVIDIA’s cuQuantum— an acceleration library designed for simulating quantum circuits, is meant for both tensor network solver and state vector solvers. It is also optimised to scale to large GPU memories, multiple GPUs and multiple DGX nodes. Developers can use it to speed up quantum circuit simulations based on state vector, density matrix, and tensor network methods by order of magnitude.

Stay tuned to AIM for more updates from GTC 2021.

Share
Picture of Peter Mathew

Peter Mathew

Passionate about all things media and communications. I love being a journalist, though you can see me read a book or watch a classic film in my free time.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.