MITB Banner

AMD Launches MI300X AI Accelerator, Takes On NVIDIA H100

AMD has come out of the gate swinging with their latest chips. Made for deploying AI models, this chip is AMD's foothold in the AI market.

Share

Listen to this story

AMD, going for a slice of the AI pie, has added another AI chip to its arsenal. This GPU, titled the MI300X, aims to tap a part of the AI accelerator market that NVIDIA is enjoying dominance over. The chip is purpose-built for AI tasks, and comes with up to 192GB of memory, perfect for big models like LLMs.

Due to its large amount of memory, large models can easily be deployed on the chip. The chip is also built from the ground up to enable generative AI workloads. Most state-of-the-art models today will fit into the MI300X’s 192GB of HBM3 (high-bandwidth memory), with AMD CEO Lisa Su stating

“With all of that additional memory capacity, we actually have an advantage for large language models because we can run larger models directly in memory. What that does is…it actually reduces the number of GPUs you need, significantly speeding up the performance, especially for inference.”

While this announcement comes off the back of NVIDIA’s market capitalisation reaching $1 trillion, it is an important step for AMD to get a handle on the growing AI market. AMD also showed off the capabilities of the chip by running Falcon 40B on the chip at the reveal, with Su stating that this is the “first time an LLM of this size can be run entirely in memory”.

During the keynote, AMD also released an update to RocM, a competing platform to NVIDIA’s CUDA language. This is an important step forward for AMD, as CUDA has been one of NVIDIA’s biggest moats, especially in AI-accelerated workloads. Moreover, the large amount of memory bandwidth available on the chip will allow companies to buy lesser GPUs, making AMD an interesting value proposition for smaller companies with light to medium AI workloads. 

While a price has not yet been revealed, AMD has positioned the MI300X to directly compete with NVIDIA’s H100 chips for a bigger slice of the AI compute market. However, NVIDIA has a more comprehensive lineup of products for the AI age, including its latest chips made specifically for generative AI tasks. With a lack of big customers on its platform, AMD has an uphill battle to fight against NVIDIA but has fired the first volley.

Share
Picture of Anirudh VK

Anirudh VK

I am an AI enthusiast and love keeping up with the latest events in the space. I love video games and pizza.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.