MITB Banner

Microsoft Joins the Chip-munks, Will Make its Own AI Chips

One of the primary reasons for Microsoft's foray into chip-making is to ​​reduce costs, which Athena could possibly slash by a third in comparison to Nvidia
Share
Listen to this story

Microsoft wants to make its own AI chips. According to a report by The Information, the company is currently engaged in the development of its AI chips, called Athena, a project the tech giant has been working on since 2019. Currently, the chips are being tested by a small group of employees from Microsoft and OpenAI. But what we don’t know yet is if and when Microsoft plans to bring the chip out for commercial use. 

So far, to run its AI models, Microsoft has been relying on chips available in the market, which is dominated by Nvidia. To help OpenAI train its GPT series large language models (LLMs), the Redmond-based company also built a supercomputer for OpenAI, powered by Nvidia A100 chips.

Given the rapid advancement in the field, AI companies such as Google, Apple and Amazon have already developed their AI chips. Google, in fact, has built a supercomputer to train its models with its TPUs (Tensor Processing Units). Amazon also has its Trainium and Inferentia processor architectures. Hence, it’s no surprise that Microsoft is walking down the same lane. 

Optimising cost 

It is estimated that just training an LLM like GPT-3 could have cost OpenAI around USD4 million. Additionally, OpenAI spends approximately USD 3 million per month to sustain ChatGPT. Besides, GPUs used to run these models are also very expensive. Nvidia, the leading producer of GPUs for the AI industry, sells its primary data centre chips for around USD10,000. Moreover, its H100 GPUs are being sold at USD40,000 on eBay. Reportedly, Microsoft’s supercomputer utilises tens of thousands of Nvidia’s H100 and A100 data centre GPUs.

One of the primary reasons for Microsoft to build its own chips is to ​​reduce its costs. As per the report, Athena could possibly reduce the cost per chip by a third in comparison to Nvidia.

Besides, Microsoft also wants to reduce its reliance on Nvidia. Hence, building chips in house could mean Microsoft can design the chips, its architecture, and compatibility according to their own needs. As per the report, Microsoft has designed Athena for both training as well as running its AI models.

Given that Microsoft is aiming to introduce AI-powered features in Bing, Office 365, and GitHub, the migration could prove advantageous for the company in terms of cost. “As needs expand and diversity of use expands as well, it’s important for Microsoft and the other hyperscalers to pursue their own optimised versions of AI chips for their own architectures and optimised algorithms (not CUDA-specific),” Jack Gold, of J Gold Associates, told VentureBeat

On device AI

Typically, AI models are run on the cloud. However, recent advancements have shown that it’s possible to run AI models entirely on devices. Very recently, a group of Qualcomm engineers managed to run text-to-image AI model Stable Diffusion on an Android device. 

Currently, there are laptops in the market that have chips designed to help train AI models. Microsoft too, makes premium laptops with hi-end hardware under the Surface brand. Microsoft is also reportedly designing its own ARM-based processors for its Surface laptops; however, these chips have not yet been released. 

Hence, we could possibly see a future generation of Surface laptops that run a ChatGPT like model completely on the device. 

Competition with Nvidia?

Currently, Nvidia is the leading supplier of AI chips holding more than 90% of the enterprise GPU market. While Nvidia is focussed on building GPUs, Microsoft’s focus lies elsewhere. It wants to bring its AI products to the enterprise. 

Currently, Microsoft is building Athena for its own internal use. Hence, it’s unlikely that Microsoft will be in competition with Nvidia in the chips market. For now, Microsoft’s aim is to lower its cloud operating costs. However, down the line, Microsoft could use Athena to enhance its cloud services and devices, offering better performance and lower costs than its competitors. 

A challenge for developers?

At present, it is uncertain if Microsoft will offer these chips to its customers using the Azure cloud service. If Microsoft decides to run Azure on Athena, the newer chips might pose a challenge for developers because the chips may differ from existing ones in terms of their design, performance, power usage, and compatibility with existing software and frameworks. 

Initially, developers may face challenges in migrating to the new AI chips due to these differences and may need to adapt their workflows and code accordingly. Transitioning their existing code to the new platform could prove to be challenging.  Further, developers might require detailed documentation and resources on the new AI chips, including tutorials, sample code, and compatibility guidelines.

While the report suggests that Microsoft’s Athena chips may be available within a year, it may take some time for them to become widely available.

PS: The story was written using a keyboard.
Picture of Pritam Bordoloi

Pritam Bordoloi

I have a keen interest in creative writing and artificial intelligence. As a journalist, I deep dive into the world of technology and analyse how it’s restructuring business models and reshaping society.
Related Posts

Download our Mobile App

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox
Recent Stories

Featured

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

AIM Conference Calendar

Immerse yourself in AI and business conferences tailored to your role, designed to elevate your performance and empower you to accomplish your organization’s vital objectives. Revel in intimate events that encapsulate the heart and soul of the AI Industry.

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed