Modular’s Attempt to Steal NVIDIA’s Mojo

Modular software enables developers to run models on non-NVIDIA servers like AMD, Intel, and Google, potentially addressing the GPU crunch
Listen to this story

AI startup Modular recently raised $100 million in a recent funding round, led by General Catalyst, Google Ventures, SV Angel, Greylock, and Factory. 

The 20-month-old startup by former Google and Apple alumnus, Tim Davis and Chris Lattner is making waves in the AI industry, offering software that aims to challenge the dominance of NVIDIA’s CUDA and similar alternatives, providing developers with a more accessible and efficient solution for building and combining AI applications.

“Modular’s technology scales across a wide range of hardware and devices, which gives users more flexibility as their needs evolve,” said Modular CEO and co-founder, Chris Lattner, in an exclusive interaction with AIM. 

NVIDIA’s CUDA software, used for machine learning, only supports its own GPUs, whereas Modular’s software aims to enable developers to run models on non-NVIDIA servers like those from AMD, Intel, and Google—potentially providing a solution to the GPU crunch as the demand by companies like Microsoft, Google, OpenAI, xAI and Meta strain supply. 

Lattner explained saying that their Modular’s AI engine and Mojo have been designed with ease of use in mind, catering to machine learning engineers by employing a Python-based approach rather than a more intricate C++ foundation, as seen in CUDA. “When compared to CUDA for GPU programming, Modular’s engine and Mojo are easier to use and more familiar to ML engineers, notably being Python-based instead of C++ based,” he added.

He said that Modular exhibits remarkable scalability across an expansive spectrum of hardware and devices giving users a higher degree of flexibility, ensuring that their AI solutions can evolve with their requirements. The AI engine builds on these strengths to provide higher performance and productivity by combining individual operations into an efficient optimised execution environment,” said Lattner. 

Modular’s tools seamlessly integrate into existing workflows, negating the need for wholesale rearchitecting or code rewriting in C++ or CUDA. This affords developers a frictionless transition and empowers them to unlock heightened productivity and performance without incurring exorbitant costs.

A cornerstone of Modular’s arsenal is the Mojo toolkit, which represents a concerted effort to simplify AI development across diverse hardware platforms. The Mojo programming language, blends the ease of use associated with Python with features like caching and adaptive compilation techniques, targeting improved performance and scalability in AI development.

Towards an AI Future 

In an era where tech alumni-founded startups command high valuations, Modular’s approach to validating its commercial momentum and proving its value proposition to investors remains crucial. 

Modular’s journey is not without challenges. The adoption of a new programming language like Mojo can be hindered by Python’s established status in the machine-learning landscape. However, Lattner’s conviction in Mojo’s distinct advantages and its potential to revolutionize AI development processes remains unshaken.

Given the duo’s experience, the venture exhibits potential to make it big. For instance, Lattner has led the creation of Swift, a programming language by Apple, while Davis has led Google’s machine-learning product efforts, focusing on getting the models working directly on devices.

With a rapidly growing community of over 120,000 developers, Modular claims that it has gauged demand from thousands of prominent enterprises that are excited to deploy its infrastructure.

“We have been able to achieve tremendous momentum in only 20 months,” said Davis, Modular co-founder and President. “The financing will allow us to accelerate our momentum even more, scaling to meet the incredible demand we have seen since our launch,” he added, talking about the recent funding that it raised. 

Competition Galore

Lattner acknowledged that while Modular faces competition, a lot of companies are offering point solutions that fail to resolve the challenges across the AI infrastructure stack for developers and enterprises. Besides NVIDIA, some of its competitors include Cerebras and Rain among others. 

“There is no solution in the market that unifies the frameworks and compute architectures, and supports all their weird and wonderful features with really minimal migration pain,” said Lattner, stating the USP of the company. 

Further, he said that while others allege that they are fast, deeper dive forces one to change application or model-specific codes, and they don’t scale across different hardware types either.

Lattner also said that Modular’s technologies are designed to complement NVIDIA’s existing AI infrastructures and that the chip giant is an important partner in this endeavour. The overarching mission is to facilitate broader adoption of hardware among AI developers by unifying technology stacks, simplifying complexities, and making the process more accessible, he said.

Simply put, Modular’s strategy hinges on its holistic approach, which seeks to unify frameworks and compute architectures with minimal migration challenges. Unlike some competitors, Modular’s solutions aim to address end-to-end challenges, fostering accessibility, innovation, and ethical considerations in AI technology.

NVIDIA vs the World 

Modular is not alone, there are several other startups challenging Nvidia’s dominance in GPU manufacturing and the associated software that binds users to its chips. Notable companies in this competition include d-Matrix, Rain Neuromorphics, and Tiny Corp. The collective aim is to transform the AI chip landscape by providing alternatives to NVIDIA’s products, which can be expensive for training and running machine-learning models. These startups are focusing on designing chips and software that they claim offer improved efficiency compared to Nvidia’s GPUs.

Rain Neuromorphics, now known as Rain AI, is addressing the high costs of training and running machine-learning models on conventional GPUs. Their approach combines memory and processing, similar to human synapses, resulting in cooler and energy-efficient operation compared to Nvidia’s GPUs, which require continuous cooling and drive up electricity costs.

Tiny Corp, founded by George Hotz, the former CEO of Comma AI, focuses on open-source deep-learning tools named tinygrad. These tools aim to accelerate the training and running of machine-learning models.

However, NVIDIA stands apart and according to Databricks CEO Naveen Rao, has separated itself from competitors. Despite the challenges and past bankruptcies of startups attempting to compete with Nvidia, these companies are betting on the transformative potential of AI to gain traction in the competitive AI chip sector.

Download our Mobile App

Shyam Nandan Upadhyay
Shyam is a tech journalist with expertise in policy and politics, and exhibits a fervent interest in scrutinising the convergence of AI and analytics in society. In his leisure time, he indulges in anime binges and mountain hikes.

Subscribe to our newsletter

Join our editors every weekday evening as they steer you through the most significant news of the day.
Your newsletter subscriptions are subject to AIM Privacy Policy and Terms and Conditions.

Our Recent Stories

Our Upcoming Events

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox