Listen to this story
|
MetaAI researchers recently released MultiRay, a new platform for running large-scale AI systems.
This new platform solves the issue with current AI models which compute large amounts of data to produce results. In its current state, each AI model is trained to perform particular tasks, which while generating high-quality output, requires huge operational costs.
However, Meta’s new platform majorly cuts down on processing costs by allowing multiple models to run on its platform. The universal models are trained to perform functions across varying tasks and domains, producing results with better quality than one’s performed by specialised models.
AIM Daily XO
Join our editors every weekday evening as they steer you through the most significant news of the day, introduce you to fresh perspectives, and provide unexpected moments of joy
Your newsletter subscriptions are subject to AIM Privacy Policy and Terms and Conditions.
Read: Meta Unveils CICERO to Achieve Human-level Performance at Diplomacy
Teams across Meta are using MultiRay to improve and iterate on ML models for wide-ranging applications. TextRay, the first model among many, has been in production since 2020, and will be used for several text understanding applications that include detecting inauthentic content, identifying hate speech, and enhancing users’ search experience.
Download our Mobile App
Similarly, PostRay, their second model, combines elements of text and image understanding, and is used for applications like topic classification, which is used in Reels.
However, instead of combining multiple models—for text, image, and videos—into one large model, MultiRay utilises large foundational models to represent an input that can be applied to multiple task-specific models. The single input that will run many of these models will be quite large, so as to convey more information.
Read: Why Meta Took Down its ‘Hallucinating’ AI Model Galactica?
The company-wide computation run by the large foundational model is therefore centralised and executed on accelerations like GPUs, with cache used to save spending on recomputation as much as possible. At the moment, the platform is powering over 125 use cases across Meta, and supports up to 20 million queries per second, while being able to serve 800 billion queries per day.