MITB Banner

Google AI Introduces Unified Language Learner to Assist Language Model Efficiency

According to the company’s blog, the UL2 frames objective functions for training language models as denoising tasks, where the model has to recover missing sub-sequences of a given input

Share

Listen to this story

One of the crucial goals of machine learning (ML) research is to build models that understand and generate natural language efficiently. This has a direct impact on building smart systems for everyday applications, where researchers keep a target of improving the quality of language models. 

Researchers at Google AI in ‘Unifying Language Learning Paradigms’, have presented a language pre-training paradigm called Unified Language Learner (UL2) that focuses on improving the performance of language models across datasets and setups around the world. 

Some of the most common paradigms to build and train language models either use autoregressive decoder-only architectures such as PaLM or GPT-3, where the model is trained to predict the next word for a given phrase. Whereas, other models such as T5, ST-MoE span corruption-based encoder-decoder architectures. However, there remains an opportunity to create an effective unified framework for pre-training models.

According to the company’s blog, the UL2 forms different objective functions for training language models as denoising tasks, where the model has to recover missing sub-sequences of a given input. 

Furthermore, a novel mixture-of-denoisers is used during pre-training – which samples from a varied set of objectives – each with different configurations. The team then demonstrates the models trained using the framework in a variety of language domains that includes models fine-tuned for down-stream tasks and prompt-based few-shot learning. 
Google AI says, “UL2 demonstrates superior performance on a plethora of fine-tuning and few-shot tasks. Additionally, we show that UL2 excels in generation, language understanding, retrieval, long-text understanding and question answering tasks. We publicly release checkpoints of our best performing UL2 model with 20 billion parameters, which we hope will inspire faster progress in developing better language models in the machine learning community as a whole.”

Share
Picture of Bhuvana Kamath

Bhuvana Kamath

I am fascinated by technology and AI’s implementation in today’s dynamic world. Being a technophile, I am keen on exploring the ever-evolving trends around applied science and innovation.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.