MITB Banner

DeepMind’s New AI Framework Helps Machines Understand Humans Better

The new framework uses reinforcement learning to build AI agents that can follow instructions, and safely perform actions in open-ended conditions

Share

Listen to this story

DeepMind recently released a framework that will enable the creation of AI agents that can understand human instructions and perform actions. 

Existing AI frameworks have been at the receiving end of criticism for ignoring the situational understanding inherent to how humans use language. For example, DALL.E 2, the text-to-image generator, received a lot of flak for failing to understand the syntax of text prompts. So, for a simple input text like ‘a spoon on a cup’, the responses would include all images consisting of a spoon and a cup in the dataset, without really knowing the relationality between the spoon and cup in the text. 

To overcome this problem and build agents that can follow instructions and safely perform actions in open-ended conditions, the researchers at DeepMind created a new model within a video game environment.

The new framework will move away from training the AI agents, keeping a score based on wins/losses calculated by computer code used in applications like StarCraft and Dota. Instead, people will create tasks and score AI agents based on their behaviour. 

Although in its infancy, the new research paradigm develops real-time agents that can navigate, talk, interact with people, search for information, ask questions, control items, and do various other tasks.

The game is conceptualised based on a child’s “playhouse”, where both humans and agents have an avatar that can interact with each other and manipulate objects around. The framework is four-stepped. First, human-human interaction facilitates the ground for training initial agents by imitation learning. Then comes the cycle of human-agent interaction with performance judgement and optimisation of these judgements, which will improve agents based on reinforcement learning (RL). 

The new research is built on DeepMind’s previously published work demonstrating the role of imitation learning in creating AI agents that can capture the diversity of human behaviour well. In the new work, a reinforcement learning model is used to improve AI systems based on scores obtained via human evaluation.   

The framework has also been described as offering utility in building digital and robotic assistants—to create safe AI.

Share
Picture of Ayush Jain

Ayush Jain

Ayush is interested in knowing how technology shapes and defines our culture, and our understanding of the world. He believes in exploring reality at the intersections of technology and art, science, and politics.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.