MITB Banner

Stanford professor wants to train robots to understand humans better

To create a model built on the idea of prospect theory, Sadigh and her team paired humans and robots dyadically to build a tower with each other.

Share

Stanford professor wants to train robots to understand humans better

Dorsa Sadigh, a professor in the computer science department at Stanford, believes developing a better understanding of human behaviour will be critical in human-robot interaction. She explained her line of research at Stanford Engineering’s The Future of Everything podcast.

Human behaviour is often unpredictable. For example, we often have priorities that aren’t easily evident to those around us. Humans are also prone to changing their minds and behaving differently in different situations. So it’s a significant engineering challenge to get robots to understand the fundamentals of human behaviour and motivation.

Sadigh believes one of the core challenges in robotics today is dealing with human behaviour. While exploring ways to make autonomous driving work, she asked cognitive scientists and psychologists about modelling human behaviour. Unfortunately, none had feasible computational models fit for roboticists and engineers, which sent Sadigh down a human-robot interaction rabbit hole. 

Multidisciplinary approach to building computational models of humans

Building computational models of humans call for a multidisciplinary approach as models provided by cognitive scientists and psychologists can’t be implemented in the planning or perception loops of computer scientists. Similarly, simply having computer scientists build data-driven models that use human data also falls short because that doesn’t account for the unpredictability of human behaviour. 

Roboticists leverage ideas such as bounded rationality and prospect theory to build such models. According to bounded rationality, humans aren’t entirely irrational–instead, their rationality is “bounded.” Meaning, humans have bounded resources and tend to make mostly rational decisions within those limitations. For example, if a person is under severe time constraints, they may not make the most optimal decision—but they’ll make the best decision they can within that limited amount of time. On the other hand, prospect theory assumes that, if an investor has to choose between two things of equal value, they will choose the one presented in terms of potential gains.

To create a model built on the idea of prospect theory, Sadigh and her team paired humans and robots dyadically to build a tower with each other. The robot isn’t perfect and is susceptible to failure, leaving their human partner open to either underestimating or overestimating the robot’s capabilities. The work demonstrates how humans make suboptimal decisions, which they were trying to model. Once the model was captured, the robot was able to recognise human suboptimality and take decisions to complement it. 

Roboticists have developed ways of modelling these bounded rational behaviours. However, this model doesn’t account for out-of-distribution data. The question of how to detect and what to do in these scenarios is an active area of research. 

What comes first: Training humans or robots? 

When you pair a human and a robot together, you’re essentially pairing up two learning agents. Since both have different learning rates, it makes sense for the robot to give humans additional time to adapt—and the robot adjusts around the human’s changing behaviour.

Sadigh and her team also have a project around active teaching, focused on educating humans on operating a robot. To that end, the team looked at what makes humans bad at playing certain games—such as “lunar lander” or chess. By determining what makes these games challenging for humans—such as bounded memory or bounded time—it is possible to teach humans particular skills that make them better at the game.

Today, many human-robot collaborations are computationally intractable and expensive because they require the robot to build a belief over the state and the strategy of the other agent (the human). However, this is computationally expensive to keep track of. However, humans tend to be pretty good at keeping track of low dimensional representations suffice for coordinating with a collaborator. That’s the idea that Sadigh and her team have been exploring: keeping low dimensional representations as a way of partner modelling. 

Share
Picture of Srishti Mukherjee

Srishti Mukherjee

Drowned in reading sci-fi, fantasy, and classics in equal measure; Srishti carries her bond with literature head-on into the world of science and tech, learning and writing about the fascinating possibilities in the fields of artificial intelligence and machine learning. Making hyperrealistic paintings of her dog Pickle and going through succession memes are her ideas of fun.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.