MITB Banner

DeepMind Researchers Develop Tools To Visualise Unfairness Using Causal Bayesian Networks

Share

Machine Learning engineers work around bias or the offsets in a model by drawing insights from the output, gauging the losses, going through tonnes of data and repeating till agreeable results have been obtained.

This is a traditional process which takes time but works decently. An alternative to this approach is the Lagrangian approach, a mathematical method to find the local maxima and local minima of a function when provided with equality constraints. This too, comes with its own set of complexities.

The unfairness of machine learning algorithms was exposed when they were deployed for manual tasks like hiring, surveillance and other such critical tasks, where the damages can be irreversible. The algorithms treated few groups more favourably than others. This discrimination was due to a number of factors like race, gender and other such sensitive attributes. Usually, a model suffers from bias when the training data has over-represented groups or when the data is scarce.

The training data can display different patterns of unfairness depending on how and why the sensitive attribute influences other variables. 

Categorising the variables based on their sensitivity is a daunting task for any data scientist. This requires a deep understanding of both technical and sociological influences. 

To identify and address the emergence of unfairness in the algorithms, the researchers at DeepMind release two papers that demonstrate the use of Causal Bayesian networks(CBNs) to visualise and quantify the degree of unfairness. 

Visualising Unfairness With CBNs

A Bayesian network is a directed acyclic graph where nodes and edges represent random variables and statistical dependencies.

When equipped with causal semantic, namely when representing the data- generation mechanism, Bayesian networks can be used to visually express causal relationships. More specifically, CBNs enable us to give a graphical definition of causes and causal effect.

Causal Bayesian Networks offer a powerful visual and quantitative tool for expressing the relationships among random variables in a dataset. Here a hypothetical admission scenario is taken as demonstrated by the researchers at DeepMind.

They offer a graphical interpretation of unfairness in a dataset as the presence of an unfair causal path in the causal Bayesian network representing the data-generation mechanism

In this hypothetical admission scenario, applicants are admitted based on qualifications Q, choice of department D, and gender G; and in which female applicants apply more often to certain departments (for simplicity’s sake, we consider gender as binary, but this is not a necessary restriction imposed by the framework.

Source: DeepMind blog

According to the definition, the path G→D→A is causal, whilst the path G→D→A←Q is non-causal.

To visualise the way unfairness emerges in the network, three scenarios were considered:

  1. Firstly, female applicants voluntarily apply to departments with low acceptance rates, and therefore the path G→D is considered fair.
  2. Now female applicants apply to departments with low acceptance rates due to systemic historical or cultural pressures, and therefore the path G→D is considered unfair (as a consequence, the path D→A becomes partially unfair).
  3. Finally, the college lowers the admission rates for departments voluntarily chosen more often by women. The path G→D is considered fair, but path D→A is partially unfair.

Quantifying Unfairness

Causal Bayesian Networks, as described above, give good insights by enabling the visual framework to represent fairness or unfairness in a model. 

Not only to visualise but these networks can also be used to measure the degree of unfairness in a model. 

Path-specific techniques enable us to estimate the influence that a sensitive attribute has on other variables along specific sets of causal paths. This can be used to measure the degree of unfairness on a given dataset in complex scenarios in which some causal paths are considered unfair whilst other causal paths are considered fair.

Most fairness definitions express properties of the model output with respect to sensitive information, without considering the relations among the relevant variables underlying the data-generation mechanism.

In spite of showing promising results, there are challenges that will surface when the above approach is deployed. One such challenge being the identification of a CBN that describes a certain dataset generation with great accuracy.

As machine learning continues to be embedded in more systems which have a significant impact on people’s lives and safety, researchers insist that it is of great responsibility for machine learning practitioners to tackle potential biases embedded in how training data sets are generated.

Share
Picture of Ram Sagar

Ram Sagar

I have a master's degree in Robotics and I write about machine learning advancements.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.