Can The ML Models Ever Be Devoid Of Labelling Biases

A machine learning model can only be as biased as the humans who frame it.

The datasets have to be prepared or have to be collected from some source which is a collateral of human interactions. The collected data will be cleaned and appended with classes. These sub-groups no matter how unbiased they were planned to be, there still lies an underwritten, underlying bias. To remove prejudices from a model might not be an impossible task but can any application which serves humans be immune to the human itself. And, even if the human element is considered, how much of it is too much.

Subscribe to our Newsletter

Join our editors every weekday evening as they steer you through the most significant news of the day, introduce you to fresh perspectives, and provide unexpected moments of joy
Your newsletter subscriptions are subject to AIM Privacy Policy and Terms and Conditions.

The typical life cycle of deployment machine learning models involves a training phase, where a typical data scientist develops a model with good predictive based on historical data. This model is put into production with the hope that it would continue to have similar predictive performance during the course of its deployment.

But there can be problems associated with the information that is deployed into the model such as:

  • an incorrect model gets pushed
  • incoming data is corrupted
  • incoming data changes and no longer resembles datasets used during training

The notions of machine learning fairness can be bottled down to the following facets of data pre-processing:

  • Demographic parity
  • Equal opportunity
  • Equalized odds
  • Disparate impact

Machine learning engineers work around bias or the offsets in a model by drawing insights from the output, gauging the losses,gouging through tonnes of data and repeating till a agreeable results have been obtained.

This is a traditional process which takes time but works decently. An alternative to this approach to this is the Lagrangian approach, a mathematical method to find the local maxima and local minima of a function when provided with equality constraints. This too, comes with its own set of complexities.

Now, the researchers at Google, tackle this problem of bias in labeling, provide a mathematical formulation how biases arise in labeling and how can this be mitigated.  They propose a new framework to model how bias can arise in a dataset, on the assumption that there exists an unbiased ground truth.

The bias is corrected by re-weighting the training examples. These approximated weights are then tweaked for generating fair classifiers.

This framework can also be applied to settings where the features too, are subjected to bias.

Source: RaniHorev

The bias correction takes place as follows:

  • It is assumed that the biased dataset (y_bias) is the result of a manipulation of a (theoretical) unbiased dataset y_true.
  • Learn the values of λk, from constraint violations, which represent the connection between y_true and y_bias.
  • The learned λk values are used to calculate the weight wk of each training sample.
  • Biased samples get low weights and unbiased samples, high.
  • These weights are used to train unbiased classifier.
  • Evaluate the constraints from the loss function and retrain the classifier with new weights.

The above figure illustrates the approach to training an unbiased classifier and the researchers assumption of an unknown label and how it has been adjusted to produce labels.

This new method doesn’t modify any labels but instead the bias is corrected by changing how sample points are distributed across the dataset.

The researchers behind this new framework Heinrich Jiang and Ofir Nachum, tested their model on data from datasets like Bank marketing, communities and crime where each datapoint represents a community and the task is to predict the crime rate in a community and other datasets for predicting credibility of customers for issuance of credit cards.

And, when trained on MNIST dataset, this model outperformed unconstrained and Lagrangian methods. This framework is platform independent and can be used to reinforce machine learning fairness.

Know more about the work here.


Ram Sagar
I have a master's degree in Robotics and I write about machine learning advancements.

Download our Mobile App


AI Hackathons, Coding & Learning

Host Hackathons & Recruit Great Data Talent!

AIM Research

Pioneering advanced AI market research

Request Customised Insights & Surveys for the AI Industry


Strengthen Critical AI Skills with Trusted Corporate AI Training

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

AIM Leaders Council

World’s Biggest Community Exclusively For Senior Executives In Data Science And Analytics.

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox