Top ML Projects To Fight Fake News Fatigue During COVID-19

The role of fake news has had violent outcomes in the past and continues to do so. Especially today, amid a pandemic, social media platforms are being used to dish out misinformation at lightning speed. One thing we can do is to avoid news altogether or use tools such as those of machine learning to fight the fatigue of fake news. Here, we list a clutch of interesting and relevant projects from GitHub based on their rating. Get hands-on with the available code: 

Use Fake News Generator As The Detector

Grover is a model developed to generate text in a controllable way. If a prompt such as `Link Found Between Vaccines and Autism’ is given, then the Grover model can generate the rest of the article. This sounds like a great premise for anyone looking to automate fake news generation. However, as the creators claim, the best defense against Grover turns out to be Grover itself. This project makes a strong case for having strong generators open-sourced. Grover produces results with 92% accuracy and can help pave the way for better detection of neural fake news.

Check the code here.

AIM Daily XO

Join our editors every weekday evening as they steer you through the most significant news of the day, introduce you to fresh perspectives, and provide unexpected moments of joy
Your newsletter subscriptions are subject to AIM Privacy Policy and Terms and Conditions.

Making A Fake News Tracker

A tool by the name FakeNewsTracker, an extension of the FakeNewsNet that was released a couple of years ago, contains a repository for collecting, analysing, and visualising fake news and the related dissemination on social media. This work by researchers at Arizona State University is a comprehensive review of detecting fake news on social media, including fake news characterisations on psychology and social theories, existing algorithms from a data mining perspective, evaluation metrics, and representative datasets. We also discuss related research areas, open problems, and future research directions for fake news detection on social media.

Check the code here.

Download our Mobile App

Fake News Detection As Natural Language Inference

The project takes sentences into three parts. The first sentence is the title of an article already known to be fake news. The second sentence is the title of another article, and the task is to decide whether it agrees with the original fake news, disagrees with it, or is unrelated. The tasks are treated as natural language inference (NLI). As illustrated above, all the strong models, such as BERT, were also incorporated during the training phase. These results are assembled and retrained with noisy labels.

Check the code here.

Fake News Detection On Twitter Dataset

For this project, a multi-modal feature extractor was used, which extracts the textual and visual features from posts. For this project, adversarial neural networks are implemented, and the feature extractor cooperates with the fake news detector to learn how to detect the key features of fake news. The discriminator network removes the event-specific features and keeps shared features among events. For this project, multimedia datasets from Weibo and Twitter were used.

Check the code here.

Fake News Detection via Reinforcement Learning

A reinforced weakly-supervised fake news detection framework was proposed that leverages users’ reports in a weakly supervised manner to enlarge the amount of training data for fake news detection. The framework consists of an annotator, the reinforced selector, and the fake news detector. 

The annotator is used to assign weak labels for unlabeled news based on users’ reports. Whereas, the reinforced selector uses reinforcement learning techniques to choose high-quality samples from the weakly labeled data and remove low-quality ones that may degrade the detector’s prediction performance. The data used to test for this project is obtained from news articles published via WeChat official accounts and associated user reports. 

Check the full work here.

Model Trained On NYT & The Guardian

This project contains scraped news from NYT API and The Guardian API to have a data set labeled as real news. Whereas, the fake news dataset has been downloaded from There are 12,000 fake news articles from and 43,000 real news. 

Real and fake news articles had to be in certain topics and the creators have decided to use: “US News,” “Politics,” “Business,” and “World,” assuming that most fake news would be from these topics.

Check the full code here.

Sign up for The Deep Learning Podcast

by Vijayalakshmi Anandan

The Deep Learning Curve is a technology-based podcast hosted by Vijayalakshmi Anandan - Video Presenter and Podcaster at Analytics India Magazine. This podcast is the narrator's journey of curiosity and discovery in the world of technology.

Ram Sagar
I have a master's degree in Robotics and I write about machine learning advancements.

Our Upcoming Events

24th Mar, 2023 | Webinar
Women-in-Tech: Are you ready for the Techade

27-28th Apr, 2023 I Bangalore
Data Engineering Summit (DES) 2023

23 Jun, 2023 | Bangalore
MachineCon India 2023 [AI100 Awards]

21 Jul, 2023 | New York
MachineCon USA 2023 [AI100 Awards]

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox

Council Post: Evolution of Data Science: Skillset, Toolset, and Mindset

In my opinion, there will be considerable disorder and disarray in the near future concerning the emerging fields of data and analytics. The proliferation of platforms such as ChatGPT or Bard has generated a lot of buzz. While some users are enthusiastic about the potential benefits of generative AI and its extensive use in business and daily life, others have raised concerns regarding the accuracy, ethics, and related issues.