MITB Banner

This USC Professor from Kolkata is on a Journey to Civilise AI

Amitava Das has been working in NLP for over 20 years and is currently hell-bent on reducing hallucinations within LLMs.

Share

The present-day AI system is smitten with hallucinations – for some, it is a feature (and not a bug), and for others, it’s frustration. For Amitava Das, a professor at the University of Southern Carolina, it was a worry that lead to the creation of what he calls ‘Civilising AI or Constitutional AI’.

“AIs [LLMs] are very smart and at the same time, they are also stupid. And we do not know how to handle these in either case. On the other hand, AI-generated content becomes so eloquent that it is impossible to detect it. No Turing test is able to do it,” said Das, in an interview with AIM, sharing the origin of Civilised AI, a framework to proactively measure the hallucination of a model, alongside mitigating and managing risks associated with it. 

With a PhD in natural language processing (NLP) from Jadavpur University in Kolkata, Das has been working in the field for more than 20 years, which he says was all because of his passion for working in AI ever since the beginning. Having two academic postdocs from Europe and the USA, he is currently a professor at the Artificial Intelligence Institute of South Carolina (AIISC).

Das started the AI research lab at Wipro and is still an advisor there. Having worked across geographies, and along with IITs and universities world over, he has also created a joint PhD program for industry practitioners to earn doctorate degrees. He’s currently focused on reducing hallucinations within language models. 

There are several research papers and models coming up to solve the AI detection problem. GPTZero, DetectGPT, and RoBERTa baseline are all trying to do the same, but a lot still needs to be done. The most recent paper, Ghostbuster, also works on solving the same problem.

That is why Das and his team coined the term ‘Civilised AI’ to explain how to handle the problem of hallucinations, and at the same time make it easier to detect if a text is AI generated, and tackle the misinformation and the possible harm it poses to the society. 

Single metric to civilise AI

Using the example of Google’s Bard and how the company lost around $100 million because of its hallucinations, Das emphasised the need to control what AI is writing. “On the one hand, we need AI to be creative when it comes to fields like journalism as it helps in writing. But on the other hand, if I am a lawyer working on a legal case, hallucination is not allowed at all,” he explained. 

Das’s paper, titled The Troubling Emergence of Hallucination in Large Language Models – An Extensive Definition, Quantification, and Prescriptive Remediations, addresses this problem of balancing hallucinations with eloquence within AI models, along with ‘Counter Turing Test CT^2: AI-Generated Text Detection is Not as Easy as You May Think — Introducing AI Detectability Index’, the paper that won the Outstanding Paper award at EMNLP 2023.

“We call this dichotomy a factual mirage and silver lining, how much of each is allowed,” explained Das, and that is what led his research team to introduce new metrics called AI detectability index, AI adversarial attack vulnerability Index, and AI hallucination index. 

“AI is just a balance between these three,” he added that there is still not a clear-cut solution for this, but the team is trying to derive a formula for this, which is called the Civilising Index, an amalgamation of all of this. 

He said it is a difficult scenario to tackle, illustrating with Einstein’s mass-energy equivalence relation, which led to the creation of the atomic bomb. He also gave examples of other AI-related research that needs to be regulated if it involves risk to human lives such as bioinformatics. 

“It is the people who created the atomic bomb and the people who created atomic energy. So there are people who should make AI models better, and then there is my team who will make better detection techniques,” he added.

“Regulation has to come to specific use case, not in a broad way,” he asserted. Each field such as medicine, military, legal, etc need to have separate licences on how they can use AI to tackle the hallucinations, leading to emerging risks within the systems. 

The need for Indic LLMs

Das is also passionate about how language models can be made for Indic languages and has released a code-mixing model called CONFLATOR for the purpose. He started working in this field around eight years back and is regarded as one of the pioneers of the field by Microsoft Research. 

He explained that the problem that comes with Indic languages and datasets is that it is very hard to understand. “Hindi is getting mixed into English or English is getting mixed into Hindi,” Das explained about how it is very difficult to understand the switching point in languages as most of the data is from the Internet.

“That is why we introduced Conflator,” he added, which is still a complicated field that deals with introducing electromagnetic waves into a neural network for code-mixing for positional encoding formulation. 

When it comes to Indic language models “we do not have a lot of monolingual data in India. This discussion has been going on between NITI Aayog and IIIE for several years, before the launch of ChatGPT and others,” Das was part of the discussions with the Indian government for making Indic language models possible. 

That is why he believes that Indian models are still far behind English models in terms of power and detectability. Das told AIM that he is closely working with the Indian government and IIIT Hyderabad and is soon going to publish a paper on the topic.

Share
Picture of Mohit Pandey

Mohit Pandey

Mohit dives deep into the AI world to bring out information in simple, explainable, and sometimes funny words. He also holds a keen interest in photography, filmmaking, and the gaming industry.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.