MITB Banner

Beyond BERT: Rethinking Search In The Age of AI

Share

“Before about 2018, technologies that could answer questions tended to be brittle, hard to build, and even harder to scale. The advent of neural language models changed that.” 

Khattab et al.

Modern day search engines are powered by incredibly large language models (eg: BERT) that do well with Q&A format. Current language models have already breached the 1 trillion parameters mark and are poised to get bigger and smarter. These language models(LMs) are trained on massive quantities of unstructured data. One of the main reasons why large pre-trained LMs are so successful is that they learn highly effective contextual representations.

But, how do we know the results are right? It is true that the internet has made it easier to retrieve information, thanks to the ranking algorithms. But the lack of provenance of these search results allows misinformation to flourish.  

A group of researchers from Stanford University have proposed models that can enable AI to comprehend and respond in the most efficient way possible. The researchers wrote that users would usually ask follow-up questions to their initial queries to arrive at actual facts. One might Google the year of establishment of Stanford and then will follow up by asking, “What is the source for this?” and it might bring up “The Stanford University About page.” This does help, but one might wonder if the model had instead produced the text from “Wikipedia”? 

Even if the larger language models find that sweet spot for retrieval and contextualisation, they might still find information removal to be challenging. We have come across news where the regulatory authorities of the state or even an individual can ask the internet companies to get rid of sensitive data. Data obsolescence can render models useless. According to the researchers at Stanford, there is no reliable way to remove specific information from language models as a direct consequence of the way information is stored. “Applying simple keyword filters on their outputs is not a panacea, as these models might still find ways of expressing the same content in new terms.”

This is where Neural IR models come in handy. According to the Stanford researchers, models such as ColBERT not only make the whole process of Q&A more relatable but also help maintain the effectiveness of the models in the event of deletion of information. Neual IR paradigm is said to be in use at least since 2019 by Google and Microsoft for their respective search engines and it is still a hot area of research.

“Neural IR is already being used by Google and Microsoft for their search engines.”

BERT vs ColBERT (Image credits: Paper by Khattab et al.,)

BERT significantly improved the search precision. But, it comes at a cost. The computation required increases latency. To avoid this latency, researchers started incorporating traditional retrieval methods into the language models. But, now precision has become a challenge. This is why models like ColBERT were introduced–to make the best of both worlds; efficiency and contextualisation. 

ColBERT is a ranking model based on contextualised late interaction over BERT. This model proposes a novel late interaction paradigm for estimating relevance between a query “q” and a document “d”. Under late interaction, wrote the researchers, q & d are separately encoded into two sets of contextual embeddings, and relevance is evaluated using cheap and pruning-friendly computations between both sets.

Image credits: Stanford HAI

Neural IR paradigm

  • Uses a pre-trained LM(eg: BERT) to encode documents and queries are encoded into numerical representations.
  • A scoring function compares representations to determine a ranked list of search results similar to that of a classical paradigm.
  • An extraction function then identifies relevant text spans to offer as direct evidence.
  • Complex queries are interpreted more deeply than in the classical paradigm.
  • Most relevant text from documents can be highlighted and the excerpts can be combined directly from multiple sources. 

Researchers are betting on the work pushing the NLP and IR frontiers. For example, Neural IR methods like ColBERT and Baleen can be used to improve search results while preserving trust and reliability. The researchers believe this area has a huge potential to revamp the way we use the internet. That said, they also admit that  language models are not, and will never be, all-knowing “infallible oracles”, even if they seem to behave that way when you ask them questions.

Share
Picture of Ram Sagar

Ram Sagar

I have a master's degree in Robotics and I write about machine learning advancements.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.