21st-may-banner design

NVIDIA Introduces ChatQA, GPT-4 Level Conversational QA Models

ChatQA-70B stands out by surpassing GPT-4 in average scores on 10 conversational QA datasets (54.14 vs. 53.90).

Share

NVIDIA researchers recently introduced ChatQA, a family of conversational question answering (QA) models, aiming to achieve GPT-4 level accuracies.

NVIDIA’s ChatQA introduces a range of models, ranging from 7B to 70B in size. Extensive evaluations across 10 conversational QA datasets reveal that the top-performing ChatQA-70B model not only outperforms GPT-3.5-turbo but also performs at par with GPT-4. What’s notable is that these achievements are attained without relying on any synthetic data from ChatGPT models.

The team behind ChatQA proposes a two-stage instruction tuning method, significantly enhancing zero-shot conversational QA results from large language models (LLMs). To address retrieval in conversational QA, a dense retriever is fine-tuned on a multi-turn QA dataset, delivering comparable results to state-of-the-art query rewriting models with reduced deployment costs.

NVIDIA showcases the effectiveness of fine-tuning a single-turn query retriever using their curated conversational QA data. This approach performs comparably to the state-of-the-art LLM-based query rewriting model, all achieved without the need for extra computational time and potential API costs associated with rewriting.

NVIDIA’s ChatQA has demonstrated a remarkable advancement in handling scenarios where answers are elusive. The incorporation of a small number of “unanswerable” samples has proven to significantly enhance the model’s capabilities. Through an evaluation of unanswerable cases, it is evident that the leading model, ChatQA-70B, exhibits only a slight performance gap when compared to the formidable GPT-4. 

For more details, refer to the paper: ChatQA Paper.

NVIDIA is not alone. Several foundational models have achieved the capabilities of GPT-4. Google is likely to launch Gemini Ultra at any moment. Meanwhile, Mistral CEO Arthur Mensch announced on French national radio that the company will unveil an open-source GPT-4-level model in 2024. 

Share
Picture of Siddharth Jindal

Siddharth Jindal

Siddharth is a media graduate who loves to explore tech through journalism and putting forward ideas worth pondering about in the era of artificial intelligence.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.