AIM Banners_978 x 90

GPT-3 has Set a Benchmark, But It’s Not What You Think

The model trained on 2.7 billion parameters is much more accurate than the model trained on 175 billion parameters.
When GPT-3 came online, it was the talk of the town, and it still is. The reason is primarily the huge dataset the large language model (LLM) was trained on. Many upcoming LLMs consider GPT-3 as a benchmark while launching their product; for instance, recently, MosaicML put out a blog titled, ‘Mosaic LLMs (Part 2): GPT-3 quality for <$500k’.  But is the good old LLM really the benchmark?  It can’t be possible for an LLM model like GPT-3, which is based on a huge dataset, to be the benchmark for the upcoming language learning models. The model is not even showing better performance in terms of truthfulness when compared to the earlier LLMs like GPT-Neo/GPT-J, GPT-2, and T5-based models.  For example, if we compare the percentage of truth from various GPT-3 mode
Subscribe or log in to Continue Reading

Uncompromising innovation. Timeless influence. Your support powers the future of independent tech journalism.

Already have an account? Sign In.

📣 Want to advertise in AIM? Book here

Picture of Lokesh Choudhary
Lokesh Choudhary
Tech-savvy storyteller with a knack for uncovering AI's hidden gems and dodging its potential pitfalls. 'Navigating the world of tech', one story at a time. You can reach me at: lokesh.choudhary@analyticsindiamag.com.
Related Posts
AIM Print and TV
Don’t Miss the Next Big Shift in AI.
Get one year subscription for ₹5999
Download the easiest way to
stay informed