MITB Banner

GPT-3 prompts: Technical progress or just AI alchemy?

Frankly, calling this engineering is offending to engineers. It's chaos alchemy.

Share

Illustration by Analytics India Magazine

Listen to this story

Patrick Krauss, professor at Friedrich-Alexander-University Erlangen-Nuremberg (FAU), has called out a paper: “Large Language Models are Zero-Shot Reasoners” on Twitter. The paper claimed prompts increase the accuracy of GPT-3. 

The chain of thought (CoT) prompting, a technique for eliciting complex multi-step reasoning through step-by-step answer examples, achieved state-of-the-art performances in arithmetics and symbolic reasoning, the paper claimed. “We create large black boxes and test them with more or less meaningless sentences in order to increase their accuracy. Where is the scientific rigor? It’s AI alchemy! What about explainable AI?” Patrick said.

With 58 papers on LLM published on arxiv.org alone in 2022 and the global NLP market projected to reach USD 35.1 billion by 2026, LLMs are one of the thriving areas of research.

Chain of Thought prompting

The idea was proposed in the paper, “Chain of Thought Prompting Elicits Reasoning in Large Language Models”. The researchers from Google Brain team utilised a chain of thought prompting – a coherent series of intermediate reasoning steps that lead to the final answer for a problem, to improve the decision making capability of the large language model. They demonstrated that sufficiently large language models could generate chains of thought if demonstrations of the chain of thought reasoning are provided in the exemplars for few-shot prompting.

Source: arxiv.org

To test their hypothesis, the researchers used 3 transformer-based language models: GPT-3 ( Generative Pre-trained Transformer), PaLM(Pathways Language Model) and LaMDA (Language Model for Dialogue Applications). The researchers explored the chain of thought prompting for various language models on multiple benchmarks. Chain of thought prompting outperformed standard prompting for various annotators and different exemplars.

Zero-Shot COT

Researchers from the University of Tokyo and Google Brain team, improved on the chain of thought prompt method by introducing Zero-shot-COT (chain of thought). LLMs become decent zero-shot reasoners with a simple prompt, the paper claimed.

Source: arxiv.org

The results were demonstrated by comparing the performances on two arithmetic reasoning benchmarks (MultiArith and GSM8K) across Zero-shot-CoT and baselines. 

AI alchemy

Patrick’s tweet sparked a huge debate. “It is an empirical result, which adds to our understanding of these black boxes. Empiricism is a standard, well established approach in science, and I find it surprising this is new to you.” @Dambski further states that this discussion is subjective to what one considers to be the definition of understanding. Anything that increases the chances of the model correctly predicting how it will behave for a given input increases the understanding of that system, whether It can be explained or not,” said Twitter handle @Dambski.

Rolan Szabo, a machine learning consultant from Romania,, gave another analogy: “From a theoretical perspective, I understand the disappointment. But from a pragmatic perspective, Github Copilot writes the boring boilerplate code for me today, even if I don’t understand how exactly it conjures it up.”

Many supported Patrick’s statement. Piotr Turek, head of engineering, OLX group said: “Frankly, calling this engineering is offending to engineers. It’s chaos alchemy”

Soma Dhavala, principal researcher at Wadhwani AI, said: While we think we solved one problem — we made it somebody else’s problem or problem resurfaces in a different avatar. Case-in-point: With DL we don’t need feature engineering, was the claim.  Well yah, but we got to do architecture engineering.”

Guillermo R Simari, a professor emeritus in Logic for Computer Science and Artificial Intelligence, said: “I’d not be entirely against the approach. My concern is: What will we’ve learned about the thinking process at the end? Will I understand the human mechanism better? Or have I just got something that “works”? Whatever that means…” To which, Patrick Krauss said that’s exactly his point.

The discussion took a turn when Andreas K Maier, a professor at Friedrich-Alexander-University Erlangen-Nuremberg (FAU), asked whether such large language models are available for public access so that one can actually observe what is happening in the latent space during inference. 

To this comment, Patrick said the unavailability of LLMs is exactly the problem. “One problem is of course that some of these models are only available as API. Without access to the actual system it might become something like AI Psychology, ” Andreas added. As of now, Meta AI’s Open Pretrained Transformer (OPT-175B), is the largest LLM with open access.

Share
Picture of Kartik Wali

Kartik Wali

A writer by passion, Kartik strives to get a deep understanding of AI, Data analytics and its implementation on all walks of life. As a Senior Technology Journalist, Kartik looks forward to writing about the latest technological trends that transform the way of life!
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.