MITB Banner

OpenAI Inches Closer to AGI, Reduces Hallucinations

OpenAI’s new process supervision training is said to improve math reasoning with human-like thinking, and reduce hallucinations. Is this a step closer to AGI?

Share

Listen to this story

A Math teacher’s keen interest in checking the steps to solve a problem rather than the result loosely forms the basis for OpenAI’s new training approach. The company announced a new technique on training the model through process supervision by rewarding each step of correct reasoning as opposed to rewarding the correct final result via outcome supervision. 

Here the output would probably be a model with reduced hallucinations and higher alignment as claimed by the company. OpenAI specifically calls out mitigating hallucinations as a crucial step towards ‘building aligned AGI’, but would any of these new training methods inch them closer to an AGI status? 

Source: Twitter

Hallucinations At Bay

OpenAI talks about how you can train models to detect hallucinations either by process supervision, a method to provide feedback for each individual step, or outcome supervision, where the feedback is based on a final result. The company claims to have improved mathematical reasoning with the former method. By rewarding the model at each correct step, the model is said to mimic ‘human reasoning’ while solving a mathematical problem. 

With an emphasis on hallucinations, the company’s move towards ‘claiming’ to make the models more robust continues. Companies are actively working on reducing hallucinations. Recently, NVIDIA released NeMo Guardrails, an open-source toolkit that will help LLM-based applications become accurate, appropriate, and secure. With hallucinations considered a persisting problem with chatbots that often make them behave illogically by generating misinformation or biases, OpenAI is working on making their models become better. 

With the new training method, the company is hoping to keep a check on hallucinations as they believe a process-oriented method that involves feedback at each step, will control the irrational outputs generated by the chatbots. 

Alignment — Closer to AGI? 

OpenAI’s reference to ‘building an aligned AGI’ is hinting at the company’s long-term plans for achieving it. Looking back, Sam Altman has made multiple mentions of AGI and how the future will look with it. A few months ago, he laid out an elaborate AGI roadmap for OpenAI where its dangers were called out. The company believed that AGI can be misused and lead to grave consequences in society. However, despite these risks, the potential and benefits of it are far-reaching therefore, the company will develop it in a ‘responsible way’. AI expert Gary Marcus predicts that AGI will not be coming soon. 

It is interesting to note that Altman’s stance on AGI and its development is not clear-cut. In yesterday’s tweet, Altman seemingly downplayed the risk of AGI by predicting how ‘a much faster rate of change’ is what AGI will bring in. He believes that with AGI the future will unfold similarly to that without it, and the difference will be the speed with which things unfold – “everything happens much faster”. 

Ironically, Sam Altman along with AI scientists, Geoffrey Hinton, Yoshua Bengio, and many others, signed a statement a few days ago, which stands for safeguarding against the threat of extinction posed by AI, and considers it on par with nuclear war. If any action is to be taken on this, then the question that would arise is: how far will OpenAI go to make more advanced models reach AGI? 

The recent statement is in continuation with the open letter that was signed by over 31k people two months ago, including Elon Musk, Gary Marcus and other tech experts, who were urging for a pause on advanced AI models, which, interestingly, was not signed by Sam Altman. Though Altman had confirmed a month ago that the company will not work on building their next superior model GPT-5, and instead focus on the safety features of their existing models, his constant sway in matters pertaining to AGI threats and downplaying its scope makes it difficult to gauge where the company is headed. 


The company, often criticised for data security threats and privacy concerns, is fighting hard to prove ChatGPT as a foolproof chatbot. The company is now working on democratising AI by offering grants to those who can propose the best method for creating an AI regulatory framework — again with the hope of improving the system and seeming compliant with the world.

Share
Picture of Vandana Nair

Vandana Nair

As a rare blend of engineering, MBA, and journalism degree, Vandana Nair brings a unique combination of technical know-how, business acumen, and storytelling skills to the table. Her insatiable curiosity for all things startups, businesses, and AI technologies ensures that there's always a fresh and insightful perspective to her reporting.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.