MITB Banner

OpenAI Raises Alarm Over Open Source AI Dangers 

Jan Leike, ML researcher & alignment team lead at OpenAI raised alarm over open-source models going rogue. However, he attracted some flack

Share

Listen to this story

Imagine a world where a know-it-all entity, which is able to propagate an unfathomable amount of information, text, images, voices, and videos becomes sentient and resists attempts at being stopped. Sounds like a 1984-esque scenario, doesn’t it? 

Jan Leike, ML researcher and alignment team lead at OpenAI in a tweet, painted a doomsday picture. He highlighted a significant challenge for humanity — the decision to not release open-source LLMs that can autonomously propagate. Leike emphasised that if these LLMs spread, they could engage in various criminal activities that would be difficult to control and assign blame to.

He tweeted, “An important test for humanity will be whether we can collectively decide not to open source LLMs that can reliably survive and spread on their own.” “Once spreading, LLMs will get up to all kinds of crime, it’ll be hard to catch all copies, and we’ll fight over who’s responsible.”

Is the Threat Real?

If LLMs and artificial intelligence was to go wrong and was able to fend off any attempts to stop it, it would truly be a menace to society at large. A workshop to assess the potential impact of AI on future criminal activities, brought together 31 experts from diverse backgrounds to categorise AI-related threats and gauge their severity over a 15-year period.

The discussions’ outcomes highlighted high potential threats like audio/visual impersonation, driverless vehicles as weapons, targeted phishing, AI-controlled system disruption, and large-scale blackmail. Medium-level threats encompassed military robots, data poisoning, and autonomous attack drones, while bias exploitation and evading AI detection were classified as low-level threats.

While the dangers seem unlikely at present, there’s a need to balance the doom and gloom, while keeping an eye out for the ‘ENTITY’.

Fending off against open-source? 

But is it so or is it just a ploy to hamper the development of open-source models like LLaMA 2 which are stealing the thunder from OpenAI? Clement Delangue, co-founder & CEO of Hugging Face, responded to Leike’s post, pointing out that the tweet could be interpreted as using fear to undermine open-source practices. 

Delangue then addressed Leike’s point, noting uncertainty about the concept of technology “surviving and spreading on its own.” He went on to raise a very pertinent question of whether the act of creation itself posed a risk (or open-sourcing it was the problem) since if a technology is truly exceptional, it should naturally find avenues for widespread adoption without solely relying on open-sourcing.

In response to the ongoing discussion, another tweet contributed a perspective. It emphasised the inevitability of closed-source Language Models as well, eventually, being exposed to leaks. 

While the potential for various manipulations arises when LLMs are open-sourced, including actions that could undermine their intended purpose—by implementing a comprehensive alignment process, the act of “de-aligning” LLMs—altering their intended behaviour—could be rendered moot. The process of intentionally disrupting an aligned LLM’s behaviour could still be made significantly difficult through meticulous alignment procedures. However, there are challenges with it.

OpenAI recently released a paper addressing superalignment authored by Lieke and Ilya Sutskever, which pertains to the challenge of ensuring that superintelligent artificial intelligence systems are aligned with human values and intent. Superintelligence, the hypothetical level of AI capability far surpassing human intelligence, holds the potential to revolutionize various sectors and address critical global issues. However, it also poses substantial risks, including the potential for human disempowerment or extinction. It seems like the firm is employing a narrative to give a little nudge to it.

Two Sides of OpenAI

Nonetheless, there are others like Andrej Karpathy who is an active contributor to the open-source ecosystem. He recently built a ‘Baby Llama’ model based on Meta’s Llama 2, which has been critically celebrated in the developer ecosystem. 

It seems like there are two narratives within OpenAI itself, one that wants to capitalise on the momentum from the release of ChatGPT, make moolah and remain closed-source. Others genuinely want to make a difference and haven’t steered a long way from their initial open-source days—but later turned towards a closed source-for-profit company, in the words of one of its founders.

Share
Picture of Shyam Nandan Upadhyay

Shyam Nandan Upadhyay

Shyam is a tech journalist with expertise in policy and politics, and exhibits a fervent interest in scrutinising the convergence of AI and analytics in society. In his leisure time, he indulges in anime binges and mountain hikes.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.