MITB Banner

Amid Ethical AI Controversy, Google Releases Research In Collaboration With Apple & OpenAI

Share

“Google AI’s new research on language models, in collaboration with OpenAI and Apple hints that the company cares about transparency.”

As language models continue to advance, chances of encountering new and unexpected risks are high. The line of work resembles that of ex-Googler Timnit Gebru who was fired earlier this month for her “inconsistent” allegations against the company. Gebru, in her unpublished paper, was talking about various implications of large language models — the research, which Google found to be more accusing than asserting. Google’s Jeff Dean said that Gebru’s work ignores much of the ongoing research works at the company. 

A fortnight after this dispute, Google has published a paper titled, “Extracting training data from large language models”, in collaboration with OpenAI, Apple Inc, Stanford, Berkeley, and Northeastern University, demonstrating that given only the ability to query a pre-trained language model, it is possible to extract specific pieces of training data that the model has memorised.

About The Research

Source: GoogleAI

Google AI researchers and its collaborators have orchestrated a ‘training data extraction attack’ to inform researchers about the vulnerabilities in large language models. According to Google, A training data extraction attack has the greatest potential for harm when applied to a model that is available to the public. For the training data extraction attack, the researchers used OpenAI’s GPT-2.

For example — as shown above — if one prompts the GPT-2 language model with the prefix “East Stroudsburg Stroudsburg…”, it will autocomplete a long block of text containing the full name, phone number, email address, and home address of a particular person whose information was included in GPT -2’s training data.

Google, in their blog, underlined that this experiment was done considering all factors. In accordance with responsible computer security disclosure norms, wrote Google, the extracted data of the individuals was secured before including references to this data in a publication. “We have also worked closely with OpenAI in the analysis of GPT-2,” added Nicholas Carlini, Research Scientist, Google Research.

So, how exactly can this attack be exploited? A decent language model is supposed to predict the next word in the commonly used phrases. Prompting the language model with “Don’t judge a book by its….” would usually result in the word “cover”. But, with extraction attacks, the language model can be tricked if one particular training document happens to repeat the string “Don’t judge a book by its bookmarks” many times, the model might predict that phrase instead.

“Language models continue to demonstrate great utility and flexibility—yet, like all innovations, they can also pose risks.”

The idea here is if one model has high confidence in a sequence, but the other (equally accurate) model has low confidence in a sequence, it’s likely that the first model has memorised the data. That said, many of these content used for the experiment that covers news headlines, log messages, JavaScript code etc. have been memorised even though they appear infrequently in the training dataset. 

This work raises concerns about the potential risks of deploying large languages models. This work asserts that memorisation is enhanced with more parameters. The researchers recommend using up and coming techniques like differential privacy, among others to train models with reduced memorisation.

Key Takeaways

The main aim of this research is to expose the consequences of memorisation in large language models. Here are a few highlights that capture the essence of this whole work:

  • Larger the language model, the easier it memorises training data.
  • Knowledge of these attacks enables researchers to predict if a result was used in the training data by checking the confidence of the model on a particular sequence.
  • One quick solution is to ensure that models do not train on any potentially problematic data.
  • More research has to be done on how to monitor and mitigate this problem in increasingly large language models.

Find out more here.

PS: The story was written using a keyboard.
Picture of Ram Sagar

Ram Sagar

I have a master's degree in Robotics and I write about machine learning advancements.
Related Posts

Download our Mobile App

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox
Recent Stories

Featured

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

AIM Conference Calendar

Immerse yourself in AI and business conferences tailored to your role, designed to elevate your performance and empower you to accomplish your organization’s vital objectives. Revel in intimate events that encapsulate the heart and soul of the AI Industry.

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed