Jeffrey Ullman’s Unsettling Ultimatum 

People in AI research who could be excellent teachers are basically forced to do second and third grade research because that's how they get promoted
Listen to this story

“Looking back, it’s amazing how easy things were for researchers when I was a young man. In comparison to just how competitive the field has become,” said the 80-year-old American computer scientist Jeffrey Ullman in an exclusive interview with AIM. 

Recalling the state of research in the 60s he said, everywhere you looked, there was something new. Now there are a lot of new directions that come up, but so much of research these days is squeezing a little bit of performance out of something just to get a little improvement.

Further, talking about research he said, “A part of the problem is that too many people are led to believe that the success in life depends upon them doing research. What has happened is academics who could be excellent teachers are basically forced to do second and third grade research because that’s how they get promoted.” 

Subscribe to our Newsletter

Join our editors every weekday evening as they steer you through the most significant news of the day, introduce you to fresh perspectives, and provide unexpected moments of joy
Your newsletter subscriptions are subject to AIM Privacy Policy and Terms and Conditions.

Ullman won the Turing Award in 2020 for co-creating the building blocks of computer programming in the early 60s along with his long time friend Alfred Aho. The duo refined one of the key components of a computer: the “compiler”.  

Bias In, Bias Out 

Ullman’s contributions to computer science have a direct impact on the foundations of data science. His research bedrocked many key concepts and techniques employed in the storage, analysis, and interpretation of data, making him an influential figure.

Emphasising on the relevance of data science he said, “The thing that makes machine learning powerful is that you can feed so much data to these interesting models. For example, multi-layer neural nets only make sense if you look at millions or billions of training examples.” He also addressed that “for years people have been trying to exploit data to solve problems”.

Clarifying the bias problem in language models, he said, “Often people think the machine learning algorithms introduce bias. 50 years ago, everybody knew ‘garbage in garbage out’. In this particular case, it is ‘bias in, bias out’. 

He stated a canonical example of a company which decided to write software to sift through resumes of people applying for jobs. “If the data that comes from this company shows women have not been promoted in proportion to their ability then the AI is going to learn that being female is a negative indication of success, and they’re going to throw out females. If it’s a good algorithm, it’s going to learn and pick that up,” he said. 

The fault is not with the AI, it is with the company policy and the data that policy has generated Ullman believes. “If they built an LLM, in the time of Galileo, the chat would have said that the sun revolves around the Earth because everybody believed that,” he chuckled.

ChatGPT and Eric Schmidt

“I am surprised you didn’t ask if these language models are becoming sentient,” he said jokingly. “The deceptive thing is because the people who created the text on which these models are built are sentient. The LLM is regurgitating the consensus of what sentient beings say it is going to sound sentient,” he added. 

A couple of days ago Ullman heard Eric Schmidt talking on CNN about the problem of what happens when you ask ChatGPT what’s the best way to kill a million people? 

The former Google CEO said that it wouldn’t answer that question. It knows not to let that question be answered. But that doesn’t mean you can’t rephrase it in such a way that you can get some really dangerous information out of that. 

“It’s a very hard problem. If you understand that there is something these generative models shouldn’t be saying. You can stop them from saying it. But how do you understand everything that could cause trouble?” Ullman pondered. 

Drawing parallels, he recalled the instance when social networks tried to get rid of hate speech on their platforms. The tech giants have put efforts for decades “but how to write an algorithm that disallows anything that we would all agree as hate speech and allows anything that we would all agree is not hate speech” still remains an unsolved issue. 

Current Arguments

Ullman has recently heard the argument that the community is running out of data. Discussing the subject, he said ”I’ve heard that the newest models GPT-4 essentially use everything. “At some point, they will run out of new data. You’ve got 8 billion people. Let’s say half of the people even type around 100,000 words a year. You’ve got a couple of 100 trillion words of data and that might support significantly more than a trillion parameters model.”

Ullman believes that there are many occupations where it could be used as a tool but the idea that lawyers and physicians are going to be completely replaced by these LLMs. “I do not think it’s going to happen,” he said.  

“In fact, what I suspect will happen is exactly what happened when tools like Microsoft Word became available. You’ll spend just as much time writing or doing your job, but the output will be better quality. The jobs won’t really go away, the results may be better if these models are used,” he concluded.  

Tasmia Ansari
Tasmia is a tech journalist at AIM, looking to bring a fresh perspective to emerging technologies and trends in data science, analytics, and artificial intelligence.

Download our Mobile App

MachineHack | AI Hackathons, Coding & Learning

Host Hackathons & Recruit Great Data Talent!

AIMResearch Pioneering advanced AI market research

With a decade of experience under our belt, we are transforming how businesses use AI & data-driven insights to succeed.

The Gold Standard for Recognizing Excellence in Data Science and Tech Workplaces

With Best Firm Certification, you can effortlessly delve into the minds of your employees, unveil invaluable perspectives, and gain distinguished acclaim for fostering an exceptional company culture.

AIM Leaders Council

World’s Biggest Community Exclusively For Senior Executives In Data Science And Analytics.

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox
MOST POPULAR