MITB Banner

AI Does Not Hate You, Nor Does It Love You: Understanding The Value Alignment Problem

Share

Eminent British poet, Tom Chiver, in his book titled, The AI Does Not Hate You: Super-intelligence, Rationality and the Race to Save the World, talks at length about several reasons which could simulate the route of human extinction. Other than a terrifying looking AI, similar in appearance to the T-800 from Terminator II, one of them is the creation of genetically engineered viruses through the means of synthetic biotech. Overgrowing concerns with the latter, goes unnoticed with non-availability of sustainable investments for commencing any kind of research. 

The dystopian epoch, we all are experiencing should primarily raise alarming signals at our defunct heuristics and the failure to adopt holistic policies for a genetically engineered pandemic. As a matter of consideration, the validity of my opinions shall be realized, if we could presume that we are going to overcome the COVID-19 pandemic, otherwise this work is as good as a fictional output.

The problem of value alignment with augmented technological tools gained ethical relevance with Isaac Asimov’s Three Laws of Robotics. The laws which are now 70 years old, were able to manage effective levels of control in artificial agents like Bertie the Brain, capable of tic-tac-toe game in the 1950s. Fast forward to 2016, AlphaGo, beats Lee Sedol in a game that has more possible positions than the number of atoms in the universe, merely powered with deep learning algorithms. 

With unsurmountable development in the domain of artificial intelligence over the past two decades, the same laws appear to have lost their charm to maintain robust value aligning rigidity between the machine and us.  Consistent research backed by computational power, in the field of artificial general intelligence shall evolve the concept of intelligence in the near future. Hypothetically a machine like AlphaGo might not only be good at playing Go, it may extend its capacity of purpose in other domains of operations, like policy making.

Amidst such a transitory phase, the evolution of intelligence in artificial agents will challenge the systemic inheritance of control over them. It is a concomitant responsibility of its creators to ensure that losing control over its agents shall never lead to incorrigible consequences for us. This ideology is the foundation of the value alignment problem; in layman’s terms, we must be sure that the purpose desired by the creator is the same purpose an artificial agent seeks to achieve. The central logic revolves around artificial agents being made to act in altruistic ways so that it can never pose a threat to us. However, despite having an altruistic purpose, the agent can make decisions, backed with general intelligence, which will end up contradicting the purpose envisioned by the creator.

This shall be understood better if we take a glance at the paperclip maximizer experiment. The experiment was brought in the discussion forum by Swedish Philosopher, Nick Bostrom, stating how an artificial general intelligence agent tasked with the objective of collecting paper clips, in order to maximize its collection will eventually turn earth into a paper clip.

An artificial agent with human-level intelligence, initially, will take it upon itself to either collect those paper clips, buy those paper clips with money or manufacture them by itself. 

Inevitably, the agent shall go through a process of intelligence explosion, where it shall deduce newer and more effective methods of collecting paper clips. The agent shall never move away from its goal of collecting paper clips and in doing so, it will not be bothered about non-specific goals such as learning human sentiments or developing collective unconsciousness. Over here, we conclude that the values of an artificial agent are completely abnormal and alien to the multitude of values central to us. Therefore, in its process of optimization as a goal-seeker it will be exactly focused on the aspect of maximizing utility, which is the collection of paper clips in this instance. 

As an anthropological discourse, the value alignment problem, if viewed from the prospects of Asimov’s three laws of Robotics, appears to formulate a shield with respect to an existential threat to us by the acts of an artificial agent. This premise, irrespective of the nature of control on artificial agents, will cause serious hindrance in the practice of developing artificial agents that can co-exist with us and serve its purpose of being an augmented tool to improve our experience of existence. Beyond the norms of existential threat, the value alignment problem today, must focus on inculcating values in an artificial agent which shall not only cause an existential threat but it shall never be a cause of obstruction for us to flourish in our lifetime.

Therefore, the duty of the creator of an artificial agent is not over when it has acted in all good faith, inheriting altruistic values while devising its purposes. The duty extends to making the artificial agent aware of our behavioral competencies, so that the agent doesn’t intervene with our ability to prosper in its presence. When a social media platform mechanizes machine learning tools to scrap your private data, for targeted advertisement purposes, it might not be causing a direct existential threat to your identity but it is surely in violation of your freedom to flourish. 

Due to our heuristically challenging abilities of decision making and abysmally large population, it is important for the artificial agent to keep learning the values it should align with and not be limited to a static accord of values.

The presence of malign information from media portals, reflecting this paradigm as a war between us and artificial agents has already penetrated a pleading for greater control over these agents. However, the determinative motive shall be to accept its autonomy and secure a medium of co-existence by making them align with our ever thriving values. The process of automation can’t be rejected and therefore, any efforts to threaten its legitimacy by framing us with the idea of feeling oppressed will cause a blockade to the beautiful destiny of augmented reality. Honestly, there’s no chance of competing on a levelled pedestal with them.

A biological neuron fires, maybe, at 200 hertz, 200 times a second. But even a present-day transistor operates at the Gigahertz. Neurons propagate slowly in axons, 100 meters per second, tops. But in computers, signals can travel at the speed of light. There are also size limitations like a human brain that has to fit inside a cranium, but a computer can be the size of a warehouse or larger. So the potential for super intelligence lies dormant in the matter, much like the power of the atom lay dormant throughout human history, patiently waiting there until 1945.


REFERENCE

Nick Bostrom, Superintelligence: Paths, Dangers, Strategies, Oxford Publication, (2014).

Stuart Russell, Human Compatible: AI and the Problem of Control, Viking Press, (2019).

Stuart Russell, Peter Norvig, Artificial Intelligence: A Modern Approach, Pearson Education, (2016).

Tom Chiver, The AI Does Not Hate You: Superintelligence, Rationality and the Race to Save the World, Orion Publishing Group, (2019).

PS: The story was written using a keyboard.
Picture of Sujoy Sarkar

Sujoy Sarkar

Sujoy Sarkar is a third-year student pursuing B.A. LL.B from National Law School of India University, Bangalore. He aspires to become a Computer Programmer and experimental music producer. He's an avid supporter of cryptocurrencies and the objectives of singularitarianism.
Related Posts

Download our Mobile App

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

3 Ways to Join our Community

Telegram group

Discover special offers, top stories, upcoming events, and more.

Discord Server

Stay Connected with a larger ecosystem of data science and ML Professionals

Subscribe to our Daily newsletter

Get our daily awesome stories & videos in your inbox
Recent Stories

Featured

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

AIM Conference Calendar

Immerse yourself in AI and business conferences tailored to your role, designed to elevate your performance and empower you to accomplish your organization’s vital objectives. Revel in intimate events that encapsulate the heart and soul of the AI Industry.

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed