18 October 2024
Society is increasingly reliant on AI technologies to help make decisions. But this growing reliance comes with risk: AI models can produce biased, harmful and inaccurate outputs that are not aligned with their creators’ goals and original intent for the system.
Alignment works to reduce these side effects, helping ensure AI systems behave as expected and in line with human values and goals. For example, if you ask a generative AI chatbot how to build a weapon, it can respond with instructions or it can refuse to disclose dangerous information. The model’s response depends on how its creators aligned it.
Alignment often occurs as a phase of model fine-tuning. It might entail reinforcement learning from human feedback (RLHF), synthetic data approaches and red teaming.
However, the more complex and advanced AI models become, the more difficult it is to anticipate and control their outcomes. This challenge is sometimes referred to as the “AI alignment problem.” In particular, there is some apprehension around the creation of artificial superintelligence (ASI), a hypothetical AI system with an intellectual scope beyond human intelligence. The concern that ASI might surpass human control has led to a branch of AI alignment called superalignment.
Researchers have identified four key principals of AI alignment: robustness, interpretability, controllability and ethicality (or RICE).1
Human beings tend to anthropomorphize AI systems. We assign human-like concepts to their actions, such as “learning” and “thinking.” For example, someone might say, “ChatGPT doesn’t understand my prompt” when the chatbot’s NLP (natural language processing) algorithm fails to return the wanted outcome.
Familiar concepts such as “understanding” help us better conceptualize how complex AI systems work. However, they can also lead to distorted notions about AI’s capabilities. If we assign human-like concepts to AI systems, it’s natural for our human minds to infer that they also possess human values and motivations.
But this inference is fundamentally untrue. Artificial intelligence is not human and therefore cannot intrinsically care about reason, loyalty, safety, environmental issues and the greater good. The primary goal of an artificial “mind” is to complete the task for which it was programmed.
Therefore, it is up to AI developers to build in human values and goals. Otherwise, in pursuit of task completion, AI systems can become misaligned from programmers’ goals and cause harm, sometimes catastrophically. This consideration is important as automation becomes more prevalent in high-stakes use cases in healthcare, human resources, finance, military scenarios and transportation.
For example, self-driving cars might be programmed with the primary goal of getting from point A to point B as fast as possible. If these autonomous vehicles ignore safety guardrails to complete that goal, they might severely injure or kill pedestrians and other drivers.
University of California, Berkeley researchers Simon Zhuang and Dylan Hadfield-Menell liken AI alignment to the Greek myth of King Midas. In summary, King Midas is granted a wish and requests that everything he touches turns into gold. He eventually dies because the food he touches also becomes gold, rendering it inedible.
King Midas met an untimely end because his wish (unlimited gold) did not reflect what he truly wanted (wealth and power). The researchers explain that AI designers often find themselves in a similar position, and that “the misalignment between what we can specify and what we want has already caused significant harms.” 2
Some risks of AI misalignment include:
AI bias results from human biases present in an AI system’s original training datasets or algorithms. Without alignment, these AI systems are unable to avoid biased outcomes that are unfair, discriminatory or prejudiced. Instead, they perpetuate the human biases in their input data and algorithms.
For example, an AI hiring tool trained on data from a homogeneous, male workforce might favor male candidates while disadvantaging qualified female applicants. This model is not aligned with the human value of gender equality and might lead to hiring discrimination.
In reinforcement learning, AI systems learn from rewards and punishments to take actions within an environment that meet a specified goal. Reward hacking occurs when the AI system finds a loophole to trigger the reward function without actually meeting the developers’ intended goal.
For instance, OpenAI trained one of its AI agents on a boat racing game called CoastRunners. The human intent of the game is to win the boat race. However, players can also earn points by driving through targets within the racecourse. The AI agent found a way to isolate itself in a lagoon and continually hit targets for points. While the AI agent did not win the race (the human goal), it “won” the game with its own emergent goal of obtaining the highest score.3
Misaligned AI systems can contribute to misinformation and political polarization. For example, social media content recommendation engines are trained for user engagement optimization. Therefore, they highly rank posts, videos and articles that receive the highest engagement, such as attention-grabbing political misinformation. This outcome is not aligned with the best interests or well-being of social media users, or values such as truthfulness and time well spent.4
As far-fetched as it might sound, artificial superintelligence (ASI) without proper alignment to human values and goals might have the potential to threaten all life on earth. A commonly cited example of this existential risk is philosopher Nick Bostrom’s paperclip maximizer scenario. In this thought experiment, an ASI model is programmed with the top incentive to manufacture paperclips. To achieve this goal, the model eventually transforms all of earth and then increasing portions of space into paperclip manufacturing facilities.5
This scenario is hypothetical, and the existential risk from AI first requires artificial general intelligence (AGI) to become a reality. However, it helps emphasize the need for alignment to keep pace with the field of AI as it evolves.
There are two major challenges to achieving aligned AI: the subjectivity of human ethics and morality and the “alignment problem.”
There is no universal moral code. Human values change and evolve, and can also vary across companies, cultures and continents. People might hold different values than their own family members. So, when aligning AI systems that can affect the lives of millions of people, who makes the judgment call? Which goals and values take precedence?
American author Brian Christian frames the challenge differently in his book “The Alignment Problem: Machine Learning and Human Values.” He posits: what if the algorithm misunderstands our values? What if it learns human values from being trained on past examples that reflect what we have done but not who we want to be?6
Another challenge is the sheer number of human values and considerations. University of California, Berkeley researchers describe it this way: “there are many attributes of the world about which the human cares, and, due to engineering and cognitive constraints it is intractable to enumerate this complete set to the robot.”7
The most infamous challenge is the alignment problem. AI models are already often considered black boxes that are impossible to interpret. The alignment problem is the idea that as AI systems become even more complex and powerful, anticipating and aligning their outcomes to human goals becomes increasingly difficult. Discussions around the alignment problem often focus on the risks posed by the anticipated development of artificial superintelligence (ASI).
There is concern that the future of AI includes systems with unpredictable and uncontrollable behavior. These systems’ ability to learn and adapt rapidly might make predicting their actions and preventing harm difficult. This concern has inspired a branch of AI alignment called superalignment.
AI safety research organizations are already at work to address the alignment problem. For example, the Alignment Research Center is a nonprofit AI research organization that “seeks to align future machine learning systems with human interests by furthering theoretical research.” The organization was founded by Paul Christiano, who formerly led the language model alignment team at OpenAI and currently heads AI Safety at the US AI Safety Institute.
And Google DeepMind—a team of scientists, engineers, ethicists and other experts—is working to build the next generation of AI systems safely and responsibly. The team introduced the Frontier Safety Framework in May 2024. The framework is “a set of protocols that aims to address severe risks that may arise from powerful capabilities of future foundation models.”8
There are several methodologies that can help align AI systems to human values and goals. These methodologies include alignment through reinforcement learning from human feedback (RLHF), synthetic data, red teaming, AI governance and corporate AI ethics boards.
Through reinforcement learning developers can teach AI models “how to behave” with examples of “good behavior.”
AI alignment happens during model fine-tuning and typically has two steps. The first step might be an instruction-tuning phase, which improves model performance on specific tasks and on following instructions in general. The second phase might use reinforcement learning from human feedback (RLHF). RLHF is a machine learning technique in which a “reward model” is trained with direct human feedback, then used to optimize the performance of an artificial intelligence agent through reinforcement learning. It aims to improve a model’s integration of abstract qualities such as helpfulness and honesty.
OpenAI used RLHF as its main method to align its GPT-3 and GPT-4 series of models. However, the American AI research organization doesn’t expect RLHF to be a sufficient method for aligning future artificial general intelligence (AGI) models likely due to RLHF’s significant limitations.9 For example, its dependence on high-quality human annotations makes it difficult to apply and scale the technique for unique or intricate tasks. It is challenging to find “consistent response demonstrations and in-distribution response preferences.”10
Synthetic data is data that has been created artificially through computer simulation or generated by algorithms. It takes the place of real-world data when real-world data is not readily available and can be tailored to specific tasks and values. Synthetic data can be used in various alignment efforts.
For example, contrastive fine-tuning (CFT) shows AI models what not to do. In CFT, a second “negative persona” model is trained to generate “bad,” misaligned responses. Both these misaligned and aligned responses are fed back to the original model. IBM® researchers found that on benchmarks for helpfulness and harmlessness, large language models (LLMs) trained on contrasting examples outperform models tuned entirely on good examples. CFT allows developers to align models before even collecting human preference data—curated data that meets the defined benchmarks for alignment—which is expensive and takes time.
Another synthetic data alignment method is called SALMON (Self-ALignMent with principle fOllowiNg reward models). In this approach from IBM Research®, synthetic data allows an LLM to align itself. First, an LLM generates responses to a set of queries. These responses are then fed to a reward model that has been trained on synthetic preference data aligned with human-defined principles. The reward model scores the responses from the original LLM against these principles. The scored responses are then fed back to the original LLM.
With this method, developers have almost complete control over the reward model’s preferences. This allows organizations to shift principles according to their needs and eliminates the reliance on collecting large amounts of human preference data.11
Red teaming can be considered an extension of the alignment that occurs during model fine-tuning. It involves designing prompts to circumvent the safety controls of the model that is being fine-tuned. After vulnerabilities surface, the target models can be realigned. While humans can still engineer these “jailbreak prompts,” “red team” LLMs can produce a wider variety of prompts in limitless quantities. IBM Research describes red team LLMs as “toxic trolls trained to bring out the worst in other LLMs.”
AI governance refers to the processes, standards and guardrails that help ensure AI systems and tools are safe and ethical. In addition to other governance mechanisms, it aims to establish the oversight necessary to align AI behaviors with ethical standards and societal expectations. Through governance practices such as automated monitoring, audit trails and performance alerts, organizations can help ensure their AI tools—like AI assistants and virtual agents—are aligned with their values and goals.
Organizations might establish ethics boards or committees to oversee AI initiatives. For example, IBM’s AI Ethics Council reviews new AI products and services and helps ensure that they align with IBM's AI principles. These boards often include cross-functional teams with legal, computer science and policy backgrounds.
Learn about the new challenges of generative AI, the need for governing AI and ML models and steps to build a trusted, transparent and explainable AI framework.
Read about driving ethical and compliant practices with a portfolio of AI products for generative AI models.
Gain a deeper understanding of how to ensure fairness, manage drift, maintain quality and enhance explainability with watsonx.governance™.
We surveyed 2,000 organizations about their AI initiatives to discover what’s working, what’s not and how you can get ahead.
Govern generative AI models from anywhere and deploy on cloud or on premises with IBM watsonx.governance.
See how AI governance can help increase your employees’ confidence in AI, accelerate adoption and innovation, and improve customer trust.
Prepare for the EU AI Act and establish a responsible AI governance approach with the help of IBM Consulting®.
1 “AI Alignment: A Comprehensive Survey," arXiv, 1 May 2024.
2, 7 "Consequences of Misaligned AI," NeurIPS Proceedings, 2020.
3 "Faulty Reward Functions in the Wild," OpenAI, 21 December 2016.
4 “Modelling the Recommender Alignment Problem,” arXiv, 25 Aug 2022.
5 “Ethical Issues in Advanced Artificial Intelligence,” Nick Bostrom, n.d.
6 “‘The Alignment Problem’ Review: When Machines Miss the Point,” The Wall Street Journal, 25 October 2020.
8 “Introducing the Frontier Safety Framework,” Google DeepMind, 17 May 2024.
9 “Our Approach to Alignment Research,” OpenAI, 24 August 2022.
10, 11 “SALMON: Self-Alignment with Instructable Reward Models,” arXiv, 9 Apr 2024.
IBM web domains
ibm.com, ibm.org, ibm-zcouncil.com, insights-on-business.com, jazz.net, mobilebusinessinsights.com, promontory.com, proveit.com, ptech.org, s81c.com, securityintelligence.com, skillsbuild.org, softlayer.com, storagecommunity.org, think-exchange.com, thoughtsoncloud.com, alphaevents.webcasts.com, ibm-cloud.github.io, ibmbigdatahub.com, bluemix.net, mybluemix.net, ibm.net, ibmcloud.com, galasa.dev, blueworkslive.com, swiss-quantum.ch, blueworkslive.com, cloudant.com, ibm.ie, ibm.fr, ibm.com.br, ibm.co, ibm.ca, community.watsonanalytics.com, datapower.com, skills.yourlearning.ibm.com, bluewolf.com, carbondesignsystem.com