Read how explainable AI benefits production AI

What is explainable AI?

Explainable artificial intelligence (XAI) is a set of processes and methods that allows human users to comprehend and trust the results and output created by machine learning algorithms. Explainable AI is used to describe an AI model, its expected impact and potential biases. It helps characterize model accuracy, fairness, transparency and outcomes in AI-powered decision making. Explainable AI is crucial for an organization in building trust and confidence when putting AI models into production. AI explainability also helps an organization adopt a responsible approach to AI development.

As AI becomes more advanced, humans are challenged to comprehend and retrace how the algorithm came to a result. The whole calculation process is turned into what is commonly referred to as a “black box" that is impossible to interpret. These black box models are created directly from the data. And, not even the engineers or data scientists who create the algorithm can understand or explain what exactly is happening inside them or how the AI algorithm arrived at a specific result.

There are many advantages to understanding how an AI-enabled system has led to a specific output.  Explainability can help developers ensure that the system is working as expected, it might be necessary to meet regulatory standards, or it might be important in allowing those affected by a decision to challenge or change that outcome.¹

Professional working on laptop

Why does explainable AI matter?

It is crucial for an organization to have a full understanding of the AI decision-making processes with model monitoring and accountability of AI and not to trust them blindly. Explainable AI can help humans understand and explain machine learning (ML) algorithms, deep learning and neural networks.

ML models are often thought of as black boxes that are impossible to interpret.² Neural networks used in deep learning are some of the hardest for a human to understand. Bias, often based on race, gender, age or location, has been a long-standing risk in training AI models. Further, AI model performance can drift or degrade because production data differs from training data. This makes it crucial for a business to continuously monitor and manage models to promote AI explainability while measuring the business impact of using such algorithms. Explainable AI also helps promote end user trust, model auditability and productive use of AI. It also mitigates compliance, legal, security and reputational risks of production AI.

Explainable AI is one of the key requirements for implementing responsible AI, a methodology for the large-scale implementation of AI methods in real organizations with fairness, model explainability and accountability.³ To help adopt AI responsibly, organizations need to embed ethical principles into AI applications and processes by building AI systems based on trust and transparency.

Person looking at people in video meeting on laptop computer

Continuous model evaluation

With explainable AI, a business can troubleshoot and improve model performance while helping stakeholders understand the behaviors of AI models. Investigating model behaviors through tracking model insights on deployment status, fairness, quality and drift is essential to scaling AI. Continuous model evaluation empowers a business to compare model predictions, quantify model risk and optimize model performance. Displaying positive and negative values in model behaviors with data used to generate explanation speeds model evaluations. A data and AI platform can generate feature attributions for model predictions and empower teams to visually investigate model behavior with interactive charts and exportable documents.

Professional working on laptop

Value of explainable AI

From a Forrester study covering explainable AI and model monitoring on IBM Cloud Pak for Data

Explainable AI benefits

Operationalize AI with trust and confidence

Build trust in production AI. Rapidly bring your AI models to production. Ensure interpretability and explainability of AI models. Simplify the process of model evaluation while increasing model transparency and traceability.

Speed time to AI results

Systematically monitor and manage models to optimize business outcomes. Continually evaluate and improve model performance. Fine-tune model development efforts based on continuous evaluation.

Mitigate risk and cost of model governance

Keep your AI models explainable and transparent. Manage regulatory, compliance, risk and other requirements. Minimize overhead of manual inspection and costly errors. Mitigate risk of unintended bias.

The IBM approach to explainable AI

For more than 100 years, IBM has continuously strived for innovation capable of bringing benefits to everyone and not just a few. This philosophy is also applied to AI: we aim to create and offer reliable technology that can augment, not replace, human decision-making.

While AI holds the promise of delivering valuable insights and patterns across a multitude of applications, broad adoption of AI systems will rely heavily on the ability of people to trust the AI output. Human trust in technology is based on our understanding of how it works and our assessment of its safety and reliability. This makes explainable AI crucial.  IBM’s approach to explainable AI is to make AI reliable and fair, to make it able to be accounted for, and to help ensure that it will cause no harm.

At the heart of our innovation, IBM Research is developing diverse approaches for how to achieve fairness, robustness, explainability, accountability and value alignment, and how to integrate these throughout the entire lifecycle of an AI application. Explainable AI frameworks and tool sets by IBM Research are integrated into the IBM Cloud Pak for Data platform so that businesses can take advantage of our latest AI technology in a governed, secure and scalable fashion.

Person looking at mobile phone while standing in front of an office window

Five considerations for explainable AI

Five considerations for explainable AI

To drive desirable outcomes with explainable AI, consider the following.

Fairness and debiasing: Manage and monitor fairness. Scan your deployment for potential biases. 

Model drift mitigation: Analyze your model and make recommendations based on the most logical outcome. Alert when models deviate from the intended outcomes.

Model risk management: Quantify and mitigate model risk. Get alerted when a model performs inadequately. Understand what happened when deviations persist.

Lifecycle automation: Build, run and manage models as part of integrated data and AI services. Unify the tools and processes on a platform to monitor models and share outcomes. Explain the dependencies of machine learning models.

Multicloud-ready: Deploy AI projects across hybrid clouds including public clouds, private clouds and on premises. Promote trust and confidence with explainable AI.

Closeup of modern glass building

Improve AI explainability with IBM Cloud Pak for Data

The IBM Cloud Pak® for Data platform provides data and AI services in a unified environment so that a business can assess impact and relationships of data and models to improve AI explainability. It also helps a business gain model insights on deployments, fairness, quality and risk. The solution helps explain AI transactions, categorical models, image models and unstructured text models with tools such as contrastive explanations and Local Interpretable Model-Agnostic Explanations (LIME). Making AI explainable and transparent by automating the AI lifecycle on a modern information architecture is vital to production AI success.

Go deeper on explainable AI

How does explainable AI work?

With explainable AI – as well as interpretable machine learning – organizations can gain access to AI technology’s underlying decision-making and are empowered to make adjustments. Explainable AI can improve the user experience of a product or service by helping the end user trust that the AI is making good decisions. When do AI systems give enough confidence in the decision that you can trust it, and how can the AI system correct errors that arise?⁴

As AI becomes more advanced, ML processes still need to be understood and controlled to ensure AI model results are accurate. Let’s look at the difference between AI and XAI, the methods and techniques used to turn AI to XAI, and the difference between interpreting and explaining AI processes.

Comparison between AI and XAI

What exactly is the difference between “regular” AI and explainable AI? XAI implements specific techniques and methods to ensure that each decision made during the ML process can be traced and explained. AI, on the other hand, often arrives at a result using an ML algorithm, but the architects of the AI systems do not fully understand how the algorithm reached that result. This makes it hard to check for accuracy and leads to loss of control, accountability and auditability.

Explainable AI techniques

The setup of XAI techniques consists of three main methods. Prediction accuracy and traceability address technology requirements while decision understanding addresses human needs. Explainable AI — especially explainable machine learning — will be essential if future warfighters are to understand, appropriately trust, and effectively manage an emerging generation of artificially intelligent machine partners.⁵

Prediction accuracy
Accuracy is a key component of how successful the use of AI is in everyday operation. By running simulations and comparing XAI output to the results in the training data set, the prediction accuracy can be determined. The most popular technique used for this is Local Interpretable Model-Agnostic Explanations (LIME), which explains the prediction of classifiers by the ML algorithm.

Traceability is another key technique for accomplishing XAI. This is achieved, for example, by limiting the way decisions can be made and setting up a narrower scope for ML rules and features. An example of a traceability XAI technique is DeepLIFT (Deep Learning Important FeaTures), which compares the activation of each neuron to its reference neuron and shows a traceable link between each activated neuron and even shows dependencies between them.

Decision understanding
This is the human factor. Many people have a distrust in AI, yet to work with it efficiently, they need to learn to trust it. This is accomplished by educating the team working with the AI so they can understand how and why the AI makes decisions.

Explainability versus interpretability in AI

Interpretability is the degree to which an observer can understand the cause of a decision. It is the success rate that humans can predict for the result of an AI output, while explainability goes a step further and looks at how the AI arrived at the result.

How does explainable AI relate to responsible AI?

Explainable AI and responsible AI have similar objectives, yet different approaches. Here are the main differences between explainable and responsible AI:

  • Explainable AI looks at AI results after the results are computed.
  • Responsible AI looks at AI during the planning stages to make the AI algorithm responsible before the results are computed.
  • Explainable and responsible AI can work together to make better AI.

To learn more about explainable AI, sign up for an IBMid and start your IBM Cloud Pak for Data trial today.

How to implement explainable AI

Use these resources to learn more about how to implement explainable AI.

Online seminar: How to manage and monitor models
Learn what you can do when your models do not work.
Watch the webinar (link resides outside IBM) →

Learning path: Manage AI with trust
Learn how to track and measure outcomes from AI across its lifecycle, while adapting and governing AI to changing business conditions.
See the tutorial →

Hands-on lab: Monitor machine learning models
Explore the step-by-step processes for evaluating models for fairness, accuracy and explainability. 
See the lab →

Explainable AI use cases


Closeup photo of 3 surgeons operating

Accelerate diagnostics, image analysis, resource optimization and medical diagnosis. Improve transparency and traceability in decision-making for patient care. Streamline the pharmaceutical approval process with explainable AI.

Financial Services

Person holding a credit card while working on a laptop

Improve customer experiences with a transparent loan and credit approval process. Speed credit risk, wealth management and financial crime risk assessments. Accelerate resolution of potential complaints and issues. Increase confidence in pricing, product recommendations and investment services.

Criminal justice

Police person in front of an area marked off with crime scene tape

Optimize processes for prediction and risk assessment. Accelerate resolutions using explainable AI on DNA analysis, prison population analysis and crime forecasting. Detect potential biases in training data and algorithms.


¹ ”Explainable AI,” The Royal Society, 28 November 2019. (link resides outside IBM)

² ”Explainable Artificial Intelligence,” Jaime Zornoza, 15 April 2020. (link resides outside IBM)

³ ”Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI,” ScienceDirect, June 2020. (link resides outside IBM)

⁴ ”Understanding Explainable AI,” Ron Schmelzer, Forbes contributor, 23 July 2019. (link resides outside IBM)

⁵ ”Explainable Artificial Intelligence (XAI),” Dr. Matt Turek, The U.S. Defense Advanced Research Projects Agency (DARPA). (link resides outside IBM)