October 22, 2020 By Priya Krishnan 4 min read

It’s all about trust

Two things are true about trust: it’s at the core of all successful human relationships, and it’s not an easy thing to attain. Trust contains multitudes of nuance, and when achieved, it can lead to transformational events. The same could be said of artificial intelligence.

We have seen that a business that can trust their AI will do more and go beyond their expectations and projections. Every business uses AI differently, and that trust looks different to each industry and every use case. So what does trust mean when it comes to a technology like AI? IBM Research has broken its taxonomy of AI trust into three dynamics:

  • Ethics
  • Governance
  • Trustworthiness

To take those ideas further, trust in AI means understanding:

  • Where the data is coming from
  • How that data is being used
  • What data the training model contains
  • How all of this affects the entire lifecycle of the AI

For IBM, trust is a foundational pillar of AI. Whether you’re looking at data collected by AI or seeing how AI performs within your industry use-case guidelines, you will need those insights delivered in a trusted manner. As such, we’ve developed a multifaceted perspective around this complex topic, which helped us devise tools and capabilities for enterprise use to help businesses remain confident with their AI.

The Pillars of Trust

Our engineers at IBM Research started with the question, “What would it take to trust the output of an AI model?” The properties they came up with centered around accuracy, fairness, understandability, dependability and transparency in AI. They further developed those key takeaways into the AI Pillars of Trust:

  • Performance
  • Fairness
  • Explainability
  • Robustness
  • Transparency

Over the last several years, IBM Research has been building AI algorithms that will imbue AI with these properties of trust. They then created toolkits that embody those algorithms, and now we’ve taken those innovations and added them to Watson OpenScale capabilities inside IBM Cloud Pak for Data.

AI Governance, which is part of the overall taxonomy, is how a business operationalizes and vets AI results — so they’re getting only what’s intended. It’s also the ability to prove trustworthiness. In regulated industries, this implies audit readiness.

However, vetting results requires documenting the model’s inputs and behavior, which is manual and tedious work. It’s also not easy to share metadata about models across multiple enterprise tools and platforms, and current practices and tools are not optimized for AI.

IBM Cloud Pak for Data combines the best of IBM Research and engineering to enable a fully governed AI Lifecycle, making it easier to know your model, trust your model, and use your model.

AI Fairness

Fairness is fundamental to who we are and where we want to be as a society. As such, bias in AI has drawn much attention in the last couple of years. In our quest for unbiased AI, IBM Research has authored a pioneering algorithm for bias detection.

Imagine a credit lender who needs to approve a loan. When the lender checks the client against their risk model, a lot of information gets shipped into the modeler to help it make its recommendation to approve the loan or not. This information comes from many sources, including the lender’s data and often third-party data. In most cases, the lender cannot know if the data is free of bias.

Products like Watson OpenScale in Cloud Pak for Data provide tools that can mitigate bias and detect drift and performance invalidation, so operations personnel or data scientists can fix instances of biased outcomes by model. The idea is to give users the ability to take biased data and easily shape it into a fairer version of itself while still allowing the model to learn what it needs to learn.

AI Explainability

Explainability in AI is multifaceted. One approach does not fit all cases, because different processes require different explanations.

For example, a loan officer asks why you recommended rejection of a loan; the customer wants to know why their loan was denied; the regulator wants proof that your system isn’t discriminatory. There is no single answer that will satisfy all of these questions.

Enterprise-grade decisions must be consumable, so this concept has been integrated into Watson OpenScale to make explainability more transparent for business use cases. We’ve introduced two types of explanations to truly open up your AI black box. The first shows visually why a prediction was made by the model, showing the features or the inputs that are most important to the outcome, and how they’re stacked up. These visualizations can be generated on the fly via the OpenScale dashboard. The second type of explanation allows the user to change the inputs of the model to test the boundaries of the model’s decision-making.

This is all just a small taste of the advanced features IBM Research is working on in regard to AI governance. There’s much more to explore, and as these advances make their way into the product, we’ll be back to tell you about them.

Trusted AI is not only a strategic imperative but also an ethical one. As a result of the work we’re doing around trust and AI, clients can understand and explain how their AI models are making decisions, and why they’re making them.

Interested in seeing these capabilities in action? Check out the full Innovation panel to watch demos and hear how this transformative technology has helped clients like KPMG, IBM HR and the US Open develop innovative and trustworthy experiences for users.

Watch the Innovation panel

Was this article helpful?

More from Artificial intelligence

Getting ready for artificial general intelligence with examples

12 min read - Imagine a world where machines aren't confined to pre-programmed tasks but operate with human-like autonomy and competence. A world where computer minds pilot self-driving cars, delve into complex scientific research, provide personalized customer service and even explore the unknown. This is the potential of artificial general intelligence (AGI), a hypothetical technology that may be poised to revolutionize nearly every aspect of human life and work. While AGI remains theoretical, organizations can take proactive steps to prepare for its arrival by…

Breaking Boundaries: PostgreSQL 16 is now available on IBM Cloud

2 min read - PostgreSQL Version 16 is now available on IBM Cloud®. The latest version of IBM Cloud® Databases for PostgreSQL includes critical features that offer unmatched reliability and scalability for clients' data storage needs. Advanced features like enhanced parallel query performance and accelerated indexing provide significant performance gains to your applications. With IBM Cloud's robust ecosystem of extensions and plugins, PostgreSQL v16 empowers customers to build and manage sophisticated, high-performance applications easily while continuing to improve data management at scale. In addition, customers can continue to access powerful IBM…

IBM and TechD partner to securely share data and power insights with gen AI

3 min read - As technology expands, at TechD, we know that the quality of generative AI (gen AI) depends on accurate data sourcing. A reliable and trustworthy data source is essential for sharing information across departments. Through the implementation of generative AI we are able to expand our knowledge to many individuals easily, quickly and efficiently becoming a resource. In today's rapidly evolving digital world, immediate responses are crucial for delivering outstanding user experiences. Our partnership with IBM facilitates the delivery of scalable…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters