Data science can quickly turn data into insights and those insights can lead to decisions. And sometimes, the results are unwittingly spoiled by bias and drift, causing mistrust. This problem undoubtedly hampers AI adoption and can negatively impact people’s lives and a company’s reputation.  

Take hiring decisions. 

Tools or recruiting systems that screen candidates have long demanded attention; as research has demonstrated, they can reflect historical discrimination based on the datasets.  

Sensitive features such as gender, ethnicity, and age, even if not included in AI, could have influenced the training of the data, the source of the data, and even how the data got to the AI from a training dataset. In other words, even if there’s no intent or access to those features in the beginning, those perceptions can lead to incorrect decisions. 

A growing concern about AI’s trustworthiness has provoked worldwide conversation among data leaders and business leaders alike about how to improve the practices of trustworthy AI and govern it across the AI lifecycle.  

How do we understand what AI models are doing?  

How do we ensure AI accuracy and fairness?  

How do we speed up production and adoption of AI models?  

Can we trust the output? 

According to IBM, if a business is involved in making decisions on automation that’s driven by AI, it needs to be transparent. The business must know it’s making decisions that align with company policy — and that people who are making the decisions based on AI can trust it. 

One major U.S. company was eager to tackle the problem on a large scale and turned to IBM for help. Within this corporation’s mandate to focus on social responsibility has been an effort to drive more workforce diversity and inclusion. When it came to its hiring practices, it was critical that this employer ensure fairness and trust was in place within its AI and ML models – especially when it came to attracting and recruiting talent. 

With over 1,000 data scientists in its ranks, this industry leader has traveled far on its AI journey. Hundreds of ML models were in production, but what it lacked was an enterprise solution that assured that models could be trusted in a socially responsible manner.  

Data science leaders wanted to be able to translate the models’ decisions and results easily — in a way any hiring manager could understand. It wanted to establish fairness by accelerating the identification of any bias in hiring and “explain” decisions made by AI models. The company also knew it needed to operationalize AI governance to get more of its business users on board – so it set out to find a solution that could achieve all of these things.  

The answer was IBM Watson® Studio(TM), a AI monitoring and management tool within IBM Cloud Pak® for Data that filled a much needed gap. Once IBM’s Data Science and AI Elite team showed how the product could consistently manage AI models for accuracy and fairness, IBM’s Expert Lab services came in to drive the ongoing teamwork needed to reach the corporation’s goals. 

Over 90% of organizations say their ability to explain how their AI made a decision is critical.

So what’s the next step to put trustworthy AI into practice?

Since partnering closely with IBM, the company has been tapping IBM’s Expert Lab services to implement IBM Watson Studio on Cloud Pak for Data in several use cases, relying on IBM’s expertise for this area of the AI lifecycle. The partnership has resulted in the creation of a enterprise framework that can operate within the scale of the enormous organization. Today the customer has all the capabilities it needs to manage aspects of bias, fairness, accuracy, drift, explainability and transparency in its use of AI and machine learning. 

Now, the company is proactively monitoring for and mitigating bias in its hiring processes. Because automation has reduced the workload within DevOps, the company’s data scientists can focus more on the new model development and refinements.  

Today, companies across all industries have a clear opportunity to harness data and AI to build effective and scalable solutions while eradicating systemic racism and structural inequality. And there’s no denying the fact that there’s a relationship between higher growth and the ability to scale AI with repeatable, trustworthy processes. According to a January 2020 Forrester Consulting study commissioned by IBM, Overcome Obstacles to get to AI at scale, the companies who are the fastest growing in their industries are over 6x times more likely to have scaled AI.  

There’s no better time to address the societal relevance of AI and the need for a trustworthy AI framework based on ethics, has governed data and AI technology, and is rooted in a diverse and open ecosystem. 

Accelerate your AI journeywith a prescriptive approach. 

Was this article helpful?
YesNo

More from Artificial intelligence

AI transforms the IT support experience

5 min read - We know that understanding clients’ technical issues is paramount for delivering effective support service. Enterprises demand prompt and accurate solutions to their technical issues, requiring support teams to possess deep technical knowledge and communicate action plans clearly. Product-embedded or online support tools, such as virtual assistants, can drive more informed and efficient support interactions with client self-service. About 85% of execs say generative AI will be interacting directly with customers in the next two years. Those who implement self-service search…

Bigger isn’t always better: How hybrid AI pattern enables smaller language models

5 min read - As large language models (LLMs) have entered the common vernacular, people have discovered how to use apps that access them. Modern AI tools can generate, create, summarize, translate, classify and even converse. Tools in the generative AI domain allow us to generate responses to prompts after learning from existing artifacts. One area that has not seen much innovation is at the far edge and on constrained devices. We see some versions of AI apps running locally on mobile devices with…

Chat with watsonx models

3 min read - IBM is excited to offer a 30-day demo, in which you can chat with a solo model to experience working with generative AI in the IBM® watsonx.ai™ studio.   In the watsonx.ai demo, you can access some of our most popular AI models, ask them questions and see how they respond. This gives users a taste of some of the capabilities of large language models (LLMs). AI developers may also use this interface as an introduction to building more advanced…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters