On June 12th, IBM debuted AutoAI, a new set of capabilities for Watson Studio designed to automate critical yet time-consuming tasks associated with designing and optimizing AI in the enterprise. As a result, data scientists can be liberated to execute more data science and AI projects in their organizations. Read more about AutoAI in the announcement.

To learn more about what these developments mean for the data science community, I sat down with Alexander Gray, vice president of AI, IBM Research, to get his perspective. Alexander has more than 25 years of experience researching machine learning and AI algorithms, theoretical frameworks and designing solutions for difficult use cases across many industries.

Alexander, what is your view on the business appetite for transforming businesses through automation of AI, aka, AutoAI? And what use cases are most popular with automating AI today?

We see that the appetite is very high. Organizations are motivated by the potential of automation to dramatically reduce time to market and to increase the number of AI projects that they can take on with their available staffing. One thing that many are not as aware of, but will certainly discover later, is that automation designs can significantly increase the quality of solutions as well.

Tell me the real scoop. What routine tasks are people using automated AI for today? And what is still science fiction?

The current state of automation technology still encounters challenges around use cases that rely heavily on domain knowledge. I would say most data scientists we encounter who are using automation to their advantage today are mainly doing hyper-parameter optimization (HPO). There are many existing technologies that focus on this area. It should be noted that while it is a good place to start, such tools only address a small fraction of the data science process. The part that I would say is still science fiction – meaning too far out to put on a roadmap – is the ability to specify the actual business goals and constraints that would require AGI (artificial general intelligence), and we’re not there yet.

For businesses starting out in new data science projects, what part of data science is ideal for automating?

For those new to automating data science, the most straightforward place to start is at the end of the data science pipeline, the modeling stage. It is easy and straightforward to automate HPO because you can see immediate gains in your data science projects. Then one can move to automating the choice of machine learning model. We are focused on going beyond this stage to also address data preparation, because it is of high interest to data scientists, being where they spend most of their energy typically. We regard this as one of the most important research frontiers.

Is AutoAI going to take jobs away from data scientists?

It is commonly recognized that we have a growing demand for data scientists worldwide. This is not accounting for the potential number of AI projects within the realm of possibility but not even considered at this moment. In my experience, most data scientists have a giant backlog of both exploratory and critical AI projects that they and their organizations would love to get to, but they don’t have the bandwidth. Also, when we say “automation” here, we typically mean assistance for humans along some spectrum from light to heavy.

In terms of “taking jobs away,” automation of AI is actually a mechanization of tedious activities, a time-saving benefit that data scientists embrace because they seemingly enjoy thinking more than tedium. We are simply making data scientists’ tools smarter and more powerful. Using more powerful tools is simply a new skill, compared to using less powerful tools. As data science skills shift to use these new kinds of tools, job roles will have increased business responsibility and impact because they are able to create more value.

With AutoAI, will businesses get more from their AI and data science investments? What are the potential drawbacks and misconceptions?

We believe many small vendors and open source projects will appear in the face of automation. While automation offers the potential to do things better and faster, it also has the potential to propagate human errors if there is poor science underneath them. It is much easier for this to occur than most people think. In my experience, even teams of PhDs from top schools can commonly make errors in statistical nuances which leads to poorer models than would otherwise be possible. For this reason, a high degree of mathematical expertise behind the automation is critical in order to be able to rely on the decisions made by automated AI. And the need for data scientists to have strong understanding of the underlying principles will not go away, because human oversight will always be needed for the most important applications.

Do you have a personal prediction on how automation of AI impacts our society? Is AI “scary?”

Data science automation will actually create new opportunities for many new people to enter data science who previously had limited ways to participate. I believe it will enable the creation of entirely new job categories, allowing much wider participation in the AI revolution. We believe this is exciting rather than scary.

What interesting AutoAI research are you currently working on? Any potential breakthroughs you can share?

We are working toward treating the problems of data science automation in a much more fundamental way than I have previously observed. It begins with formalizing the problem of data science in mathematical terms, which we cannot find in existing text books. We predict that placing all of the “grungy” aspects of data science on solid mathematical foundations will have deep benefits in both error prevention and quality of solutions. It’s an exciting time in AI, so stay tuned.

Follow Alexander on LinkedIn. Watch Alexander talk about the future of AI, and explore what AutoAI can do for your business.

More from Analytics

How data stores and governance impact your AI initiatives

6 min read - Organizations with a firm grasp on how, where, and when to use artificial intelligence (AI) can take advantage of any number of AI-based capabilities such as: Content generation Task automation Code creation Large-scale classification Summarization of dense and/or complex documents Information extraction IT security optimization Be it healthcare, hospitality, finance, or manufacturing, the beneficial use cases of AI are virtually limitless in every industry. But the implementation of AI is only one piece of the puzzle. The tasks behind efficient,…

IBM and ESPN use AI models built with watsonx to transform fantasy football data into insight

4 min read - If you play fantasy football, you are no stranger to data-driven decision-making. Every week during football season, an estimated 60 million Americans pore over player statistics, point projections and trade proposals, looking for those elusive insights to guide their roster decisions and lead them to victory. But numbers only tell half the story. For the past seven years, ESPN has worked closely with IBM to help tell the whole tale. And this year, ESPN Fantasy Football is using AI models…

Data science vs data analytics: Unpacking the differences

5 min read - Though you may encounter the terms “data science” and “data analytics” being used interchangeably in conversations or online, they refer to two distinctly different concepts. Data science is an area of expertise that combines many disciplines such as mathematics, computer science, software engineering and statistics. It focuses on data collection and management of large-scale structured and unstructured data for various academic and business applications. Meanwhile, data analytics is the act of examining datasets to extract value and find answers to…

Financial planning & budgeting: Navigating the Budgeting Paradox

5 min read - Budgeting, an essential pillar of financial planning for organizations, often presents a unique dilemma known as the “Budgeting Paradox.” Ideally, a budget should give the most accurate and timely idea of anticipated revenues and expenses. However, the traditional budgeting process, in its pursuit of precision and consensus, can take several months. By the time the budget is finalized and approved, it might already be outdated.In today's rapid pace of change and unpredictability, the conventional budgeting process is coming under scrutiny.It's…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters