Automated machine learning (AutoML) is the practice of automating the end-to-end development of machine learning models (ML models). AutoML enables non-experts to create and implement artificial intelligence (AI) systems while streamlining AI workflows for data scientists and developers.
AutoML tools simplify the process of building ML models. Users benefit from an intuitive interface through which they can create, train, validate and deploy generative AI models and other deep learning systems. AutoML facilitates AI implementation in regulated industries with its explainable and reproducible results.
Without AutoML, every step in the machine learning (ML) workflow—data preparation, data preprocessing, feature engineering and hyperparameter optimization—must be manually carried out. AutoML democratizes machine learning by making it accessible to anyone who is interested in exploring its potential. Meanwhile, experienced MLOps teams and data science professionals can automate the routine aspects of machine learning workflows while focusing on more demanding learning tasks.
AutoML solutions work by constructing numerous machine learning pipelines to handle the intended task, then identifying the optimal choice. Model evaluation and model selection are automated as part of the iterative process of choosing the best model for the job. Data visualization tools bring even more ease-of-use to the AutoML process.
The difference between AutoML and traditional machine learning is that AutoML automates nearly every stage of the machine learning pipeline. Traditional pipelines are time-consuming, resource-intensive and prone to human error. By comparison, advancements in AutoML have led to greater efficiency and better results.
A typical machine learning pipeline consists of the following steps:
Data preparation is the process of collecting raw data and integrating it into a training dataset. Data preparation helps ensure that training data is free from bias and is what sets a model up for success: accurate data leads to accurate predictions and insights. As enterprises link AI systems with proprietary data stores, such as through retrieval-augmented generation (RAG), data preparation is critical for reliable AI implementation.
Users connect the AutoML platform with the source of the training data—ideally a large dataset containing data that is ready for use in training. The data preparation phase occurs before an AutoML solution is deployed.
The AutoML solution steps in to further preprocess and clean the data. More thorough data preprocessing leads to better AI model performance.
When manually building models for supervised learning and semi-supervised learning tasks, the training data must be manually labeled. Features and outputs must be selected based on the model’s intended use case. AutoML solutions can handle feature engineering on behalf of users to select the data features that are most likely to improve model performance.
Data features or variables are the attributes of a dataset that machine learning models use to make decisions and predictions. For example, for a computer vision model built to identify plant species, data features might include leaf shape and color.
Feature engineering is the transformative process by which a data scientist draws new information from input data and prepares it for machine learning. Good engineering and feature selection can determine the difference between acceptable and high-quality model performance.
Automated feature engineering automates the process of exploring the feature space, filling missing values and selecting features to use. Manually building a single feature can take hours, and the number of features required for a bare minimum accuracy score—let alone a production-level accuracy baseline—can reach into the hundreds. Automated feature engineering reduces this phase from days to minutes.
In addition to the efficiency benefits, automated feature efficiency also increases AI explainability—important for strictly regulated industries such as healthcare or finance. Greater feature clarity makes models more compelling and actionable by discovering new organizational KPIs.
What type of model is best for the intended use case? With traditional machine learning, model selection requires expert knowledge of AI model types along with their respective capabilities and limitations.
AutoML tools improve on traditional processes by automatically building and training several models simultaneously with a range of algorithms and hyperparameter configurations. Many AutoML solutions combine multiple models in a process known as ensemble learning.
One of the most complicated, error-prone and time-consuming tasks when building deep neural networks is the creation of the neural architecture. Advanced tasks require multi-layered networks with complex hyperparameter configurations.
Neural architecture search (NAS) automates this process, reducing the time spent and potential for error. With the use of advanced algorithms, NAS identifies the optimal architecture based on the context and dataset. Recent advancements in NAS focus on the development of more efficient techniques to reduce the associated computational costs.
Hyperparameters are the rules that govern the model’s learning process. Unlike the internal parameters that a model updates during training, hyperparameters are external to the model and are configured by data scientists. Neural network structure is also defined by hyperparameters.
In small-scale data modeling contexts, hyperparameters can be manually configured and optimized through trial and error. But with deep learning applications, the number of hyperparameters grows exponentially. Automated hyperparameter optimization allows teams to iterate and experiment to discover the best hyperparameters across features and models.
Hyperparameter tuning is automated through advanced algorithms such as Bayesian optimization. Automated hyperparameter tuning frees data scientists to focus on the why of model creation rather than the how during the machine learning process. Analytics teams can instead focus on optimizing models for designated use cases—for example, to minimize false negatives in medical testing.
Data scientists need to validate a machine learning algorithm’s progress during training. After training, the model is tested with new data to evaluate its performance before real-world deployment. The model’s performance is evaluated with metrics including a confusion matrix, F1 score, ROC curve and others.
When training is complete, the AutoML tool tests each model to identify which performs best on the training and test datasets, then automatically selects the top-performing model for deployment.
Model creation is just the first step in the product timeline. Completed models need to be made available to users, monitored for performance and maintained over time to help ensure reliability and accuracy. Without automation, development teams must write scripts and build systems to integrate the model into their operations and deliver it to its user base.
Many AutoML solutions include deployment tools for seamless real-world integration. Models can be deployed as a service accessible through a website, app or API connection. AutoML platforms can automate model deployment into pre-existing product offerings, manage scaling, updates and versioning, and increase explainability with data visualization.
The diverse array of AutoML tools means that the technique can be applied to a wide range of machine learning tasks, including:
Classification
Regression
Computer vision
Natural language processing
Classification is the machine learning task of assigning data inputs into designated categories. Predictive models use input data features to predict the correct labels, or outputs. AutoML systems can build and test an array of algorithms, such as random forests and support vector machines (SVM), to process tabular data.
AutoML tools automatically detect patterns in labeled datasets and can design models for common classification tasks such as fraud detection and email spam filtering.
Regression in machine learning is the challenge of using historical data to predict future values. Linear regression predicts the value of a dependent variable based on one or more independent variables—for example, with risk analysis or market forecasting. Logistic regression predicts the probability of a future event, such as a patient’s likelihood of contracting an illness, as opposed to a discrete value.
AutoML streamlines the process of establishing relationships between input variables and the target variables, notably with complex multivariate tasks.
Computer vision is the use of computers to process visual data, such as images and video. AutoML systems can generate models geared for vision-based classification tasks including object detection, image classification and intelligent optical character recognition. Use cases can cover content moderation and filtering, image tagging and other related tasks.
AutoML systems can also fine-tune models for use in more advanced computer vision contexts, such as with self-driving automobiles.
Natural language processing (NLP) enables AI systems to interpret textual inputs, such as user prompts and legal documents. Chatbot creation, multi-class and multi-label text classification, customer sentiment analysis, named entity recognition and language translation are all examples of complex NLP tasks that can be easily handled with AutoML.
Data scientists can create custom models with AutoML that are automatically optimized for strong performance in their intended use cases. Otherwise, when building NLP models manually, data scientists must either start from scratch or base their models on previous ones that might not perform as well as a tailored, automatically generated model.
While AutoML brings many benefits to AI developers, it is not a wholesale replacement for human knowledge, experience, skills and creativity. The limitations of AutoML include:
High costs: The more demanding the task, the more advanced the corresponding model must be. AutoML costs can quickly spiral out of control when the technique is applied to creating large, complex models.
Lack of interpretability: AutoML-generated models can sometimes fall into the trap of “black box AI,” where the model’s inner workings are obtuse. Human developers can build models that are designed in accordance with the principles of explainable AI, but this isn’t guaranteed with AutoML solutions.
Risk of overfitting: Overfitting—where a trained model hews too closely to its training data and fails to transfer its learning to real-world data—can be mitigated with human intervention and careful monitoring of the learning process.
Limited control: Developers sacrifice control for efficiency with automation. In niche cases where highly customized models are needed, AutoML solutions can struggle to deliver an appropriate model.
Data reliance: An AI model is as strong as its training data. Both human-made and AutoML-created models cannot perform well if they are not provided with high-quality data.
AI model creators have a wide range of AutoML tools at their fingertips. Options include:
AutoKeras: An open source tool built on the Keras library and TensorFlow.
Auto-PyTorch: An AutoML solution designed to automate machine learning projects created with PyTorch.
Google Cloud AutoML: Google’s AutoML solution available on its Cloud platform for machine learning.
Lale1: An open source semi-automated Python library that integrates seamlessly with scikit-learn pipelines.
Microsoft Azure AutoML: Developers using Microsoft Azure can benefit from its AutoML capabilities.
Auto-Sklearn: An open source AutoML platform based on the scikit-learn library.
IBM® Granite™ is our family of open, performant and trusted AI models, tailored for business and optimized to scale your AI applications. Explore language, code, time series and guardrail options.
We surveyed 2,000 organizations about their AI initiatives to discover what's working, what's not and how you can get ahead.
Learn how to incorporate generative AI, machine learning and foundation models into your business operations for improved performance.
Learn how to select the most suitable AI foundation model for your use case.
Machine learning is a branch of AI and computer science that focuses on using data and algorithms to enable AI to imitate the way that humans learn.
Dive into the 3 critical elements of a strong AI strategy: creating a competitive edge, scaling AI across the business and advancing trustworthy AI.
Train, validate, tune and deploy generative AI, foundation models and machine learning capabilities with IBM watsonx.ai, a next-generation enterprise studio for AI builders. Build AI applications in a fraction of the time with a fraction of the data.
Put AI to work in your business with IBM's industry-leading AI expertise and portfolio of solutions at your side.
Reinvent critical workflows and operations by adding AI to maximize experiences, real-time decision-making and business value.