Analytics Analytics Engine IBM Analytics Engine
Get back control – separate compute from storage to flexibly build, scale and maintain your analytics applications
Try today for free
 internal view of an engine
Now available:

Scale AI workloads, for all your data, anywhere

What is Analytics Engine?


IBM Analytics Engine provides Apache Spark environments a service that decouples the compute and storage tiers to control costs, and achieve analytics at scale. Instead of a permanent cluster formed of dual-purpose nodes, IBM Analytics Engine enables users to store data in an object storage layer such as IBM Cloud Object Storage and spins up clusters of compute notes when needed. For added flexibility and cost predictability, usage-based consumption is available for Apache Spark environments.

IBM Analytics Engine Serverless plan is now available on IBM Cloud
Put your focus back on analytics
Improve cluster utilization

Consume instances only when jobs are running

Control costs

Pay solely for what you use

Scale flexibly

Optimize resources by separating compute and storage

Analytics Engine features
Leverage open-source power

Build on an ODPi-compliant stack with pioneering data science tools with the broader Apache Spark ecosystem.

Spin up and scale on demand

Define clusters based on your application's requirement. Choose the appropriate software pack, version, and size of the cluster. Use as long as required and delete as soon as application finishes jobs.

Customize and configure analytics

Configure clusters with third-party analytics libraries and packages as well as IBM’s own enhancements. Deploy workloads from IBM Cloud services, such as machine learning.

Analytics Engine benefits
Compute and storage are no longer bound

Spin up compute-only clusters on demand. Because no data is stored in the cluster, clusters never need to be upgraded.

I/O-heavy clusters are more cost-effective

Provision more IBM Cloud Object Storage (or other data stores) on demand with no extra costs for compute cycles not used.

Clusters are more elastic

Adding and removing data nodes based on live demand is possible via REST APIs. Also, overhead costs remain low because there is no data stored in the compute cluster.

Security is more cost-effective

Using a multilayered approach significantly simplifies the individual cluster security implementation, while enabling access management at a more granular level.

Vendor lock-in is avoided

Clusters are spun up to meet the needs of the job versus forcing jobs to conform to a single software package/version. Multiple different versions of software can be run in different clusters.

Control costs with Serverless Spark

If you’re working with Apache Spark, but unsure how much resource is needed, provision a Serverless Spark instance that only consumes compute resource when running an app. Pay only for what you use.

Related products
IBM Watson IBM Cloud Pak for Data

Collect, organize, and analyze data on an open, multicloud data and AI platform.

IBM Knowledge Catalog

Intelligent data and analytic asset discovery, cataloging and governance to fuel AI apps

Get started on Analytics Engine in minutes

Get the combined service that provides an environment for developing and deploying advanced analytics applications

Sign up