Provides a massively scalable, shared-nothing, in-memory data integration engine running natively in a Hadoop cluster to help bring enterprise robust capabilities to the data lake.
Delivers a rich set of data profiling capabilities to understand the assets that are moved into Hadoop.
Uses metadata management to help make sense of the enormous quantities of information in the data lake.
Delivers big data-related governance features such as impact analysis and data lineage on any integration points, enabling scalable analytics without sacrificing organizational insight.
Transforms big data projects with real-time analytical processing. Integrates with IBM Streams. Uses standard data integration conventions to gather and pass data to powerful big data analytics.