AI and Analytics: Insights at scale and AIOps
Infuse AI in real-time into every business transaction and automate your IT infrastructure through AI infused applications
Overview
Infuse AI in real-time into every business transaction, driving top-line growth and bottom-line savings, for your mission critical applications while meeting the most stringent SLA’s. Leverage both IBM and open-source solutions to enable your data scientists and engineers to use the applications they know and trust.
Automate your IT infrastructure through AI infused applications, while enhancing your security, monitoring, data privacy and traditional processes. Stay current on innovations and enhancements to constantly improve core IT operations and infrastructure management with AI based functionality.
- Enable problem identification, isolation and resolution on IBM Z through analysis of structured and unstructured operational data.
- Leverage anomoly detection and problem identification to reduce time for root cause analysis, anticipate, and adjust.
- Visualize IBM Z data in the same context as the rest of your hybrid ecosystem, enabling focus on progressing AIOps transformation.
Value
20x
Lower inferencing response time vs sending the same inferencing operations off platform 1
19x
Higher throughput with inferencing vs sending the same inferencing operations off platform 1
Footnotes
- IBM z16 with z/OS delivers up to 20x lower response time and up to 19x higher throughput when co-locating applications and inferencing versus sending the same inferencing operations to a compared x86 cloud server with 60ms average network latency.*Disclaimer: Performance result is extrapolated from IBM internal tests running local inference operations in a z16 LPAR with 48 IFLs and 128 GB memory on Ubuntu 20.04 (SMT mode) using a synthetic credit card fraud detection model (https://github.com/IBM/ai-on-z-fraud-detection) exploiting the Integrated Accelerator for AI. The benchmark was running with 8 parallel threads each pinned to the first core of a different chip. The lscpu command was used to identify the core-chip topology. A batch size of 128 inference operations was used. Results may vary.