AI on IBM Z® brings real-time insights by applying machine learning directly to transactional data—eliminating the need for data movement.
Leveraging the advanced hardware and software stack of IBM z17, businesses can scale multiple AI models to power predictive use cases such as fraud detection and retail automation. With high throughput, low latency, and industry-leading cyber-resilience, IBM Z is built for mission-critical AI.
With IBM z17, process up to 450 billion inference operations per day with 1 ms response time for real-time use cases.1
Infuse AI into every transaction—no data movement—while meeting the most stringent SLAs and response times.
Route inference requests to any idle Integrated Accelerator for AI to boost throughput up to 7.5x over IBM z16.2
Run AI where your data already resides to safeguard sensitive information and support regulatory compliance.
Unlock the potential of generative AI with IBM watsonx Code Assistant for Z and watsonx Assistant for Z . These tools enable hybrid or on-premise AI solutions, with future capabilities planned through the Spyre Accelerator3—expanding the reach of AI across your enterprise infrastructure.
A generative AI-powered tool that provides an end-to-end application developer lifecycle. It includes application discovery and analysis, automated code refactoring and COBOL to Java conversion.
Consists of IBM® Elite Support and IBM Secure Engineering. These vet and scan open-source AI, serving frameworks and IBM-certified containers for security vulnerabilities, and validate compliance with industry regulations.
An AI solution for users to build machine learning models by using any platform of choice and deploy those models within transactional applications while maintaining SLAs.
A family of artificially generated datasets designed to enhance predictive AI model training and LLMs to benefit IBM Z enterprises in financial services to gain quick access to relevant and rich data for AI projects.
IBM watsonx Assistant for Z delivers secure, AI-powered virtual agents at scale on IBM Z for smarter customer interactions.
IBM Db2 for z/OS powers secure, agile data serving for hybrid cloud, transactional, and analytics workloads.
The Python AI Toolkit for IBM z/OS gives you key open-source tools to run AI and ML workloads on IBM Z.
IBM ZDNN Plug-in for TensorFlow lets you deploy AI models near your core apps on IBM Z using the Integrated Accelerator for AI.
IBM Z Platform for Apache Spark supports fast in-memory analytics using Java, Scala, Python, and R on IBM Z.
IBM Z Deep Learning Compiler runs ONNX AI models on IBM Z with low dependency using the Integrated Accelerator for AI.
Anaconda on IBM Z and LinuxONE runs Scikit-learn, NumPy, and more in zCX containers for efficient data science.
Discover how a multiple AI model approach accelerates Anti-Money Laundering (AML) on IBM z17 and drives accuracy improvements to detect illicit activity, simplify regulatory compliance, and safeguard economic transparency.
Learn about the key elements of current and next-generation mainframes and how they are transforming the AI landscape.
Discover how mainframes are more relevant than ever for hybrid cloud optimization, AI innovation, and digital transformation.
Learn how to enable AI solutions in business-critical use cases, such as fraud detection and credit risk scoring, on the platform.
Discover low-latency AI on a highly trustworthy and secure enterprise system: the modernized IBM Mainframe.
¹ DISCLAIMER: Performance result is extrapolated from IBM® internal tests running on IBM Systems Hardware of machine type 9175. The benchmark was executed with 1 thread performing local inference operations using a LSTM based synthetic Credit Card Fraud Detection model (https://github.com/IBM/ai-on-z-fraud-detection) to exploit the Integrated Accelerator for AI. A batch size of 160 was used. IBM Systems Hardware configuration: 1 LPAR running Red Hat® Enterprise Linux® 9.4 with 6 IFLs (SMT), 128 GB memory. 1 LPAR with 2 CPs, 4 zIIPs and 256 GB memory running IBM z/OS® 3.1 with IBM z/OS Container Extensions (zCX) feature. Results may vary.
2 DISCLAIMER: Performance results are based on internal tests exploiting the IBM Integrated Accelerator for AI for inference operations on IBM z16 and z17. On IBM z17, each IBM Integrated Accelerator for AI allows any CPU within a drawer to direct AI inference request to any of the 8 idle AI accelerators on the same drawer. The tests involved running inference operations on 8 parallel threads with batch size of 1. Both IBM z16 and z17 were configured with 2 GCPs, 4 zIIPs with SMT and 256 GB memory on IBM z/OS V3R1 with IBM Z Deep Learning Compiler 4.3.0, using a synthetic credit card fraud detection model (https://github.com/IBM/ai-on-z-fraud-detection). Results may vary.
3 Upon Spyre Accelerator availability. The IBM Spyre Accelerator is current in tech preview. https://www.ibm.com/docs/en/announcements/z17-makes-more-possible