AI-powered innovation to fuel business growth
Infuse AI into every transaction without data movement while meeting stringent service level agreements (SLAs) and response times.
Run AI where your data resides to protect sensitive information and meet regulatory requirements.
With IBM z17, process up to 450 billion inference operations per day with 1 ms response time for real-time use cases.1
Route inference requests to any idle Integrated Accelerator for AI to boost throughput up to 7.5x over IBM z16.2
IBM watsonx Assistant for Z delivers secure, AI-powered virtual agents at scale on IBM Z for smarter customer interactions and agentic workflows.
Machine learning for IBM z/OS allows users to deploy machine learning models within transactional applications while maintaining SLAs.
AI Toolkit for IBM Z is a family of supported open source AI frameworks optimized for the Telum processor and use on-chip AI acceleration in IBM z16® and z17 systems.
IBM Synthetic Data Sets is a family of artificially generated datasets designed to enhance predictive AI model training and LLMs.
¹ DISCLAIMER: Performance result is extrapolated from IBM® internal tests running on IBM Systems Hardware of machine type 9175. The benchmark was executed with 1 thread performing local inference operations using a LSTM based synthetic Credit Card Fraud Detection model to exploit the Integrated Accelerator for AI. A batch size of 160 was used. IBM Systems Hardware configuration: 1 LPAR running Red Hat® Enterprise Linux® 9.4 with 6 IFLs (SMT), 128 GB memory. 1 LPAR with 2 CPs, 4 zIIPs and 256 GB memory running IBM z/OS® 3.1 with IBM z/OS Container Extensions (zCX) feature. Results may vary.
2 DISCLAIMER: Performance results are based on internal tests exploiting the IBM Integrated Accelerator for AI for inference operations on IBM z16 and z17. On IBM z17, each IBM Integrated Accelerator for AI allows any CPU within a drawer to direct AI inference request to any of the 8 idle AI accelerators on the same drawer. The tests involved running inference operations on 8 parallel threads with batch size of 1. Both IBM z16 and z17 were configured with 2 GCPs, 4 zIIPs with SMT and 256 GB memory on IBM z/OS V3R1 with IBM Z Deep Learning Compiler 4.3.0, using a synthetic credit card fraud detection model. Results may vary.