Designed to enable a turnkey solution for AI enterprise workloads
Enterprises are facing real obstacles when it comes to scaling AI across the organization, such as data and integration complexity, shortages of the right skills, security and compliance risks, and more. IBM® Power® AI offerings and IBM Spyre™ Accelerator for Power remove these barriers through full stack optimization. The result? IBM Power provides an accelerated, flexible, and secured platform designed for enterprise AI workloads.
Chose from turnkey solutions that deploy in one click to simplify setup and accelerate workloads with IBM Spyre Accelerator for Power.
Run traditional and GenAI models on Power servers with Spyre, watsonx.data, OpenShift AI, and Red Hat AI Inference Server optimized for enterprise workloads.
Harness both on- and off-chip acceleration and enterprise-class infrastructure optimized for both traditional and generative AI workloads.
Embed AI directly into enterprise knowledge bases. Greater than 8 million document embeddings for knowledge base integration every hour using Spyre Accelerator for Power with batch and prompt sizes of 128.3
Predict IT issues, detect and fix incidents, and forecast and plan capacity.
Analyze medical images and automate claims and EHR matching with digital assistants.
Detect fraud, enable anti-money laundering, and accelerate risk and underwriting processes.
Manage claims, prevent fraud, optimize underwriting, and support customer interactions.
Learn how a hospital used AI and IBM technology to streamline cancer diagnostics, improving speed, accuracy, and focus on high-risk cases.
Learn how a family-owned distributor quickly recovered from a ransomware attack using IBM i and Power, restoring critical operations and driving modernization with generative AI tools.
Learn how a logistics provider used AI on IBM i and Power to cut order processing time by 80% and speed up handling 5X — boosting accuracy and customer response.
1. AI service of the IBM-supported catalog is delivered as one or a set of containers that can be deployed with a single deployment command. The provided UI for the catalog executes such commands in the backend based on a single click within the UI page of the respective AI service.
2. Single configuration enabled by exposed industry standard APIs to decouple services at the top and the backing inferencing service for all AI services that are part of the IBM-supported catalog. Any service that requires AI inferencing capabilities can connect inferencing services that provide OpenAI API or watsonx.ai API compliant inferencing endpoints (Spyre endpoint, RH AI Inferencing Server, IBM Cloud, OpenAI, Azure, AWS, GCP, ...). Services can run either on IBM Power or on IBM Power Virtual Server.
3. Based upon internal testing running 1M unit data set with prompt size 128, batch size 128 using 1-card container. Individual results may vary based on workload size, use of storage subsystems and other conditions.