Deployment profiles
The following sections describe basic, production, and HA set up options available for an operator.
StarterPak
StarterPak is a basic setup that does not require any specific configuration. It is intended for the nonproduction environments only.
Custom Resource Definition
For more information about the basic setup process, see Basic setup section.
Node resources
The following table shows specifications for IBM Process Mining and IBM Task Mining:
Software | Memory (GB) | CPU (cores) | Disk (GB) | Nodes |
---|---|---|---|---|
IBM Process Mining | 64 | 16 | 100 | 1 |
IBM Task Mining | 16 | 4 | 100 | 1 |
Total | 80 | 20 | 200 | 2 |
Production Setup
In this section, you can learn about the configuration for the production installation.
Custom Resource Definition
For more information about the production setup process, see the Production setup section.
Node Resources
The following information is needed to set up a production environment:
- Number of projects that are managed in the application and their classification. You can classify the project as:
- Flat
Only one business entity is involved. - Multi-level
From 2 to 5 involved and mapped business entities.
- Flat
Multi-level processes require more complex and resource-consuming algorithms than flat processes.
- Number of events per each project in the range:
- Up to 10 million.
- From 10 million to 50 million.
- Over 50 million.
The same number of events requires more computing resources for a multi-level process, especially when there are high cardinalities between intersecting business entities, for example 1:10000 between contracts and invoices.
Number of
currently active users in the application and what they use it for can impact the right sizing of the system.
In IBM Process Mining, the quality of data is more important than the quantity of collected events. The number of events must
be small enough to provide a full range of variants and reliable statistics for the relevant period. The presence of stale data relating to previous periods can impact exploration and statistics because they can reflect workflows that no
longer exist.
- Number of mapped custom fields per project in the range:
- Up to 20.
- From 20 to 50.
- From 50 to 80.
No more than 80 custom fields can be used.
Custom fields affect the complexity of the processing.
- The number of users that have access to the application and the number of users who additionally have access to the data analytics.
You need to know how many snapshots of the project you must store and manage.
Analytics is the most resource-consuming component of the suite, especially for RAM usage.
Following tables show sizing ranges based on quantity of events with 10 active users in the same time:
-
Up to 10 million events
Software Memory (GB) CPU (cores) Disk (GB) Nodes IBM Process Mining 64 16 300 1 -
Up to 50 million events
Software Memory (GB) CPU (cores) Disk (GB) Nodes IBM Process Mining 128 32 600 1 -
Up to 100 million events
Software Memory (GB) CPU (cores) Disk (GB) Nodes IBM Process Mining 192 48 1000 1
Task mining
For IBM Task Mining you can use a common cofiguration:
Software | Memory (GB) | CPU (cores) | Disk (GB) | Nodes |
---|---|---|---|---|
IBM Task Mining | 32 | 8 | 300 | 1 |
HA Setup
In this section, you can learn about the configuration for the HA installation.
CRD
For more details on how to increase the number of pods duplicate, see the Custom setup section.
Node Resources
The following table is an example of a configuration for an installation that can manage up to 50 million events:
Software | Memory (GB) | CPU (cores) | Disk (GB) | Nodes |
---|---|---|---|---|
IBM Process Mining | 128 | 48 | 200 | 3 |
IBM Task Mining | 32 | 8 | 300 | 3 |
Total | 160 | 56 | 500 | 6 |