Sizing your hardware for scalable data collection

Before you implement this architecture, consider the hardware requirements for the components that make up your scalable data collection architecture.

The servers can be metal or virtualized. If you use a virtualized server, ensure that hyper threading is disabled and that physical cores are allocated and reserved.

For an example deployment, see Example sizing.

This document outlines the hardware requirements for your scalable data collection components. For more information about the general hardware requirements see the information on Hardware requirements.

The following table summarizes the hardware requirements for each server and the recommended number of servers.

Table 1. Hardware requirements
Component Daily data ingestion per server in GB Processors per server RAM per server in GB Recommended number of servers
IBM® Tivoli® Monitoring Log File Agent 200 8 4 2
HAProxy 400 8 4 1
Receiver cluster 200 12 32 2
Apache Kafka 200 12 16 2
Sender cluster 200 12 32 2
Log Analysis server 400 8 32 1
Indexing Engine server 400 8 48 1
Note the following settings when tuning your hardware:
  • For both the Sender and Receiver servers, set the heap size to 16 GB.
  • For your Apache Kafka brokers, set the heap size to 4 GB. Set the heap size for Apache ZooKeeper to 1 GB.
  • Apache Kafka caches data for 7 days on disk. This means that if you plan to stream 400 GB of data daily, reserve 3.2 Terabytes (TB) of disk space for this cache.

For more information, see Tuning considerations.