Preparing to install the service (Data Virtualization)
Plan and prepare to deploy the Data Virtualization service.
About this task
c-db2u-dv-db2u-0 pod runs the Data Virtualization head component (also known as the
The term worker pod in Data Virtualization refers to the worker service component that
runs on each
c-db2u-dv-db2u-x pod, where x
starts at 1. You can allocate multiple worker pods, which are effectively multiple
c-db2u-dv-db2u-x pods, to the Data Virtualization service
You must not confuse Data Virtualization worker pods with compute pods, which are the physical nodes that compose the Red Hat® OpenShift® cluster. For more information about cluster compute nodes, see Architecture for Cloud Pak for Data.
Before you install the Data Virtualization service, you must meet the following requirements.
Ensure that you complete the pre-installation tasks for Cloud Pak for Data installations on Red Hat OpenShift.
The Data Virtualization service runs on x86_64 hardware only.
Service resource requirements
Ensure that you meet the service requirements listed in System requirements.
* You can configure the settings that are marked with an asterisk (*). For more information about editing a service instance, see Provisioning the service.
The Data Virtualization service is provisioned to any compute node in the Cloud Pak for Data cluster that has the specified resources (cores and memory) available.
Work with IBM Sales to get a more accurate sizing based on your expected workload.
Data Virtualization supports NFS, OpenShift container storage, IBM Cloud File Storage (on Red Hat OpenShift Kubernetes Service), and Portworx for persistent storage. For information about storage requirements for Cloud Pak for Data, see Storage considerations.
- Persistent volume for Data Virtualization engine node: 50Gi.
- Persistent volume for Data Virtualization caching: 100Gi.
- XFS formatted
- NFSRequired storage class:
- PortworxRequired storage class:
- OpenShift Container StorageRequired storage class:
- IBM Cloud File Storage (on Red Hat
OpenShift Kubernetes Service)Required storage class:
- External libraries
- External libraries (that is, libraries that are not included in the Data Virtualization service) are stored on a persistent volume.
Data Virtualization automatically creates persistent
volume claims during the provisioning process.
The persistent volume claim for external libraries must have at least 50 GB available.
- Cache storage
- A data cache holds temporary data that is used frequently. By using data cache, you can reduce processing and loading time that is required when you use this data.
Kernel parameter settings
- Complete the steps in Kernel parameter settings to specify the following parameters:
- Virtual memory limit (
- Message limits (
- Shared memory limits (
- Virtual memory limit (
- If the Linux®
Kernel version on the nodes is less than 4.6, you must set the kernel semaphore limits. In
Data Virtualization, the kernel semaphore limit must meet
the minimum required values on all cluster compute
kernel.sem="250 256000 100 4096"To obtain the Linux Kernel version on each compute node in the cluster, run the uname -r command on each compute node. Alternatively, run the following command to see the kernel version of each OpenShift compute node:
oc describe node Compute-node-name | grep -i kernel
To set the kernel semaphore parameter on compute nodes in the cluster, see Changing node settings.