Installing Db2 Data Gate
An instance administrator can install Db2 Data Gate on IBM Cloud Pak® for Data Version 4.7.
- Who needs to complete this task?
-
Instance administrator To install Db2 Data Gate, you must be an instance administrator. An instance administrator has permission to install software in the following projects:
- The operators project for the instance
-
The operators for this instance of Cloud Pak for Data are installed in the operators project.
In the installation commands, the
${PROJECT_CPD_INST_OPERATORS}
environment variable refers to the operators project. - The operands project for the instance
-
The Cloud Pak for Data control plane and the services for this instance of Cloud Pak for Data are installed in the operands project.
In the installation commands, the
${PROJECT_CPD_INST_OPERANDS}
environment variable refers to the operands project.
- When do you need to complete this task?
-
Review the following options to determine whether you need to complete this task:
- If you want to install the Cloud Pak for Data control plane and one or more services at the same time, follow the process in Installing an instance of Cloud Pak for Data instead.
- If you didn't install Db2 Data
Gate when you installed the Cloud Pak for Data control plane, complete this task to add Db2 Data
Gate to your environment.
Repeat as needed If you are responsible for multiple instances of Cloud Pak for Data, you can repeat this task to install more instances of Db2 Data Gate on the cluster.
Information you need to complete this task
Review the following information before you install Db2 Data Gate:
- Version requirements
-
All of the components that are associated with an instance of Cloud Pak for Data must be installed at the same release. For example, if the Cloud Pak for Data control plane is installed at Version 4.7.4, you must install Db2 Data Gate at Version 4.7.4.
- Environment variables
-
The commands in this task use environment variables so that you can run the commands exactly as written.
- If you don't have the script that defines the environment variables, see Setting up installation environment variables.
- To use the environment variables from the script, you must source the environment variables
before you run the commands in this task. For example,
run:
source ./cpd_vars.sh
- Security context constraint
-
Db2 Data Gate works with the default Red Hat® OpenShift® Container Platform security context constraint:
- On Version 4.10, the default SCC is
restricted
. - On Version 4.12, the default SCC is
restricted-v2
- On Version 4.10, the default SCC is
- Storage requirements
- You must specify storage classes when you install Db2 Data Gate. The following storage classes are recommended. However, if you don't use these storage classes on your cluster, ensure that you specify a storage class with an equivalent definition.
Before you begin
This task assumes that the following prerequisites are met:
Prerequisite | Where to find more information |
---|---|
The cluster meets the minimum requirements for installing Db2 Data Gate. | If this task is not complete, see System requirements. |
The workstation from which you will run the installation is set up as a client workstation
and includes the following command-line interfaces:
|
If this task is not complete, see Setting up a client workstation. |
The Cloud Pak for Data control plane is installed. | If this task is not complete, see Installing an instance of Cloud Pak for Data. |
For environments that use a private container registry, such as air-gapped environments, the Db2 Data Gate software images are mirrored to the private container registry. | If this task is not complete, see Mirroring images to a private container registry. |
For environments that use a private container registry, such as air-gapped
environments, the cpd-cli is configured to pull the olm-utils-v2 image from the private container registry. |
If this task is not complete, see Pulling the olm-utils-v2 image from the private container registry. |
The node settings are adjusted for Db2 Data Gate. | If this task is not complete, see Changing required node settings. |
Procedure
Complete the following tasks to install Db2 Data Gate:
Installing the service
To install Db2 Data Gate:
-
Run the
cpd-cli manage login-to-ocp
command to log in to the cluster as a user with sufficient permissions to complete this task. For example:cpd-cli manage login-to-ocp \ --username=${OCP_USERNAME} \ --password=${OCP_PASSWORD} \ --server=${OCP_URL}
Tip: Thelogin-to-ocp
command takes the same input as theoc login
command. Runoc login --help
for details. - Run the following command to create the required OLM objects for Db2 Data
Gate in the
operators project for the
instance:
cpd-cli manage apply-olm \ --release=${VERSION} \ --cpd_operator_ns=${PROJECT_CPD_INST_OPERATORS} \ --components=datagate
Wait for thecpd-cli
to return the following message before you proceed to the next step:[SUCCESS]... The apply-olm command ran successfully
If the
apply-olm
fails, see Troubleshooting the apply-olm command during installation or upgrade. - Create the custom resource for Db2 Data
Gate.
The command that you run depends on the storage on your cluster.
Red Hat OpenShift Data Foundation storage
Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
IBM Storage Fusion Data Foundation storage
Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
IBM Storage Fusion Global Data Platform storage
Remember: When you use IBM Storage Fusion Global Data Platform storage, both${STG_CLASS_BLOCK}
and${STG_CLASS_FILE}
point to the same storage class, typicallyibm-spectrum-scale-sc
.Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
IBM Storage Scale Container Native storage
Remember: When you use IBM Storage Scale Container Native storage, both${STG_CLASS_BLOCK}
and${STG_CLASS_FILE}
point to the same storage class, typicallyibm-spectrum-scale-sc
.Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
Portworx storage
Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --storage_vendor=portworx \ --license_acceptance=true
NFS storage
Remember: When you use NFS storage, both${STG_CLASS_BLOCK}
and${STG_CLASS_FILE}
point to the same storage class, typicallymanaged-nfs-storage
.Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
AWS with EFS storage only
Remember: When you use EFS storage, both${STG_CLASS_BLOCK}
and${STG_CLASS_FILE}
point to the same storage class, typicallyefs-nfs-client
.Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
AWS with EFS and EBS storage
Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
NetApp Trident
Remember: When you use NetApp Trident storage, both${STG_CLASS_BLOCK}
and${STG_CLASS_FILE}
point to the same storage class, typicallyontap-nas
.Run the following command to create the custom resource.
cpd-cli manage apply-cr \ --components=datagate \ --release=${VERSION} \ --cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \ --block_storage_class=${STG_CLASS_BLOCK} \ --file_storage_class=${STG_CLASS_FILE} \ --license_acceptance=true
Validating the installation
apply-cr
command
returns:[SUCCESS]... The apply-cr command ran successfully
If you want to confirm that the custom resource status is
Completed
, you can run the cpd-cli
manage
get-cr-status
command:
cpd-cli manage get-cr-status \
--cpd_instance_ns=${PROJECT_CPD_INST_OPERANDS} \
--components=datagate
What to do next
- You must complete the post-installation setup before users can access Db2 Data Gate.
- You must give users access to Db2 Data Gate before they can use it. For details, see Managing user access to Db2 Data Gate.