Preparing for migration in IBM Cloud Pak for Data
Before you import any data to Cloud Pak for Data, complete a set of setup tasks.
Tasks to complete on the Cloud Pak for Data system before migrating
- Installing the migration toolkit
- Installing the latest version of the cpd-cli command-line interface
- Setting up the export-import utility
- Setting environment variables
- Increasing the expiry time for tokens
- Increasing the ephemeral storage for the wkc-data-rules pod
- Providing connection information for storing rule output
Installing the migration toolkit
Download and install the latest version of the migration toolkit for Cloud Pak for Data.
- Download the migration toolkit for Cloud Pak for Data
from this support page:
This document is updated when a new version of the toolkit is released and also contains information about any prerequisite patches that you might need to install.
- Unpack the downloaded file and follow the installation instructions in the support page from where you downloaded the file.
Installing the latest version of the cpd-cli command-line interface
Ensure that you have the export and import utility set up. The latest version of the
cpd-cli command-line interface (CLI) and related modules must be installed. For
more information, Installing the Cloud Pak
for Data command-line interface (cpd-cli).
Setting up the export-import utility
cpd-cli export-import list aux-modules --namespace=${NAMESPACE} --profile=${PROFILE_NAME} --arch=${CPU_ARCH}Setting environment variables
CP4D_HOST=<cp4d host>
CP4D_USERNAME=<default platform administrator user: cpadmin or admin>
CP4D_PASSWORD=<cp4d password>
Increasing the expiry time for tokens
- Check whether a token expiry time is set by running the following
command:
oc get configmap product-configmap --namespace=${NAMESPACE} -o custom-columns=test:.data.TOKEN_EXPIRY_TIME | sed '1d'If the parameter
TOKEN_EXPIRY_TIMEis set, note the original setting to be able to reset the value after the migration is complete. - Change the
TOKEN_EXPIRY_TIMEsetting to a large value, such as 48 hours, by running the following command:oc patch configmap product-configmap --namespace=${NAMESPACE} --type=merge --patch="{\"data\": {\"TOKEN_EXPIRY_TIME\": \"48\"}}" - Check whether a token refresh period is set by running the following
command:
oc get configmap product-configmap --namespace=${NAMESPACE} -o custom-columns=test:.data.TOKEN_REFRESH_PERIOD | sed '1d'If the parameter
TOKEN_REFRESH_PERIODis set, note the original setting to be able to reset the value after the migration is complete. - Change the
TOKEN_REFRESH_PERIODsetting to a large value, such as 48 hours, by running the following command:oc patch configmap product-configmap --namespace=${NAMESPACE} --type=merge --patch="{\"data\": {\"TOKEN_REFRESH_PERIOD\": \"48\"}}" - Restart the
usermgmtpods by running the following command:oc delete pods --namespace=${NAMESPACE} -l component=usermgmt
Increasing the ephemeral storage for the wkc-data-rules pod
wkc-data-rules pod when
migrating data quality rules, increase the pod's ephemeral storage. You can revert this setting
after migration or continue working with the increased storage settings.- Log in to the Red Hat® OpenShift® Container Platform as a user with administrator rights.
- Check whether a value for the
ephemeral-storageparameter is set for thewkc-data-rulespod and whether the value is less than2Gi.oc get deployment wkc-data-rules --output="jsonpath={.spec.template.spec.containers[*].resources.limits.ephemeral-storage}" && echo -e "\n" - If the
ephemeral-storageparameter values is less than2Gi, run the following command to set the value to2Gi:oc patch wkc wkc-cr -n ${NAMESPACE} --type merge -p '{"spec":{"wkc_data_rules_resources":{"requests":{"cpu":"100m","memory":"800Mi","ephemeral-storage":"50Mi"},"limits":{"cpu":1,"memory":"2048Mi","ephemeral-storage": "2Gi" }}}}'
Providing connection information for storing rule output
A platform connection to the location and datasource to which the data rule output tables can be persisted must exist prior to migration.
Create a platform connection for storing rule output tables and save the connection name to the
DQ_RULES_CONNECTION_NAME variable. To create such connection, complete the steps
described in Connecting to data sources at the platform level.
The connection must be configured to use shared credentials.
Only Db2, Oracle, or Microsoft SQL Server data sources are supported.
The connection details must be made available as input to the export on the InfoSphere Information Server system.
If the InfoSphere Information Server engine tier and Cloud Pak for Data system can communicate over port 443, the information is later retrieved as part of the export process. For more information, see step 5 in Exporting data from the InfoSphere Information Server system.
${TOOLKIT_PATH}/get_dq_rule_connection.sh --url https://${CP4D_HOST} -u ${CP4D_USERNAME} -p ${CP4D_PASSWORD} -name ${DQ_RULES_CONNECTION_NAME}Transfer this file to the InfoSphere Information Server system as part of the export process. For more information, see step 5 in Exporting data from the InfoSphere Information Server system.
What to do next
Now, you are ready to migrate user information from the InfoSphere Information Server system to Cloud Pak for Data.