Replacing an operational VMware node on installer-provisioned infrastructure
Use this information to replace an operational VMware node on an installer-provisioned infrastructure.
Procedure
What to do next
Verify that the new node and all pods are running.
- Verify that the new node is present in the
output:
oc get nodes --show-labels | grep cluster.ocs.openshift.io/openshift-storage= |cut -d' ' -f1 - and confirm
that at least the following pods on the new node are in a Running state:
csi-cephfsplugin-*csi-rbdplugin-*
- Verify that all the other required Fusion Data Foundation pods are in Running state.
- Verify that the new Object Storage Device (OSD) pods
are running on the replacement
node:
oc get pods -o wide -n openshift-storage| egrep -i <new_node_name> | egrep osd - If cluster-wide encryption is enabled on the
cluster, verify that the new OSD devices are encrypted.
For each of the new nodes identified in the previous step, do the following:
- Create a debug pod and open a chroot environment for the one or more selected
hosts:
oc debug node/<node_name>chroot /host - Display the list of available block devices:, using the lsblk
command.
Check for the
cryptkeyword beside the one or moreocs-devicesetnames.
- Create a debug pod and open a chroot environment for the one or more selected
hosts:
-
If the verification steps fail, contact IBM Support.