Replacing a failed Google Cloud Platform node on installer-provisioned infrastructure
Use this information to replace a failed GCP node on an installer-provisioned infrastructure.
Procedure
What to do next
Verify that the new node and all pods are running.
- Verify that the new node is present in the
output:
oc get nodes --show-labels | grep cluster.ocs.openshift.io/openshift-storage= |cut -d' ' -f1
csi-cephfsplugin-*
csi-rbdplugin-*
and confirm
that at least the following pods on the new node are in a Running state:- Verify that all the other required Fusion Data Foundation pods are in Running state.
- Verify that the new Object Storage Device (OSD) pods
are running on the replacement
node:
oc get pods -o wide -n openshift-storage| egrep -i <new_node_name> | egrep osd
- If cluster-wide encryption is enabled on the
cluster, verify that the new OSD devices are encrypted.
For each of the new nodes identified in the previous step, do the following:
- Create a debug pod and open a chroot environment for the one or more selected
hosts:
oc debug node/<node_name>
chroot /host
- Display the list of available block devices:, using the lsblk
command.
Check for the
crypt
keyword beside the one or moreocs-deviceset
names.
- Create a debug pod and open a chroot environment for the one or more selected
hosts:
-
If the verification steps fail, contact IBM Support.