Troubleshooting Hosted Control Plane clusters
Use these troubleshooting information to know the problem and workaround for the installation of spoke cluster and
Hosted Control Plane in Red Hat documentation
For Hosted Control Plane troubleshooting documentation by Red Hat®, see Red Hat Documentation.
Labels do not work
- Cause
- As the necessary components for the add-on deployment were not properly implemented in version 2.8.2, the add-ons are unusable.
- Resolution
- Do not use add-ons in version 2.8.2. To deploy Fusion services, log in to the Hosted Control Plane cluster and install IBM Storage Fusion services using the following instructions:
- Install IBM Storage Fusion on Bare Metal. See Installing IBM Storage Fusion on On-premisesBare Metal.
- Install IBM Storage Fusion spoke. See Install Backup & Restore spoke.
- Deploy storage client in the Hosted Control Plane cluster. See Deploying storage client in the Hosted Control Plane clusters.
Add-on packages are not deployed automatically
- Cause
- In the 2.8.1 release, the default catalog source is incorrect, and as a result the deployment of IBM Storage Fusion services on the managed clusters do not work.
- Resolution
-
- Log in to the OpenShift® console on the hub cluster where the Red Hat Advanced Cluster Management for Kubernetes is installed.
- Go to the
ibm-spectrum-fusion-ns
and open the configmap namedfusion-addon-config
. - In the
odf-catalog-source-image
field, add the value: icr.io/cpopen/isf-data-foundation-catalog:v4.16 - In the
sds-catalog-source-image
field, add the value: icr.io/cpopen/isf-operator-software-catalog:2.8.1 - Save the configmap.
As a resolution, apply the labels in the correct catalog that is used on the managed clusters.
Issues in installation of IBM Storage Fusion in a hosted cluster
- Resolution
-
- Check the status of
manifestwork
addon-fusion-base-deploy-0
inclusternamespace
of the hub cluster:oc login to the hub oc get manifestwork addon-fusion-base-deploy-0 -n <spoke_cluster_name> -o yaml
- If no error exists in the status of the
manifestwork
, check the status of IBM Storage Fusion installation in the spoke cluster:oc login to the spoke oc get csv -n ibm-spectrum-fusion-ns
- Check the status of
Issues in the installation of Fusion Data Foundation in the hosted cluster
- Resolution
- As a resolution, check the status of
manifestwork
addon-fusion-odf-deploy-0
inclusternamespace
in the hub cluster:oc login to the hub oc get manifestwork addon-fusion-odf-deploy-0 -n <spoke_cluster_name> -o yaml
Backup issues in Hosted Control Plane with Fusion Data Foundation
- Problem statement
- Concurrent backups fail during a Velero snapshot with the following error message:
The operation has timed out because Velero has failed to report status.' However, the backup phase is updated as 'DataTransferfailed.
The timeout for Velero to pickup request is set to 30 minutes in the transaction manager.
- Resolution
- To resolve the issue, increase it to 60 minutes.
Hosted Control Plane cluster does not get created because of unavailability of IP addresses
- Resolution
- Check whether the installed load balancer has sufficient IP addresses for the Hosted Control Plane cluster. Check the
IPAddressPool
object for the IP range. Run the following command to check whether an IP is available:oc get svc -A | grep LoadBalancer
Known issues and limitations
- Random "image pull" failure can occur on the Hosted Control Plane due to secrets.
- Sometimes, the hosted cluster status goes in offline mode after deletion. Contact IBM Support to resolve the issue.
- The hcp destroy command can cause the hosted cluster to get stuck indefinitely during cleanup. Contact IBM Support to resolve the issue.
- The YAML tab view on the OpenShift Container Platform console is not working as expected during the installation of the multi cluster engine operator.
- The hosted clusters expect the
storageprofile
of the used storage class, but it is not available during cluster creation. - The following issues occur whenever you remove the disks and place them back in the rack:
- Disks are not reflected in the node.
- LVM cluster and Data Foundation go into a degraded state.