GUI events
The following table lists the events that are created for the GUI component.
Event | Event Type |
Severity | Call Home | Details |
---|---|---|---|---|
bmc_connection_error | STATE_CHANGE DEGRADED |
ERROR | no | Message: Unable to connect to BMC of POWER server {0} because an error occurred when running the /opt/ibm/ess/tools/bin/esshwinvmon.py -t check -n {1} command. |
Description: The GUI checks the connection to the BMC of the POWER server. | ||||
Cause: The GUI cannot query the BMC of the POWER server because of an error that occurred in the esshwinvmon.py script. | ||||
User Action: Run the /opt/ibm/ess/tools/bin/esshwinvmon.py -t check -n [node_name] command to check error. | ||||
bmc_connection_failed | STATE_CHANGE DEGRADED |
ERROR | no | Message: Unable to connect to BMC of POWER server {0}. |
Description: The GUI checks the connection to the BMC of the POWER server. | ||||
Cause: The GUI cannot connect to the BMC of the POWER server. | ||||
User Action: Check whether the BMC IPs and passwords are correct defined in the /opt/ibm/ess/tools/conf/hosts.yml configuration file on the GUI node. Run the /opt/ibm/ess/tools/bin/esshwinvmon.py -t check -n [node_name] command to check connection to BMC. | ||||
bmc_connection_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The connection to the BMC of POWER server {0} is OK. |
Description: The GUI checks the connection to the BMC of the POWER server. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
bmc_connection_unconfigured | STATE_CHANGE DEGRADED |
ERROR | no | Message: Unable to query health state of POWER server {0} from the BMC. The /opt/ibm/ess/tools/conf/hosts.yml configuration file does not contain a section for node {1}. |
Description: The GUI checks the connection to the BMC of the POWER server. | ||||
Cause: The GUI cannot connect to the BMC of the POWER server because of wrong configuration. | ||||
User Action: Add a section for the specified node to the /opt/ibm/ess/tools/conf/hosts.yml configuration file on the GUI node. Run the /opt/ibm/ess/tools/bin/esshwinvmon.py -t check -n [node_name] command to check connection to BMC. | ||||
gui_cluster_down | STATE_CHANGE DEGRADED |
ERROR | no | Message: The GUI detected that the cluster is down. |
Description: The GUI checks the cluster state. | ||||
Cause: The GUI calculated that an insufficient number of quorum nodes is up and running. | ||||
User Action: Check for the reason that led the cluster to lost quorum. | ||||
gui_cluster_state_unknown | STATE_CHANGE DEGRADED |
WARNING | no | Message: The GUI cannot determine the cluster state. |
Description: The GUI checks the cluster state. | ||||
Cause: The GUI cannot determine whether enough quorum nodes are up and running. | ||||
User Action: N/A | ||||
gui_cluster_up | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI detected that the cluster is up and running. |
Description: The GUI checks the cluster state. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_config_cluster_id_mismatch | STATE_CHANGE FAILED |
ERROR | no | Message: The cluster ID of the current cluster '{0}', and the cluster ID in the database do not match ('{1}'). It seems that the cluster was re-created. |
Description: When a cluster is deleted and created again, the cluster ID changes, but the GUI database still references the old cluster ID. | ||||
Cause: N/A | ||||
User Action: Clear the GUI database of the old cluster information by dropping all
psql postgres postgres -c tables by using the drop schema fscc
cascade command. Then, restart the GUI by using the systemctl restart
gpfsgui command. |
||||
gui_config_cluster_id_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The cluster ID of the current cluster '{0}' matches the cluster ID in the database. |
Description: No problems that are regarding the current configuration of the GUI and the cluster were found. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_config_command_audit_off_cluster | STATE_CHANGE DEGRADED |
WARNING | no | Message: Command Audit is turned off at the cluster level. |
Description: Command Audit is turned off at the cluster level. This configuration leads to lags in the refresh of data displayed in the GUI. | ||||
Cause: Command Audit is turned off at the cluster level. | ||||
User Action: Change the cluster configuration option commandAudit to on by using the mmchconfig commandAudit=on command, or syslogonly by using the mmchconfig commandAudit=syslogonly command. This way the GUI refreshes the data that it displays automatically when IBM Storage Scale commands are run by using the CLI on other nodes in the cluster. | ||||
gui_config_command_audit_off_nodes | STATE_CHANGE DEGRADED |
WARNING | no | Message: Command Audit is turned off on the following nodes: {1}. |
Description: Command Audit is turned off on some nodes. This configuration leads to lags in the refresh of data that is displayed in the GUI. | ||||
Cause: Command Audit is turned off on some nodes. | ||||
User Action: Change the cluster configuration option commandAudit to on by using the mmchconfig commandAudit=on -N [node name]) command, or syslogonly by using the mmchconfig commandAudit=syslogonly -N [node name] command for the affected nodes. This way the GUI refreshes the data that it displays automatically when IBM Storage Scale commands are run by using the CLI on other nodes in the cluster. | ||||
gui_config_command_audit_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: Command Audit is turned on at the cluster level. |
Description: Command Audit is turned on at the cluster level. This way the GUI refreshes the data that it displays automatically when IBM Storage Scale commands are run by using the CLI on other nodes in the cluster. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_config_sudoers_error | STATE_CHANGE FAILED |
ERROR | no | Message: There is a problem with the /etc/sudoers configuration. The secure_path of the IBM Storage Scale management user 'scalemgmt' is not correct. Current value: {0} / Expected value: {1}. |
Description: There is a problem with the /etc/sudoers configuration. | ||||
Cause: N/A | ||||
User Action: Ensure that #includedir /etc/sudoers.d directive is set in /etc/sudoers, so the sudoers' configuration drop-in file for the IBM Storage Scale management user 'scalemgmt' (which the GUI process uses) is loaded from /etc/sudoers.d/scalemgmt_sudoers. Also, make sure that the #includedir directive is the last line in the /etc/sudoers configuration file. | ||||
gui_config_sudoers_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The /etc/sudoers configuration is correct. |
Description: The /etc/sudoers configuration is correct. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_database_cleared_cluster_change | INFO | WARNING | no | Message: The cluster ID has changed. |
Description: The cluster ID stored in the database no longer matches the installed cluster ID. | ||||
Cause: A new GPFS cluster has been installed. | ||||
User Action: Events that are marked as read are now displayed as unread. Mark all notices as read if they are no longer valid after the cluster change. | ||||
gui_database_cleared_downgrade | INFO | WARNING | no | Message: The GUI version read from the database({0}) is later than the GUI code version({1}). |
Description: The GUI might have been moved to an older version. | ||||
Cause: The GUI version that is stored in the database is greater than the GUI code version. | ||||
User Action: Events that are marked as read are now displayed as unread. Mark all notices as read if they are no longer valid after the GUI is moved to an older version. | ||||
gui_database_dropped | INFO | WARNING | no | Message: The database version ({0}) mismatches the PostgreSQL version({1}). |
Description: The PostgreSQL internal storage format might have changed following a major version upgrade. | ||||
Cause: There is a mismatch between the database version and the PostgreSQL server version. | ||||
User Action: Events that are marked as read are now displayed as unread. Mark all notices as read if they are no longer valid after the upgrade. | ||||
gui_db_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI reported correct connection to postgres database in the cluster. |
Description: The connection to postgres database works properly. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_db_warn | STATE_CHANGE DEGRADED |
WARNING | no | Message: The GUI reported incorrect connection to postgres database. |
Description: The connection to postgres database cannot be established. | ||||
Cause: The GUI reported incorrect connection to postgres database. | ||||
User Action: Check if postgres container works properly in the GUI pod. | ||||
gui_down | STATE_CHANGE FAILED |
ERROR | no | Message: The GUI service should be {0}, but it is {1}. If there are no other GUI nodes up and running, then no snapshots are created and email notifications are not sent anymore. |
Description: The GUI service is down. | ||||
Cause: The GUI service is not running on this node, although it has the 'GUI_MGMT_SERVER_NODE' node class. | ||||
User Action: Restart the GUI service or change the node class for this node. | ||||
gui_email_server_reachable | STATE_CHANGE HEALTHY |
INFO | no | Message: The email server {0} is reachable. |
Description: The specified email server is reachable. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_email_server_unreachable | STATE_CHANGE DEGRADED |
ERROR | no | Message: The email server {0} is unreachable {1}. |
Description: The specified email server does not respond to any messages. | ||||
Cause: The configuration or firewall setting is wrong. | ||||
User Action: Check the email server configuration (hostname, port, username, and password). Ensure that the email server is up and running and no firewall is blocking the access. | ||||
gui_external_authentication_failed | INFO | ERROR | no | Message: The GUI cannot connect to the external LDAP or AD server: {0}. |
Description: The GUI cannot connect to one or more of the specified LDAP or AD servers. | ||||
Cause: The LDAP or AD server is not reachable because it is not running or due to a network issue. | ||||
User Action: Verify that the configured LDAP or AD servers are up and running and reachable from the GUI node. | ||||
gui_login_attempt_failed | INFO | WARNING | no | Message: A login attempt failed for the user {0} from the source IP address {1}. |
Description: A login attempt for the specified user failed. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_mount_allowed_on_gui_node | STATE_CHANGE HEALTHY |
INFO | no | Message: Mount operation is allowed for all file systems on the GUI node. |
Description: Mount operation is allowed for all file systems on the GUI node. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_mount_prevented_on_gui_node | STATE_CHANGE DEGRADED |
WARNING | no | Message: Mount operation is prevented for {1} file systems on the GUI node {0}. |
Description: Mount operation for specific file systems is prevented on the GUI node. | ||||
Cause: Mount operation is prevented for specific file systems on the GUI node. | ||||
User Action: Run the fix procedure or go to the file system panel, and allow mount operation for mentioned file systems on the GUI node. | ||||
gui_node_update_failure | STATE_CHANGE FAILED |
ERROR | no | Message: GUI node class cannot be updated. |
Description: The node class update got failed. | ||||
Cause: The node class update got failed. | ||||
User Action: N/A | ||||
gui_node_update_successful | STATE_CHANGE HEALTHY |
INFO | no | Message: GUI node class got updated successfully. |
Description: The node class got updated successfully. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_out_of_memory | INFO | ERROR | no | Message: The GUI reported an internal out-of-memory state. Restart the GUI. |
Description: A GUI internal process ran into an out-of-memory state, which might impact GUI functions partially or fully. | ||||
Cause: The Java virtual machine of the GUI reported an internal out-of-memory state. | ||||
User Action: Restart the GUI or recreate the liberty container. | ||||
gui_pmcollector_connection_failed | STATE_CHANGE DEGRADED |
ERROR | no | Message: The GUI cannot connect to the pmcollector that is running on {0} using port {1}. |
Description: The GUI checks the connection to the pmcollector. | ||||
Cause: The GUI cannot connect to the pmcollector. | ||||
User Action: Check whether the GUI node has the gpfs.gss.pmcollector package installed and is specified for the 'colCandidates' attribute in the mmperfmon config show command and check whether the pmcollector service is running. Check for event pmcollector_up with mmhealth node show perfmon -v command. Check pmcollector has many expired keys with mmperfmon query --list=expiredKeys command and delete expired keys if their perfmon data is no longer needed. | ||||
gui_pmcollector_connection_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI can connect to the pmcollector that is running on {0} using port {1}. |
Description: The GUI checks the connection to the pmcollector. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_pmsensors_connection_failed | STATE_CHANGE DEGRADED |
ERROR | no | Message: The performance monitoring sensor service 'pmsensors' on node {0} is not sending any data. |
Description: The GUI checks whether data can be retrieved from the pmcollector service for this node. | ||||
Cause: The performance monitoring sensor service 'pmsensors' is not sending any data. The service might be down or the time of the node is more than 15 minutes away from the time on the node hosting the performance monitoring collector service 'pmcollector'. The connection probe uses data from the CPU sensor to check whether nodes send data. Also, check whether the CPU sensor was disabled. If the CPU sensor was disabled, then enable the sensor with a period of 10 or smaller value. | ||||
User Action: Check by using the systemctl status pmsensors command. If pmsensors service is 'inactive', then run the systemctl start pmsensors command. Check whether the CPU sensor was disabled. If it was disabled, then run the mmperfmon config update CPU.period=1 CPU.restrict=all command. Check the time of the GUI node and all perfmon pmsensors nodes is in sync. | ||||
gui_pmsensors_connection_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The state of performance monitoring sensor service 'pmsensor' on node {0} is OK. |
Description: The GUI checks whether data can be retrieved from the pmcollector service for this node. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_quorum_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI reported correct quorum in the cluster. |
Description: Quorum is reached in the cluster. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_quorum_warn | STATE_CHANGE DEGRADED |
WARNING | no | Message: The GUI reported quorum loss in the cluster. |
Description: The GUI reported quorum loss in the cluster. | ||||
Cause: The GUI reported quorum loss in the cluster. | ||||
User Action: Check if quorum is correct in your cluster. | ||||
gui_reachable_node | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI can reach the node {0}. |
Description: The GUI checks the reachability of all nodes. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_refresh_task_failed | STATE_CHANGE DEGRADED |
WARNING | no | Message: The following GUI refresh task(s) failed: {0}. |
Description: One or more GUI refresh tasks failed, which means that data in the GUI is outdated. | ||||
Cause: There can be several reasons. | ||||
User Action: Check whether there is additional information available by using the '/usr/lpp/mmfs/gui/cli/lstasklog [taskname]' command. Then, run the specified task manually on the CLI by using the '/usr/lpp/mmfs/gui/cli/runtask [taskname] --debug' command. Check the GUI logs under '/var/log/cnlog/mgtsrv' and contact IBM Support if this error persists or occurs more often. | ||||
gui_refresh_task_successful | STATE_CHANGE HEALTHY |
INFO | no | Message: All GUI refresh tasks are running fine. |
Description: All GUI refresh tasks are running fine. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_response_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI is responsive to the test query. |
Description: The GUI is responsive to the test query. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_response_warn | STATE_CHANGE DEGRADED |
WARNING | no | Message: The GUI is unresponsive to the test query. |
Description: The GUI is unresponsive to the test query. | ||||
Cause: The GUI did not respond with the expected data for the test query (debug platform). | ||||
User Action: Restart the GUI by using the systemctl restart gpfsgui command or wait for the liberty container to be re-created. | ||||
gui_snap_create_failed_fs | INFO | ERROR | no | Message: A snapshot creation invoked by rule {1} failed on file system {0}. |
Description: The snapshot was not created according to the specified rule. | ||||
Cause: A snapshot creation invoked by a rule fails. | ||||
User Action: Try to create the snapshot again manually. | ||||
gui_snap_create_failed_fset | INFO | ERROR | no | Message: A snapshot creation that is invoked by rule {1} failed on file system {2}, fileset {0}. |
Description: The snapshot was not created according to the specified rule. | ||||
Cause: A snapshot creation that is invoked by a rule fails. | ||||
User Action: Try to create the snapshot again manually. | ||||
gui_snap_delete_failed_fs | INFO | ERROR | no | Message: A snapshot deletion that is invoked by rule {1} failed on file system {0}. |
Description: The snapshot was not deleted according to the specified rule. | ||||
Cause: A snapshot deletion that is invoked by a rule fails. | ||||
User Action: Try to manually delete the snapshot. | ||||
gui_snap_delete_failed_fset | INFO | ERROR | no | Message: A snapshot deletion that is invoked by rule {1} failed on file system {2}, fileset {0}. |
Description: The snapshot was not deleted according to the specified rule. | ||||
Cause: A snapshot deletion that is invoked by a rule fails. | ||||
User Action: Try to manually delete the snapshot. | ||||
gui_snap_rule_ops_exceeded | INFO | WARNING | no | Message: The number of pending operations exceeds {1} operations for rule {2}. |
Description: The number of pending operations for a rule exceed a specified value. | ||||
Cause: The number of pending operations for a rule exceed a specified value. | ||||
User Action: N/A | ||||
gui_snap_running | INFO | WARNING | no | Message: Operations for rule {1} are still running at the start of the next management of rule {1}. |
Description: Operations for a rule are still running at the start of the next management of that rule. | ||||
Cause: Operations for a rule are still running. | ||||
User Action: N/A | ||||
gui_snap_time_limit_exceeded_fs | INFO | WARNING | no | Message: A snapshot operation exceeds {1} minutes for rule {2} on file system {0}. |
Description: The snapshot operation that is resulting from the rule is exceeding the established time limit. | ||||
Cause: A snapshot operation exceeds a specified number of minutes. | ||||
User Action: N/A | ||||
gui_snap_time_limit_exceeded_fset | INFO | WARNING | no | Message: A snapshot operation exceeds {1} minutes for rule {2} on file system {3}, fileset {0}. |
Description: The snapshot operation that is resulting from the rule is exceeding the established time limit. | ||||
Cause: A snapshot operation exceeds a specified number of minutes. | ||||
User Action: N/A | ||||
gui_snap_total_ops_exceeded | INFO | WARNING | no | Message: The total number of pending operations exceeds {1} operations. |
Description: The total number of pending operations exceed a specified value. | ||||
Cause: The total number of pending operations exceed a specified value. | ||||
User Action: N/A | ||||
gui_ssh_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The GUI reported correct ssh connection in the cluster. |
Description: The ssh connection works properly. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_ssh_warn | STATE_CHANGE DEGRADED |
WARNING | no | Message: The GUI reported incorrect ssh connection. |
Description: The connection over ssh cannot be established. | ||||
Cause: The GUI reported incorrect ssh connection. | ||||
User Action: Check if ssh connection works properly in your cluster. | ||||
gui_ssl_certificate_expired | STATE_CHANGE FAILED |
ERROR | no | Message: The SSL certificate that is used by the GUI expired. Expiration date was {0}. |
Description: SSL certificate expired. | ||||
Cause: The SSL certificate that is used by the GUI expired. | ||||
User Action: On the CLI, run the '/usr/lpp/mmfs/gui/cli/rmhttpskeystore' command to return to the default certificate. On the GUI, go to 'Service' and select 'GUI' to create or upload a new certificate. | ||||
gui_ssl_certificate_is_about_to_expire | STATE_CHANGE DEGRADED |
WARNING | no | Message: The SSL certificate that is used by the GUI is about to expire. Expiration date is {0}. |
Description: SSL certificate is about to expire. | ||||
Cause: The SSL certificate that is used by the GUI is about to expire. | ||||
User Action: Go to the Service panel and select 'GUI'. On the 'Nodes' tab, select an option to create a new certificate request, self-signed certificate, or upload your own certificate. | ||||
gui_ssl_certificate_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The SSL certificate that is used by the GUI is valid. Expiration date is {0}. |
Description: GUI SSL certificates are valid. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_unreachable_node | STATE_CHANGE DEGRADED |
ERROR | no | Message: The GUI cannot reach the node {0}. |
Description: The GUI checks the reachability of all nodes. | ||||
Cause: The specified node cannot be reached by the GUI node. | ||||
User Action: Check your firewall or network setup, and whether the specified node is up and running. | ||||
gui_up | STATE_CHANGE HEALTHY |
INFO | no | Message: The status of the GUI service is {0} as expected. |
Description: The GUI service is running. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
gui_warn | INFO | INFO | no | Message: The GUI service returned an unknown result. |
Description: The GUI service returned an unknown result. | ||||
Cause: The service or systemctl command returned unknown results about the gpfsgui service. | ||||
User Action: Check whether the gpfsgui is in the expected status in the service or systemctl command. Also, check whether there is no gpfsgui service, although the node has the 'GUI_MGMT_SERVER_NODE' node class. For more information, see the IBM Documentation. Otherwise, monitor the issue if this warning appears more often. | ||||
host_disk_filled | STATE_CHANGE DEGRADED |
WARNING | no | Message: A local file system on node {0} reached a warning level {1}. |
Description: The GUI checks the fill level of the local file systems. | ||||
Cause: The local file systems reached a warning level. | ||||
User Action: Delete data on the local disk. | ||||
host_disk_full | STATE_CHANGE DEGRADED |
ERROR | no | Message: A local file system on node {0} reached a nearly exhausted level {1}. |
Description: The GUI checks the fill level of the local file systems. | ||||
Cause: The local file systems reached a nearly exhausted level. | ||||
User Action: Delete data on the local disk. | ||||
host_disk_normal | STATE_CHANGE HEALTHY |
INFO | no | Message: The local file systems on node {0} reached a normal level. |
Description: The GUI checks the fill level of the local file systems. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
host_disk_unknown | STATE_CHANGE DEGRADED |
WARNING | no | Message: The fill level of local file systems on node {0} is unknown. |
Description: The GUI checks the fill level of the local file systems. | ||||
Cause: Cannot determine fill state of the local file systems. | ||||
User Action: N/A | ||||
sudo_admin_not_configured | STATE_CHANGE FAILED |
ERROR | no | Message: Sudo wrappers are enabled on the cluster '{0}', but the GUI is not configured to use Sudo wrappers. |
Description: Sudo wrappers are enabled on the cluster, but the value for GPFS_ADMIN in '/usr/lpp/mmfs/gui/conf/gpfsgui.properties' was either not set or is still set to root. The value of 'GPFS_ADMIN' is set to the username for which sudo wrappers were configured on the cluster. | ||||
Cause: N/A | ||||
User Action: Ensure that sudo wrappers were correctly configured for a user that is available on the GUI node and all other nodes of the cluster. This username is set as the value of the 'GPFS_ADMIN' option in the '/usr/lpp/mmfs/gui/conf/gpfsgui.properties' file. After the restart, the GUI starts by using the systemctl restart gpfsgui command. | ||||
sudo_admin_not_exist | STATE_CHANGE FAILED |
ERROR | no | Message: Sudo wrappers are enabled on the cluster '{0}', but there is a misconfiguration that is regarding the user '{1}' that was set as 'GPFS_ADMIN' in the GUI properties file. |
Description: Sudo wrappers are enabled on the cluster, but the username that was set as GPFS_ADMIN in the GUI properties file at '/usr/lpp/mmfs/gui/conf/gpfsgui.properties' does not exist on the GUI node. | ||||
Cause: N/A | ||||
User Action: Ensure that sudo wrappers were correctly configured for a user that is available on the GUI node and all other nodes of the cluster. This username is set as the value of the 'GPFS_ADMIN' option in the '/usr/lpp/mmfs/gui/conf/gpfsgui.properties'. After that restart, the GUI starts by using the systemctl restart gpfsgui command. | ||||
sudo_admin_set_but_disabled | STATE_CHANGE DEGRADED |
WARNING | no | Message: Sudo wrappers are not enabled on the cluster '{0}', but 'GPFS_ADMIN' was set to a non-root user. |
Description: Sudo wrappers are not enabled on the cluster, but the value for 'GPFS_ADMIN' in the '/usr/lpp/mmfs/gui/conf/gpfsgui.properties' was set to a non-root user. The value of 'GPFS_ADMIN' is set to 'root' when sudo wrappers are not enabled on the cluster. | ||||
Cause: N/A | ||||
User Action: Set 'GPFS_ADMIN' in '/usr/lpp/mmfs/gui/conf/gpfsgui.properties' to 'root'. After that restart, the GUI starts by using the systemctl restart gpfsgui command. | ||||
sudo_connect_error | STATE_CHANGE FAILED |
ERROR | no | Message: Sudo wrappers are enabled on the cluster '{0}', but the GUI cannot connect to other nodes with the username '{1}' that was defined as 'GPFS_ADMIN' in the GUI properties file. |
Description: When sudo wrappers are configured and enabled on a cluster, the GUI does not run commands as root, but as the user for which sudo wrappers were configured. This user is set as 'GPFS_ADMIN' in the GUI properties file at '/usr/lpp/mmfs/gui/conf/gpfsgui.properties'. | ||||
Cause: N/A | ||||
User Action: Ensure that sudo wrappers were correctly configured for a user that is available on the GUI node and all other nodes of the cluster. This username is set as the value of the 'GPFS_ADMIN' option in the '/usr/lpp/mmfs/gui/conf/gpfsgui.properties'. After that restart, the GUI starts by using the systemctl restart gpfsgui command. | ||||
sudo_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: Sudo wrappers were enabled on the cluster and the GUI configuration for the cluster '{0}' is correct. |
Description: No problems were found with the current GUI and cluster configurations. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
time_in_sync | STATE_CHANGE HEALTHY |
INFO | no | Message: The time on node {0} is in sync with the cluster median. |
Description: The GUI checks the time on all nodes. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
time_not_in_sync | STATE_CHANGE DEGRADED |
ERROR | no | Message: The time on node {0} is not in sync with the cluster median. |
Description: The GUI checks the time on all nodes. | ||||
Cause: The time on the specified node is not in sync with the cluster median. | ||||
User Action: Synchronize the time on the specified node. | ||||
time_sync_unknown | STATE_CHANGE DEGRADED |
WARNING | no | Message: The time on node {0} cannot be determined. |
Description: The GUI checks the time on all nodes. | ||||
Cause: The time on the specified node cannot be determined. | ||||
User Action: Check whether the node is reachable from the GUI. | ||||
xcat_nodelist_missing | STATE_CHANGE DEGRADED |
ERROR | no | Message: The node {0} is unknown by xCAT. |
Description: The GUI checks whether xCAT can manage the node. | ||||
Cause: The xCAT does not know about the node. | ||||
User Action: Add the node to xCAT. Ensure that the hostname that is used in xCAT matches the hostname that is known by the node itself. | ||||
xcat_nodelist_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The node {0} is known to the xCAT. |
Description: The GUI checks whether xCAT can manage the node. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
xcat_nodelist_unknown | STATE_CHANGE DEGRADED |
WARNING | no | Message: State of the node {0} in xCAT is unknown. |
Description: The GUI checks whether xCAT can manage the node. | ||||
Cause: The state of the node within xCAT cannot be determined. | ||||
User Action: N/A | ||||
xcat_state_error | STATE_CHANGE HEALTHY |
INFO | no | Message: The xCAT on node {1} failed to operate properly on cluster {0}. |
Description: The GUI checks the xCAT state. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
xcat_state_invalid_version | STATE_CHANGE DEGRADED |
WARNING | no | Message: The xCAT service has not the recommended version ({1} actual/recommended). |
Description: The GUI checks the xCAT state. | ||||
Cause: The reported version of xCAT is not compliant with the recommendation. | ||||
User Action: Install the recommended xCAT version. | ||||
xcat_state_no_connection | STATE_CHANGE DEGRADED |
ERROR | no | Message: Unable to connect to xCAT node {1} on cluster {0}. |
Description: The GUI checks the xCAT state. | ||||
Cause: Cannot connect to the node specified as xCAT host. | ||||
User Action: Check whether the IP address is correct and ensure that root has key-based SSH set up to the xCAT node. | ||||
xcat_state_ok | STATE_CHANGE HEALTHY |
INFO | no | Message: The availability of xCAT on cluster {0} is OK. |
Description: The GUI checks the xCAT state. | ||||
Cause: N/A | ||||
User Action: N/A | ||||
xcat_state_unconfigured | STATE_CHANGE DEGRADED |
WARNING | no | Message: The xCAT host is not configured on cluster {0}. |
Description: The GUI checks the xCAT state. | ||||
Cause: The host where xCAT is located is not specified. | ||||
User Action: Specify the hostname or IP where xCAT is located. | ||||
xcat_state_unknown | STATE_CHANGE DEGRADED |
WARNING | no | Message: Availability of xCAT on cluster {0} is unknown. |
Description: The GUI checks the xCAT state. | ||||
Cause: The availability and state of xCAT cannot be determined. | ||||
User Action: N/A |