IBM PowerHA SystemMirror cluster migration
IBM Power Systems high availability
The purpose of this article is to provide a step-by-step guide for migrating an existing PowerHA cluster (at PowerHA 6.1.0) to PowerHA SystemMirror 7.1.2. This article helps in understanding how to plan for and accomplish a successful migration. It provides an overview of cluster variants in the PowerHA 7.1.2 and provides an overview of the PowerHA migration process, various migration methodologies, and the requirements for the migration process. I will discuss some migration limitations and prerequisites along with the planning process and also introduce the clmigcheck utility, which checks the current cluster configuration for any unsupported element in the cluster as well as collecting additional information required for the migration. The actual migration steps are presented in detail for use by customers to seamlessly migrate their two-node PowerHA 6.1 (single-site clusters) to PowerHA 7.1.2.
Cluster Aware AIX
The Cluster Aware AIX (CAA) is a built-in clustering capability of the IBM AIX® operating system. Using CAA, administrators can create a cluster of AIX nodes and take advantage of the capabilities of cluster. The CAA has many capabilities, and some of them are listed below:
- Cluster-wide event management
- Communication and storage events such as node up and down, network adapter up and down, network address changes, and disk up and down
- Predefined and user-defined events
- Cluster-wide storage naming service
- Cluster-wide command distribution
- Commands and application programming interfaces (APIs) to create clusters across a set of AIX systems: Kernel-based heartbeats and messages provide a robust cluster infrastructure and by default, use multichannel communication between nodes using the network and storage area network (SAN) physical links
Cluster repository disk
A cluster repository disk is a storage device shared across all the cluster nodes. This disk is used as a central repository. You can have only one cluster repository disk. In PowerHA 7.1.2, you can define a backup repository disk, which can be used in case the primary repository disk fails. For a linked cluster (true XD cluster), each PowerHA site will have its own repository. The repository disk cannot be mirrored using AIX Logical Volume Manager (LVM), and therefore, plan to have Redundant Array of Independent Disks (RAID) mirroring for the disk. The minimum space required for a cluster repository disk is 1 GB. Refer to the PowerHA SystemMirror Admin Guide for information on how to define backup repository disk.
Multicast IP addresses
CAA uses multicast addresses for cluster communication between the nodes in the cluster. It is mandatory to have multicast enabled in your cluster network infrastructure.
Differences between clcomdES and clcomd
Starting with AIX 6.1 TL6 and AIX 7.1, the cluster communication daemon has been integrated into AIX as part of the CAA infrastructure. Some of the differences between the clcomdES subsystem (used by previous versions of PowerHA) and the new clcomd daemon of CAA and PowerHA 7.1 and later are provided in this section.
- Install: The clcomdES subsystem is part of the PowerHA SystemMirror install media. Whereas, clcomd is part of AIX installed with Base AIX Enterprise Edition (delivered with the bos.cluster.rte file set).
- Name: The subsystem name of traditional cluster communication daemon is clcomdES; the new subsystem name is clcomd.
- Run ability: The clcomdES daemon is always running on the nodes installed with PowerHA SystemMirror (run from /etc/inittab). The clcomd daemon is always running on nodes even if PowerHA SystemMirror is not installed (this is run from /etc/inittab as well).
- Port: The clcomdES subsystem uses port 6191 (/etc/services). The clcomd daemon uses port 16191 (/etc/services); also uses the clcomdES port 6191 if PowerHA SystemMirror migration is detected.
- Cluster definition: The clcomdES subsystem uses the
/usr/es/sbin/cluster/etc/rhosts file for initial cluster
definition. It can be populated with IP addresses for all available adapters on
the node. Whereas, clcomd uses /etc/cluster/rhosts for initial
cluster definition. The file
/etc/cluster/rhosts should be populated
with IP addresses, only one per line, of members in this file. Then, refresh
clcomd using the
refresh –s clcomdcommand.
- Definition query: The clcomdES subsystem gets the cluster definition from PowerHA SystemMirror configuration data, whereas clcomd queries the definition of the cluster using kernel API (making use of the CAA infrastructure).
Differences between PowerHA 6.1 and PowerHA 7.1 and later
With the introduction of the CAA feature in AIX 7.1 and AIX 6.1 TL6, PowerHA SystemMirror has undergone a lot of architectural changes. Due to architectural changes in PowerHA 7.1 and later with the advent of CAA, PowerHA 7.1 and later expects the communication path for cluster node be set to the IP address mapped to the host name. Some of the differences between PowerHA 6.1 and PowerHA 7.1 and later are:
- PowerHA 7.1 and later releases are based upon CAA where monitoring and event management is built into the AIX kernel providing robust foundation not prone to job scheduling. In the previous releases, PowerHA monitored soft and hard errors within the cluster from various event sources using Reliable Scalable Cluster Technology (RSCT).
- In PowerHA 6.1 and lower releases, the main communication path goes from PowerHA to group services (grpsvcs subsystem of RSCT) and then to topology services (topsvcs subsystem of RSCT) and back. In PowerHA 7.1 and later releases, the main communication path goes from PowerHA to group services (cthags) and then to CAA.
- With PowerHA 7.1, event management is handled by using a new pseudo file system architecture called Autonomic Health Advisor File System (AHAFS). This is used by CAA as its monitoring framework.
- PowerHA 7.1 uses the cluster repository disk, Fibre Channel (FC)/SAN adapters and multicasting for heartbeating. Heartbeat is performed by sending and receiving special gossip packets across the network using the multicast protocol. The gossip packets are always replied to by other nodes. In older releases of PowerHA, IP and non-IP networks participated in heartbeats and detection or diagnosis of network, node, or network adapter failures. These heartbeat packets were never acknowledged.
- PowerHA 7.1 and later releases use a special gossip protocol over the multicast address to determine node information and implement scalable reliable multicast. Older releases use traditional cluster communication daemon (clcomdES subsystem) which gets information from PowerHA Object Data Manager (ODM) and uses the heartbeat mechanism provided by RSCT for node information processing.
- PowerHA 7.1 and later releases, introduced Systems Events, which are handled by the clevmgrdES subsystem. The root volume group (rootvg) system event allows the monitoring of loss of access to the rootvg. Loss of access to rootvg results in log entry in the system error log and system reboot. Older releases of PowerHA do not handle rootvg failures.
PowerHA 7.1.2 cluster variants
PowerHA SystemMirror 7.1.2 allows customers to configure three different styles of clusters namely local, stretched, and linked clusters.
Local cluster – It is a simple, multinode, single-site or local cluster configured using node or logical partitions (LPARs) within a single data center. This is the most typical cluster configuration providing for local PowerHA cluster fallover. Local fallover provides a faster transition onto another machine than a fallover going to a geographically dispersed site. Local clusters can benefit from advanced functions such as IBM PowerVM Live Partition Mobility (LPM) between machines within the same site. This combination of IBM PowerVM functions and IBM PowerHA SystemMirror clustering is useful for helping to avoid any service interruption for a planned maintenance event while protecting the environment in the event of unforeseen outage.
Stretched cluster – The term denotes a cluster that has sites defined within the same geographic location. This provides for a campus style disaster recovery and high availability cluster with cluster nodes separated by a shorter distance. The sites can be near enough to have shared logical unit numbers (LUNs) in the same SAN. The key aspect about stretched cluster is that it uses a shared repository disk. Stretched clusters can support cross-site LVM mirroring, IBM HyperSwap®, and Geographic Logical Volume Manager (GLVM). Extended distance sites with IP-only connectivity are not possible with this configuration.
Figure 1. Example of a stretched cluster
A stretched cluster configuration can also be used with PowerHA 7.1.2 Standard Edition with the use of LVM cross-site Mirroring. The stretched cluster is capable of using all three levels of cluster communication (TCP/IP, SAN heartbeat and repository disk). The distance can be up to 15 km, with direct SAN links and up to 120 km with dense wavelength division multiplexing (DWDM) or coarse wavelength division multiplexing (CWDM) or other SAN extenders. This provides for synchronous replication or mirroring.
Linked cluster – The term denotes a cluster that has sites defined across geographic locations allowing configuration of a traditional extended distance cluster between two sites, for example Brisbane and Singapore. The key aspect of a linked cluster that makes it different from extended distance clusters in previous versions is the use of SIRCOL in CAA. This means that each site has its own CAA repository disk, which is replicated automatically between sites by CAA. Linked cluster sites communicate with each other using unicast and not multicast as it is the case with stretched cluster or normal cluster. However, local sites internally use multicast, and therefore, multicast still must be enabled in the network at each site.
Figure 2. Illustration of a linked cluster
All the interfaces are defined in this type of configuration as CAA gateway addresses. CAA maintains the repository information automatically across sites through unicast address communication.
Unlike previous PowerHA SystemMirror migration methods, there will be some cases where migration will have to be done manually by the customer resulting in a complete cluster outage. These conditions can be detected at the time we run /usr/sbin/clmigcheck. There are three supported migration paths for PowerHA SystemMirror 6.1 migration to PowerHA SystemMirror 7.1.2. Each one requires an AIX upgrade, migration to AIX 6.1 TL8 SP1 or later, or AIX 7.1 TL2 SP1. Migration to PowerHA SystemMirror 7.1.2 is a two-phase process.
- Phase I: AIX migration or upgrade based on the current AIX level.
- Phase II: PowerHA SystemMirror migration
Refer to the AIX information center for steps on how to migrate or upgrade AIX.
PowerHA SystemMirror provides the following three different migration options.
- Offline migration: As the name suggests, this type of migration involves bringing down the entire PowerHA cluster, installing PowerHA SystemMirror 7.1.2, and restarting cluster services for one node at a time.
- Rolling migration: During rolling migration, the workload is moved from the node where it is currently running to another node in the system. This is followed by the installation of PowerHA 7.1.2 and the starting of cluster services. These steps are followed on all the remaining nodes.
- Snapshot migration: This really is not a migration at all. Customers would remove the previous version of PowerHA SystemMirror and install the newer version of PowerHA SystemMirror 7.1.2. Customer would then use the PowerHA SystemMirror 7.1.2 configuration interface, either the Director GUI, System Management Interface Tool (SMIT), or command line to install the same configuration as they previously had, that is, restoring from cluster snapshot.
Before you start migrating the cluster nodes, ensure that the following tasks are completed:
- Back up all the application and system data.
- Create a back out or reversion plan. A back out plan allows for easy restoration of cluster and AIX configuration in case migration runs into some problem. System backup should be created using the mksysb and savevg utilities.
- Ensure that the Communication Path to Node option in the PowerHA cluster nodes is set to the IP address mapping to the hostname.
- Save the existing cluster configuration. Also, save any user provided scripts, most commonly custom events, pre and post event scripts, notification scripts, and application controller scripts.
Some migration requirements are as follows:
- All cluster nodes have one shared disk that will be used for cluster repository,
having at least 1 GB size. The list of supported FC and SAS adapters for
connection to the repository disk:
- 4 GB Single-Port Fibre Channel PCI-X 2.0 DDR Adapter (FC 1905; CCIN 1910)
- 4 GB Single-Port Fibre Channel PCI-X 2.0 DDR Adapter (FC 5758; CCIN 280D)
- 4 GB Single-Port Fibre Channel PCI-X Adapter (FC 5773; CCIN 5773)
- 4 GB Dual-Port Fibre Channel PCI-X Adapter (FC 5774; CCIN 5774)
- 4 Gb Dual-Port Fibre Channel PCI-X 2.0 DDR Adapter (FC 1910; CCIN 1910)
- 4 Gb Dual-Port Fibre Channel PCI-X 2.0 DDR Adapter (FC 5759; CCIN 5759)
- 8 Gb PCI Express Dual Port Fibre Channel Adapter (FC 5735; CCIN 577D)
- 8 Gb PCI Express Dual Port Fibre Channel Adapter 1Xe Blade (FC 2B3A; CCIN 2607)
- 3 Gb Dual-Port SAS Adapter PCI-X DDR External (FC 5900 and 5912; CCIN 572A)
For the most current list of supported storage adapters, refer to the IBM PowerHA SystemMirror for AIX web page.
- Ensure that the current network infrastructure supports multicast. Enable multicast traffic on all network switches connected to all cluster nodes.
- Ensure that the /etc/cluster/rhosts file is properly filled with hostnames or IP addresses of all cluster nodes (IP addresses mapping to the host name), else cluster communication will fail and migration will not take place.
- Ensure that all cluster nodes have the requisite version of AIX installed. Refer
to the following table.
PowerHA 7.1.0 AIX 6.1 TL6 SP1 or AIX 7.1 TL0 SP1 PowerHA 7.1.1 AIX 6.1 TL7 SP2 or AIX 7.1 TL1 SP2 PowerHA 7.1.2 AIX 6.1 TL8 SP1 or AIX 7.1 TL2 SP1
- Ensure that Virtual I/O Server (VIOS) 220.127.116.11-FP24-SP01 or later is installed.
- The following additional file sets are required:
- devices.common.IBM.storfwork (required for SAN heartbeat)
- RSCT version:
- rsct.core.rmc 18.104.22.168
- rsct.basic 22.214.171.124
- rsct.compat.basic.hacmp 126.96.36.199
- rsct.compat.clients.hacmp 188.8.131.52
There are certain limitations in migrating to PowerHA 7.1.2 because of the structural changes. The limitations are listed below:
- Not all configurations can be migrated.
- Configurations with FDDI, ATM, X.25 or Token Ring cannot be migrated and must be removed before migration.
- Configurations with IP Address Takeover (IPAT) using replacement or Hardware Address Takeover (HWAT) cannot be migrated, and must be removed from configuration.
- Configurations with heartbeat over IP aliasing must be removed before migration.
- Non-IP networking is accomplished differently.
PowerHA 7.1.2 (and the underlying CAA) use multicast, FC/SAN and the cluster repository disk for heartbeating. Traditional non-IP networks such as rs232, diskhb, mndhb, tmscsi, tmssa are not supported. These will be removed during migration.
The clmigcheck utility is part of base AIX, included with AIX 6.1 TL6 or later. It is an interactive tool that verifies the current cluster configuration, checks for unsupported elements, and collects additional information required for migration. You must run this command on all cluster nodes, one node at a time, before installing PowerHA 7.1.2. The initial screen is as follows:
----------[PowerHA System Mirror Migration Check] ------------- Please select one of the following options: 1 = Check ODM configuration. 2 = Check snapshot configuration. 3 = Enter repository disk and multicast IP addresses. Select one of the above, "x" to exit or "h" for help:
Note that at any prompt, you can type h for help about that data entry prompt.
- Option 1 checks SystemMirror configuration data (/etc/es/objrepos) and provides errors and warnings if there are any elements in the configuration that must be removed manually. In that case, the flagged elements must be removed, cluster configuration verified and synchronized, and this command must be re-run until the SystemMirror configuration data check completes without errors.
- Option 2 checks a snapshot (present in /usr/es/sbin/cluster/snapshots) and provides error information if there are any elements in the configuration that will not migrate. Because PowerHA SystemMirror provides no tools to edit a snapshot, any errors checking this snapshot means that it cannot be used for migration. In this case, the customer might have to apply the snapshot on the back-level PowerHA SystemMirror and update the configuration manually. Save the new snapshot and start the procedure all over again.
- Option 3 queries the customer for additional configuration needed and saves it in a file in /var on every node in the cluster. When option 3 is selected from the main screen, you will be prompted for repository disk and multicast dotted decimal IP addresses. This data will be stored in a file (/var/clmigcheck/clmigcheck.txt) on every node in the cluster. When PowerHA SystemMirror 7.1.2 is installed, this file is read and the SystemMirror configuration data is populated. The customer must use either option 1 or option 2 successfully before running option 3, which collects and stores configuration data.
/usr/sbin/clmigcheck command is run on the last
node of the cluster before installing PowerHA SystemMirror 7.1.2, the CAA
infrastructure will be started. This can be verified by running the
/usr/sbin/lscluster –m command.
FC/SAN based heartbeat mechanism
The cluster communication in PowerHA 7.1 and later (and CAA) is achieved by communicating over multiple redundant paths. This includes the important process of sending and processing the cluster heartbeats by each participating node. The following redundant paths provide robust clustering foundation:
- TCP/IP (basically using multicast address)
- Optional SAN or FC adapters
- Repository disk
SAN-based path is a redundant, high-speed path of communication established between
the hosts by using the SAN fabric that exists in any data center between hosts.
Discovery-based configuration reduces the burden of configuring the links. PowerHA
7.1.2 supports SAN-based heartbeat within a site. It is not mandatory to set up FC
or SAN-based heartbeat path, if the configured SANComm (sfwcomm as seen in
lscluster –i output) provides additional heartbeat path for
The SAN heartbeat infrastructure can be accomplished in several ways:
- Using real adapters on the cluster nodes and enabling the storage framework capability (sfwcomm device) of the host bus adapters (HBAs). Currently, FC and SAS technologies are supported. The Setting up cluster storage communication link provides more details about supported HBAs and the required steps to set up the storage framework communication.
- In a virtual environment using N-Port ID Virtualization (NPIV) or virtual Small Computer System Interface (vSCSI) with a VIOS instance, enabling the sfwcomm interface requires activating the target mode (the tme attribute) on the real adapter in the VIOS instance and defining a private virtual LAN (VLAN) (ID 3358) for communication between the partition containing the sfwcomm interface and VIOS. The real adapter on the VIOS must be a supported HBA.
The target mode enabled (
tme) attribute for a supported
adapter is only available when the minimum AIX level for CAA is installed. The
configuration steps are as follows:
- Configure the FC adapters for SAN heartbeat on the VIOS instances. Use the
chdevcommand to enable the
# chdev –l fcsX –a tme=yes -perm
- Run the
chdevcommand to enable dynamic tracking and fast failure recovery on all FSCSI adapters.
# chdev –l fscsiX –a dyntrk=yes –a fc_err_recov=fast_fail
- Restart the VIOS instances.
- On the Hardware Management Console (HMC) create a new virtual Ethernet adapter for each cluster LPAR and VIOS. Set the VLAN ID to 3558 (no other VLAN ID is allowed).
- On the VIOS, run the
cfgmgrcommand and check for the virtual Ethernet adapter and sfwcomm device using the
# lsdev –C | grep sfwcomm sfwcomm0 Available 01-00-02-FF Fibre Channel Storage Framework Comm sfwcomm1 Available 01-01-02-FF Fibre Channel Storage Framework Comm
- On the cluster nodes, run the
cfgmgrcommand and check for the virtual Ethernet and sfwcomm device using the
- No other configuration is required in PowerHA. When the cluster is up and
running, you can check the status of SAN heartbeat using the
You can run
/usr/lib/cluster, as shown below, to check if sfwcomm and dpcomm are
working or not.
0) root @ <nodename>: /usr/lib/cluster # ./clras sancomm_status +---------------------------------------------------------------+ | NAME | UUID | STATUS | +---------------------------------------------------------------+ | servr2.abcdefg.xxx.com | 6c3af126-d8d4-11e2-9c7a-00145ee770e9 | UP | +---------------------------------------------------------------+ (0) root @ <nodename>: /usr/lib/cluster # ./clras dpcomm_status +---------------------------------------------------------------+ | NAME | UUID | STATUS | +---------------------------------------------------------------+ | servr1.abcdefg.xxx.com | 54119a46-d8d4-11e2-ac6b-00145ee770e9 | UP| +---------------------------------------------------------------+ | servr2.abcdefg.xxx.com | 6c3af126-d8d4-11e2-9c7a-00145ee770e9 | UP| +---------------------------------------------------------------+ (0) root @ <nodename>: /usr/lib/cluster #
Before migrating PowerHA to PowerHA 7.1 and later, test whether the nodes in your environment support
multicast-based communication. To test end-to-end multicast communication for all
nodes used to create the cluster on your network, run the
mping command, which is part of the CAA framework of AIX. You can run
mping with a specific multicast address; otherwise
the command uses a default address. The following is an example of the
mping command for the multicast address 184.108.40.206,
where nodeA is the receiver and nodeB is the sender. You must run the following
commands from both the nodes at the same time:
- From nodeA, run
mping –r –v –c 5 –a 220.127.116.11
- From nodeB, run
mping –s –v –c 5 –a 18.104.22.168
Repeat the steps this time by reversing the sender and receiver.
You can choose to stop cluster services on all nodes, and then install PowerHA 7.1. After all the checks are successful, the clconvert utility runs from installp to convert the configuration represented in the back level PowerHA configuration data classes to PowerHA 7.1 and later version, including the running of mkcluster and creating the CAA version of cluster in addition to removing any discovered interface not in the previous version of PowerHA (such as SAN/FC heartbeat sfwcomm). After AIX has been migrated, follow these steps to migrate the PowerHA level to version PowerHA 7.1 and later.
- Stop cluster services on all cluster nodes. Use the
smitty clstopcommand and select the Bring a Resource Group Offline option.
- Ensure that the cluster services have been stopped. Use the
lssrc –ls clstrmgrEScommand to check the cluster state. It should be ST_INIT.
/usr/sbin/clmigcheckon the first node and select option 1.
- If the cluster cannot be migrated, the clmigcheck utility will indicate that in error messages. Remove the unsupported elements. If no errors are reported, skip step 5.
- Perform a verification and then synchronize.
clmigcheckonce again and select option 1. The
clmigcheckcommand says The ODM has no unsupported elements, as shown in the following figure.
- Now select option 3 to enter the repository disk information and optionally provide the multicast IP address. The data is saved in the /var/clmigcheck/clmigcheck.txt file on each node. You need to enter this information only on the first node.
- Populate the /etc/cluster/rhosts file on this node with the IP addresses of all the cluster nodes (addresses corresponding to hostname command).
- Refresh the clcomd daemon, and run the
refresh –s clcomdcommand.
- Install PowerHA 7.1 and later on the first node.
- Run the following steps on all remaining nodes, one at a time.
- Run /usr/sbin/clmigcheck. It prompts you to install new
version of PowerHA, as shown in the message in the following
- Add the IP addresses of all the cluster nodes in the
/etc/cluster/rhosts file and refresh the
- Install PowerHA 7.1.2.
- Run /usr/sbin/clmigcheck. It prompts you to install new version of PowerHA, as shown in the message in the following figure.
- /usr/sbin/clmigcheck detects the last node when it runs and it
creates a cluster-aware infrastructure, that is, a CAA cluster on all the nodes.
This can be verified by running the
- Update the /etc/cluster/rhosts file, refresh
clcomdand install PowerHA SystemMirror 7.1.
- Start the cluster services, one node at a time, and ensure that each node successfully joins the cluster. After the last node has joined the cluster, your migration is successful.
In rolling migration, the newer version of PowerHA is installed (one node at a time), while the remaining nodes continue to run cluster services and host the workload. In this mixed version state, PowerHA continue to respond to cluster events. In a rolling migration, you stop cluster services on the target node with the Move Resource Groups option. There is a brief interruption while the application moves to the backup or fallover node, and a second interruption while the application moves back to the primary or home node after it has been migrated. The steps to migrate are as follows:
- Run /usr/sbin/clmigcheck on the first node, and select option 1.
- If the cluster cannot be migrated, error messages will be displayed. In that case, remove all unsupported elements.
- Verify and sync the corrected cluster definition from the first node.
- Populate the /et/cluster/rhosts file on this node, and refresh the clcomd daemon.
- Run clmigcheck again to verify that there are no further unsupported elements. The The ODM has no unsupported elements message is displayed, as shown in the following figure.
- Stop the cluster services with the Move Resource Groups option.
- Run clmigcheck again and select option 3. Enter the shared repository disk, and optionally, provide the multicast IP address. The information is saved in /var/clmigcheck/clmigcheck.txt
- Install the newer version of PowerHA on this node.
- After the installation is complete, start the cluster services.
- On the remaining nodes, follow these steps, one node at a time, after stopping
the cluster services with the Move Resource Groups option.
- Populate the /etc/cluster/rhosts file with the IP addresses of all cluster nodes.
- Refresh the clcomd daemon.
- Run /usr/sbin/clmigcheck. A message, as shown in the following figure is displayed.
- Install PowerHA 7.1.2 on the node.
- Start cluster services.
- /usr/sbin/clmigcheck will detect the last node when it runs and
will create a CAA cluster on all the nodes. Run the
/usr/sbin/lscluster –mcommand to verify this.
- Start cluster services on the last node. After the last node joins the cluster, your migration is complete.
The snapshot migration path requires cluster services to be down on all the nodes, thus calling for a cluster outage or application downtime. To migrate a cluster using this path, you need to perform the following steps.
- Create a cluster snapshot. By default, the snapshot is saved in the /usr/es/sbin/cluster/snapshots directory. Save a copy of it in /tmp or some other location.
- Stop the cluster services on all nodes using the Bring Resource Groups Offline option.
- Run /usr/sbin/clmigcheck on the first node, and then select option 2. Enter the snapshot name.
- If the utility reports errors for unsupported elements, the snapshot cannot be
migrated. In this case, remove all unsupported elements reported by
clmigcheck. If no errors are reported, go to step 7.
- Take a new cluster snapshot and save a copy of it in /tmp.
- Run /usr/sbin/clmigcheck again with option 2 to ensure that there are no unsupported elements.
- Choose option 3 in /usr/sbin/clmigcheck to enter the shared disk (repository disk) and optionally, the multicast address.
- Remove the existing version of PowerHA software on all cluster nodes.
- In the /etc/cluster/rhosts file, fill in the IP addresses of all cluster nodes (IP addresses corresponding to the host name command).
- Refresh clcomd using the
refresh –s clcomdcommand.
- Install a newer version of PowerHA.
- Convert the snapshot
# /usr/es/sbin/cluster/conversion/clconvert_snapshot –v 6.1.0 –s <snapshot file name>
- Restore the converted snapshot. Use the path: smitty sysmirror -> Cluster Nodes and Networks -> Manage the Cluster -> Snapshot Configuration -> Restore the Cluster Configuration from a Snapshot.
- After the restoration is done, run verification and synchronization. This
creates and enables the CAA infrastructure. You can verify this using the
- Start the cluster services, one node at a time. After the last node joins the cluster, the migration is complete.
- What's New in PowerHA
- PowerHA version compatibility matrix
- Cluster Aware AIX
- PowerHA Information Center
- How to test multicast
- AIX 7.1 migration
- AIX 6.1 migration
Get products and technologies
- Find and download service packs from Fix Central.
- Participate in the discussion forum
- Get involved in the My developerWorks community. Connect with other developerWorks users while exploring the developer-driven blogs, forums, groups and wikis.