Accessing a remote GPFS file system

Learn about accessing the files in a GPFS file system from another cluster.

The ability to access and mount GPFS file systems that are owned by other clusters in a network of sufficient bandwidth is accomplished by using the mmauth, mmremotecluster, and mmremotefs commands. Each site in the network is managed as a separate cluster, while allowing shared file system access.

The cluster owning the file system is responsible for administering the file system and granting access to other clusters on a per cluster basis. After access to a particular file system has been granted to nodes in another GPFS cluster, the nodes can mount the file system and perform data operations as if the file system were locally owned.

Each node in the GPFS cluster requiring access to another cluster's file system must be able to open a TCP/IP connection to every node in the other cluster.

Nodes in two separate remote clusters mounting the same file system are not required to be able to open a TCP/IP connection to each other. For example, if a node in clusterA mounts a file system from clusterB, and a node in clusterC wants to mount the same file system, nodes in clusterA and clusterC do not have to communicate with each other.

Each node in the GPFS cluster requiring file system access must have one of the following:
  • A virtual connection to the file system data through an NSD server (refer to Figure 1).
  • A physical connection to the disks containing file system data (refer to Figure 2).

In this example, network connectivity is required from the nodes in clusterB to all the nodes in clusterA even if the nodes in clusterB can access the disks in clusterA directly.

Figure 1. Remote mount of a file system using NSD server access
Remote mount of a file system using NSD server access. This graphic shows two GPFS clusters named cluster A and cluster B, and a wide area network connecting them. Cluster A consists of three nodes and one disk. One node runs AIX and the other two run Linux. All three nodes are connected to the disk. One of the Linux nodes is also the NSD server, and this node is connected to the wide area network. The other two nodes are not connected to the wide area network. Cluster B consists of three nodes, two of which run AIX and one of which runs Linux. All three nodes are connected to the wide area network. There are no disks in cluster B.
Figure 2. Remote mount of a file system using SAN-attached disks
Remote mount of a file system using SAN-attached disks. This graphic shows two GPFS clusters named cluster A and cluster B. Cluster A consists of three nodes and one disk. One node runs AIX and the other two run Linux. All three nodes are connected to the disk. Cluster B consists of three nodes, two of which run AIX and one runs Linux. All three nodes are connected to the disk associated with cluster A. Cluster B has no disks of its own.
Note: SAN access of disks from different hardware platforms and operating systems are not supported. This limitation is because of the differences in how disks are labeled, how partitions are created and managed, and how multi-pathing managers react to error conditions among various operating systems and hardware platforms.
Note: If the cluster is configured with direct access to LUNs, then the disks might silently fall back to NSD server and results in the degradation of the network performance and slower I/O. For more information, see Changing NSD server usage and failback.
Figure 3 illustrates a multi-cluster configuration with multiple NSD servers. In this configuration:
  • The two nodes in Cluster 1 are defined as the NSD servers (you can have up to eight NSD server nodes).
  • All three clusters are connected with Gigabit Ethernet.
  • Cluster 1 shares an InfiniBand switch network with Cluster 2 and an InfiniBand switch network with Cluster 3.
To take advantage of the fast networks and to use the nodes in Cluster 1 as NSD servers for Cluster 2 and Cluster 3, you must configure a subnet for each of the supported clusters. For example, issuing the command:
  • mmchconfig subnets="<IB_Network_1> <IB_Network_1>/Cluster1" in Cluster 2 allows nodes N2 through Nx to use N1 as an NSD server with InfiniBand Network 1 providing the path to the data.
  • mmchconfig subnets="<IB_Network_2> <IB_Network_2>/Cluster1" in Cluster 3 allows nodes N2+x through Ny+x to use N1+x as an NSD server with InfiniBand Network 2 providing the path to the data.
Figure 3. Multi-cluster configuration with multiple NSD servers
Illustration shows three clusters with two of them being remote clusters. Cluster 1 contains the two NSD servers for this configuration. Cluster 2 and Cluster 3 use separate networks to access the NSD servers in Cluster 1. With this configuration, Cluster 2 and Cluster 3 can access the data stored on the NSDs without having a direct connection to those disks.