Upgrading memory in the nodes of an IBM Storage Fusion HCI System System
Use these installation instructions to upgrade the compute-only nodes (9155-C00 and
9155-C04) and/or compute-storage nodes (9155-C01 and 9155-C05) of a IBM Storage Fusion HCI System system by installing the additional memory
modules (AHJK or AHJN) or replacement modules (RPQ 8S1881). The 9155-C00 and 9155-C01 can be factory
upgraded to 512GB of RAM and field upgraded to either 512GB of RAM or 1024GB of RAM. The two memory
choices for 9155-C04 are 16 GB/core and 32GB/core (1024GB and 2048GB) per 64-core server. The same
instructions are printed and attached to the hardware.
Before you begin
Installing the memory upgrade requires temporarily shutting down each of the compute nodes that
are being upgraded. Ensure all the nodes in the IBM Storage Fusion HCI System system are functioning properly so that the
system can tolerate the shutting down of compute nodes without impact on running applications or the
storage cluster.
The system must have sufficient capacity to run all applications with one compute node shutting
down. If not, close some applications until the upgrade is complete.
Before you begin the task of installing the memory upgrade, identify the compute nodes in the
rack enclosure to be upgraded. Ensure that each compute node has enough DIMMs for the upgrade. Only
9155-C00, 9155-C01, 9155-C04, 9155-C05, 9155-C10, and 9155-C14 compute nodes are supported by the
memory upgrade. The 9155-G01 GPU nodes and the 9155-F01 AFM nodes do not support this memory
upgrade. It is likely that more than one compute node in the IBM Storage Fusion HCI System system is being upgraded. Because each
compute node that is upgraded must be powered off before the update. Also, it is important that only
one compute node is upgraded at a time to maintain the integrity of the IBM Storage Scale ECE storage cluster and the OpenShift® Container Platform control plane. The only exception is when the entire system
is powered down to perform upgrades.
A Philips screwdriver is used to turn the locking screw on the cover latches of the compute-only
nodes (9155-C10, 9155-C00, 9155-C04) and storage-compute nodes (9155-C14, 9155-C01, 9155-C05).
Observe all normal safety precautions. Refer to the safety notices provided in the feature
kit.
Before powering off the compute nodes, move the node to maintenance mode. For the steps to move
to maintenance mode, see Administering the node.
On the node details page and on the Overview tab, view the health of
events, disks, and ports. The green (normal) tick mark indicates that the status is healthy. If a
red or yellow indicator appears, do not proceed. You must correct the error or warning to safely
continue with the upgrade.
Note: Note the location for easy finding of the compute node in the rack
in future steps.
About this task
For actual memory expansion orders, contact IBM Support. Follow the procedure to install the
memory upgrade.
Procedure
After successfully placing a compute node in maintenance mode follow the steps to
upgrade.
Power down the compute node. For the steps to move to power down, see Administering the node.
Move the compute node into position for the upgrade.
Open the rear door of the IBM Storage Fusion HCI System
system and identify the compute node that was shut down.
Check that the four optical Ethernet network cables (2x 100GbE and 2x 25GbE), two RJ45
copper Ethernet cables (one in an OCP port, the other in the IMM), and the two power cables have
correct labels. If the labels are damaged or missing, add replacement labels that indicate the
compute node and port where the cable is attached. When all cables have the correct labels,
disconnect the cables from the compute node. Be careful not to dislodge any cables or power cords
from any other component in the system during this process.
Note: It is possible that there are no 25GbE optical network cables attached to the compute node.
Move to the front of the system, open the door and locate the powered down compute
node.
Unlatch the catches on the sides of the compute node that hold the rails, and then
pull the compute node forward until the rails are fully extended.
Remove the compute node from the rails and then place the compute node securely on the
workbench. To remove the compute node from the rails, follow the instructions that are shown in the
“System Removal” illustration as a guide.
Note: It is not necessary to remove the compute node fully from the rack f upgrade, but you can
choose to do so if it is more convenient.
Remove the compute node top cover as follows:
Note: Ensure you wear an ESD wrist strap, while removing the compute node top cover.
Turn the lock screw on the cover latch to the open position.
Press the blue button.
Lift the cover latch.
Slide the cover toward the back of the compute node until it detaches from the
chassis, remove it and place the cover in a safe place.
Install the Feature Code AHJK or AHJN memory module as follows:
The FC AHJK or
AHJN memory upgrade requires only adding more DIMMs to the compute node. The compute node already
has 16 x16GB DIMMs installed. At the end of the installation, ensure that all 32 memory slots of the
compute node are filled with 16GB DIMMs.
To open the server and add more memory, six network
cables (two Ethernet RJ45 cables, two 25G split cables and two100G fiber cables) and two power cords
must be removed. After memory chips are added, these eight cables must be rewired:
25G ports- split cables have to be firmly pressed in order for them to be securely seated to the
cage. If the connection become loose, then the port does not work.
Handle 100G fiber cables with great care as the fiber cables bend and break easily. Take care
when you work with existing memory chips as it may get loose and wires inside the server may get
disturbed. Some times, one has to use force to install the memory chip that can disturb near by
wires as well.
The following picture showing 25 G ports and 100 G ports:
Locate the components to be installed. You need 16 of the 16GB DIMMs for each of the
compute nodes to be upgraded.
Verify that 16 DIMMs are already in the compute node and occupy the following slots as
follows: 1, 3, 5, 7, 10, 12, 14, 16, 17, 19, 21, 23, 26, 28, 30, 32
Remove the air baffle as shown in figure 5 and 6.
Note: Ensure you wear an ESD wrist strap, while removing the air baffle.
Add 16 more 16GB DIMMs to the empty slots in the following slot order: 13, 29, 15, 31,
4, 20, 2, 18, 9, 25, 11, 27, 8, 24, 6, 22.
For each slot, open the retaining clips at
each end of the slot and then firmly press the new DIMM at both ends down into the slot, causing the
tabs to snap to the locked position.
Refer to the DIMM installation order label shown in
the figure.
After adding the 16 DIMMs, verify that all DIMM slots are now
populated.
Install the air baffle.
Install the RPQ 8S1881 memory module as follows:
The RPQ 8S1881 memory upgrade
removes all 16 of the 16GB DIMMs from the compute node and replace them with 16 x 64GB DIMMs in the
same slots where the 16GB DIMMs removed.
Locate the components to be installed. You need 16 of the 64GB DIMMs for each of the
compute nodes to be upgraded.
Verify that 16 DIMMs are already in the compute node and occupy the following slots as
follows: 1, 3, 5, 7, 10, 12, 14, 16, 17, 19, 21, 23, 26, 28, 30, 32.
Remove the air baffle as shown in figure 8 and 9.
Note: Ensure you wear an ESD wrist strap, while removing the air baffle.
Remove the 16 x 16GB DIMMs that are already in the compute node and should occupy the
following slots: 1, 3, 5, 7, 10, 12, 14, 16, 17, 19, 21, 23, 26, 28, 30, 32.
Open the
retaining clips at each end of the DIMM slot and then lift the DIMM straight up while holding both
ends of the DIMM.
Add the 16 x 64GB DIMMs to the empty DIMM slots in the following slot order: 14,
30,16, 32, 3, 19, 1 ,17, 10, 26, 12, 28, 7, 23, 5, 21. For each slot, open the retaining clips (if
not already open) at each end of the slot and then firmly press the new DIMM at both ends down into
the slot, causing the tabs to snap to the locked position.
Place the compute node back into its operating position.
If the compute node is
removed from the rack, place the compute node back on the rails using the “System Installation”
illustration shown on the label on top of the compute node.
Slide the compute node back into position in the rack enclosure and reconnect the
power cables. Labels on the power cables indicate the rack position of the compute node where the
cables should be installed.
If required, verify the memory using the rack-mounted console.
Connect the video, keyboard, and mouse of the rack-mounted console to the compute
node.
Press F1 while powering up the compute node to go into the system configuration
utility.
Select System Summary from the left menu and scroll down to
view the DIMM information and verify that the DIMM Total Count and DIMM Total Capacity have the
expected values.
Power down the compute node.
Reconnect the network cables. Ensure that compute nodes are installed on the dedicated
racks using labels. Labels on the network cables indicate the rack position of the compute node
where the cables should be installed.
Power up the compute node as follows:
After a compute node has been
successfully upgraded, power up using these steps:
In the Manage resource window, select the Power
up option from the Resource action list. (This may take a few minutes.)
Take the compute node out of maintenance mode.
Ensure that the success
notification is displayed.
Repeat the steps for the other compute nodes in the rack enclosure.
Repeat
from the step “Before powering off the compute node” for each of the compute nodes in the IBM Storage Fusion HCI System system that is being upgraded.
Verify the new configuration.
Check the amount of memory shown in the
management GUI for each compute node and compare it to the estimated amount, which should be 512GB
(FC AHJK), 1024GB (RPQ 8S1881) or 2048GB (AHJN).