ESS software deployment preparation

Install an ESS software package and deploy the storage servers by using the following information. The goal is to create a cluster that allows client or protocol nodes to access the file systems.

ESS 3200 ESS 3000 ESS 5000 ESS Legacy
Runs on POWER9™ EMS POWER8® or POWER9 EMS POWER9 EMS POWER8 or POWER9 EMS
I/O node OS Red Hat® Enterprise Linux® 8.8 x86_64 Red Hat Enterprise Linux 8.8 (Ootpa) Red Hat Enterprise Linux 8.8 PPC64LE Red Hat Enterprise Linux 8.8 (Ootpa)
Architecture x86_64 Ootpa PPC64LE Ootpa
IBM Spectrum® Scale Start of change5.1.2.14 End of change Start of change5.1.2.14 End of change Start of change5.1.2.14End of change Start of change5.1.2.14 End of change
Start of changeKernelEnd of change Start of change4.18.0- 477.27.1.el8_8End of change Start of change4.18.0- 477.27.1.el8_8End of change Start of change4.18.0- 477.27.1.el8_8End of change Start of change

Start of change3.10.0-1160.99.1.el7.ppc64leEnd of change

End of change
Systemd Start of change239-74.el8_8.5.x86_64End of change Start of change239-74.el8_8.5End of change 239-74.el8_8.5.ppc64le 219-78.el7_9.7.ppc64le
Network manager 1.40.16-4.el8 1.40.16-4.el8 1.40.16-4.el8_8.ppc64le 1.18.8-2.el7_9
GNU C Library glibc-2.28-225.el8 glibc-2.28-225.el8 glibc-2.28-225.el8 glibc-2.17-326.el8
OFED Start of changeMLNX_OFED_LINUX-23.10-0.5.5.5End of change Start of changeMLNX_OFED_LINUX-23.10-0.5.5.5End of change

Start of changeMLNX_OFED_LINUX-23.10-0.5.5.5End of change

Start of changeMLNX_OFED_LINUX-4.9-6.0.6.1End of change
Start of changeFirmware RPMEnd of change Start of change6.0.0.53End of change Start of change6.0.0.53End of change Start of change6.0.0.53End of change Start of change6.0.0.53End of change
Mpt3sas N/A N/A

Start of change46.00.01.00 End of change - 5U92 (not in box)

Start of change46.00.01.00End of change - 4U106 (not in box)

34.00.00.00

(not in box)

Podman Start of change1.6.4-28End of change

Start of change1.6.4-28 RHEL8End of change

1.6.4-28

Start of change1.4.4-4 RHEL7End of change

Container version Start of changeRed Hat UBI 8.8End of change Start of changeRed Hat UBI 8.8End of change Start of changeRed Hat UBI 8.8End of change Start of changeRed Hat UBI 8.8End of change
Start of changeAnsible®End of change Start of changeStart of change2.14.2End of changeEnd of change Start of changeStart of change2.14.2End of changeEnd of change Start of change2.14.2End of change Start of change2.14.2End of change
xCAT 2.16.3

Not used in customer shipped image - only for SCT

Start of change2.16.3End of change

Start of change2.16.3 (for SCT only)End of change

Start of change2.16.3 (for SCT only)End of change

OPAL Start of changeopal-prd-ess.v4-1.el8.x86_64.rpm End of change N/A

Start of changeopal-prd-ess.v4.1-1.el8.ppc64le.rpmEnd of change

N/A
System firmware
  • BMC: 12.53
  • Server0FPGA: Start of change0140End of change
  • Server1FPGA: Start of change0140End of change
  • Midplane1PrimaryFPGA: 0343
  • Midplane1SecondaryFPGA: 0343
  • Midplane2PrimaryFPGA: 0343
  • Midplane2SecondaryFPGA: 0343
  • Midplane3PrimaryFPGA: 0343
  • Midplane3SecondaryFPGA: 0343
  • DriveplanePrimaryFPGA: 0326
  • DriveplaneSecondaryFPGA: 0326

FW 2.02.000_0B0G_1.73_FB300052_0C32)

Boot drive firmware: 9F26

Start of changeFW950.90 (VL950_138)
NVDIMM ver: Bundled
BPM ver: BundledEnd of change

Start of changeSV860_245 (FW860.B3)End of change

Start of changeBoot driveEnd of change Start of change
  • Bootdrive1_Micron_7300_MTFDHBA960TDF: 954300P0
  • Bootdrive2_Micron_7300_MTFDHBA960TDF: 954300P0
End of change
Start of change
  • SMART:
    Prod ID: SRM2S86Q800GQT51IM
    P/N: 01LL447IBM
    FRU: 01LL447
    FW: 1361
  • Micron: MTFDDAV960TDS
    P/N: 01LL446IBM
    FRU: 01LL587
    FW: ML32
End of change
Start of change

FRU: 00E9962
Firmware Version: 39463236 (9F26)

End of change
Start of changeStart of changeE700End of changeEnd of change
Network adapter
  • MT4119 = 16.35.3006
  • MT4120 = 16.35.3006
  • MT4121 = 16.35.3006
  • MT4122 = 16.35.3006
  • MT4123 = 20.39.1002
  • MT4125 = 22.39.1002
  • MT4129 = 28.39.1002
  • MT4119 = 16.35.3006
  • MT4120 = 16.35.3006
  • MT4121 = 16.35.3006
  • MT4122 = 16.35.3006
  • MT4123 = 20.39.1002
  • MT4125 = 22.39.1002
  • MT4129 = 28.39.1002
  • MT4119 = 16.35.3006
  • MT4120 = 16.35.3006
  • MT4121 = 16.35.3006
  • MT4122 = 16.35.3006
  • MT4123 = 20.39.1002
  • MT4125 = 22.39.1002
  • MT4129 = 28.39.1002

MT4120 CX-5 EN 01FT741

MT4121 CX-5 VPI 01LL584

MT4122 CX-5 SRIOV VF 01LL584

Start of change
  • MT27500 = 10.16.1020
  • MT4099 = 2.42.5000
  • MT26448 = 2.9.1326
  • MT4103 = 2.42.5000
  • MT4113 = 10.16.1200
  • MT4115 = 12.28.2006
  • MT4117 = 14.28.0016
  • MT4118 = 14.28.0016
  • MT4119 = 16.35.2000
  • MT4120 =16.35.2000
  • MT4121 = 16.35.2000
  • MT4122 =16.35.2000
  • MT4123 = 20.36.1010
  • MT4125 = 22.36.1010
  • MT4129 = 28.36.1202
End of change
ESA

esagent.pLinux-4.5.7-0

esagent.pLinux-4.5.7-0

esagent.pLinux-4.5.7-0

esagent.pLinux-4.5.7-0

BIOS 12.16.00 52 N/A N/A
HAL Start of changeibm-ess-hal-2.0.0.0-7.2.x86_64End of change N/A N/A N/A

Changes in this release

Start of change
  • Support for IBM Storage Scale 5.1.2.14
  • Bug fixes and improvements
  • Updated security patches
End of change
Start of change

POWER9 EMS stack

Item Version
IBM Spectrum Scale IBM Spectrum Scale 5.1.2.14
Operating system Red Hat Enterprise Linux 8.8
ESS ESS 6.1.2.8
Kernel 4.18.0-477.27.1.el8_8.ppc64le
Systemd 239-74.el8_8.5
Network Manager 1.40.16-4.el8_8
GNU C Library glibc-2.28-225.el8_8.6.ppc64le.rpm
Mellanox OFED Start of changeMLNX_OFED_LINUX-23.10-0.5.5.5End of change
ESA 4.5.7-0
Ansible 2.14.2
Podman 1.6.4-28
Container OS Start of changeRed Hat UBI 8.8End of change
xCAT 2.16.3 (Not used in customer-shipped image; only for SCT)
Firmware RPM Start of changegpfs.ess.firmware-6.0.0-53.ppc64le.rpmEnd of change
System firmware Start of changeFW950.90 (VL950_138)End of change

Start of changeNVDIMM ver: Bundled End of change

Start of changeBPM ver: BundledEnd of change

Boot drive adapter IPR 19512c00
Boot drive firmware Start of change
  • Firmware: 9F26
  • Host adapter driver: 46.00.01.00
  • Host adapter firmware: 16.00.11.00
End of change
1Gb NIC firmware
  • Driver: tg3
  • Version: 3.137
  • Firmware version: 5719-v1.24i
Support RPM
  • Start of changegpfs.gnr.support-essbase-6.1.2-0.noarch.rpmEnd of change
  • gpfs.gnr.support-ess3000-6.1.2-0.noarch.rpm
  • Start of changegpfs.gnr.support-ess3200-6.1.2-0.noarch.rpmEnd of change
  • gpfs.gnr.support-ess5000-6.1.2-0.noarch.rpm
Network adapter
  • MT4119 = 16.35.3006
  • MT4120 = 16.35.3006
  • MT4121 = 16.35.3006
  • MT4122 = 16.35.3006
  • MT4123 = 20.39.1002
  • MT4125 = 22.39.1002
  • MT4129 = 28.39.1002
Start of change
Adapter FC PN FRU Vendor ID FW level
Lassen 2-Port EDR IB EC64 00WT174 00WT176 MT4121 16.34.1002
Lassen 2-Port EDR IB EC64 01FT810 01FT742 MT4121 16.34.1002

PCIe4 LP 2-port 100Gb ROCE EN LP

EC67 01FT741 01FT742 MT4120 16.34.1002
PCIe3 LP 2-Port 25/10Gb NIC&ROCE SR/Cu EC2T 01FT752 01FT753 MT4117 14.32.1010
PCIe3 LP 2-Port 25/10Gb NIC&ROCE SR/Cu EC2T 01FT755 01FT756 MT4117 14.32.1010
PCIe3 LP 2-Port 25/10Gb NIC&ROCE SR/Cu AJZR 01YM281 01YM283 MT4117 14.32.1010
End of change
End of change

Support matrix

Release OS Runs on Can upgrade or deploy
ESS 3200 Start of change6.1.2End of change Red Hat Enterprise Linux 8.8 Start of changex86_64End of change
  • POWER9 EMS
  • ESS 3200 nodes
  • POWER9 EMS
  • POWER9 protocol nodes
ESS 3000 Start of change6.1.2End of change Red Hat Enterprise Linux 8.8 (Ootpa)
  • POWER9 EMS
  • ESS 3000 nodes
  • POWER8 EMS
  • POWER9 EMS
  • POWER8 protocol nodes
  • POWER9 protocol nodes
ESS 5000 Start of change6.1.2End of change Red Hat Enterprise Linux 8.8 (PPC64LE)
  • POWER9 EMS
  • ESS 5000 nodes
  • POWER9 EMS
  • POWER9 protocol nodes
ESS Legacy Start of change6.1.2End of change
  • Red Hat Enterprise Linux 8.8 (Ootpa)
  • Red Hat Enterprise Linux 7.9 (PPC64LE)
  • POWER8 EMS
  • POWER9 EMS
  • ESS POWER8 I/O nodes (PPC64LE)
  • ESS POWER8 protocol nodes (PPC64LE)
  • ESS POWER9 protocol nodes (PPC64LE)*
  • POWER8 EMS
  • POWER9 EMS

Prerequisites

  • This document (ESS Software Quick Deployment Guide)
  • SSR completes physical hardware installation and code 20.
    • SSR uses Worldwide Customized Installation Instructions (WCII) for racking, cabling, and disk placement information.
    • SSR uses the respective ESS Hardware Guide (ESS 3000 or ESS 5000 or ESS 3200) for hardware checkout and setting IP addresses.
  • Worksheet notes from the SSR
  • Latest ESS tgz downloaded to the EMS node from Fix Central (If a newer version is available).
    • Data Access Edition or Data Management Edition: Must match the order. If the edition does not match your order, open a ticket with the IBM® Service.
  • High-speed switch and cables have been run and configured.
  • Low-speed host names are ready to be defined based on the IP addresses that the SSR have configured.
  • High-speed host names (suffix of low speed) and IP addresses are ready to be defined.
  • Container host name and IP address are ready to be defined in the /etc/hosts file.
  • Host and domain name (FQDN) are defined in the /etc/hosts file.
  • ESS Legacy 6.1.x.x Only: You must convert to mmvdisk before deploying the ESS Legacy 6.1.x.x container if you are coming from a non-container version such as ESS 5.3.x.x. If you have not done so already, convert to mmvdisk by using the following steps:
    1. Check whether there are any mmvdisk node classes.
      mmvdisk nodeclass list

      There should be one node class per ESS Legacy building-block. If the command output does not show mmvdisk for your ESS Legacy nodes, convert to mmvdisk before running the ESS Legacy 6.1.0.x container.

    2. Convert to mmvdisk by running the following command from one of the POWER8 I/O nodes or from the POWER8 EMS node.
      gssgenclusterrgs -G gss_ppc64 --suffix=-hs --convert 
      You can also use -N with a comma-separated list of nodes.
      Note: Wait for 5 minutes for daemons to recycle. The file system remains up.

What is in the /home/deploy directory on the EMS node?

  • ESS 5000 tgz used in manufacturing (may not be the latest)
  • ESS 3000 tgz used in manufacturing (may not be the latest)
  • ESS Legacy tgz used in manufacturing (may not be the latest)
  • ESS 3200 tgz used in manufacturing (may not be the latest)

Support for signed RPMs

ESS or IBM Spectrum Scale RPMs are signed by IBM.

The PGP key is located in /opt/ibm/ess/tools/conf.
-rw-r-xr-x 1 root root 907 Dec 1 07:45 SpectrumScale_public_key.pgp
You can check whether an ESS or IBM Spectrum Scale RPM is signed by IBM as follows.
  1. Import the PGP key.
    rpm --import  /opt/ibm/ess/tools/conf/SpectrumScale_public_key.pgp
  2. Verify the RPM.
    rpm -K RPMFile

ESS 3000, ESS 5000, ESS 3200, and ESS Legacy networking requirements

In any scenario you must have an EMS node and a management switch. The management switch must be split into two VLANs.
  • Management VLAN
  • Service/FSP VLAN
    Note: Start of changeTo future proof your environment for ESS 3200, modify any existing management switches to the new VLAN configuration. For more information, see Switch VLAN configuration instructions.End of change
You also need a high-speed switch (IB or Ethernet) for cluster communication.

ESS 3000

POWER8 or POWER9 EMS

It is recommended to buy a POWER9 EMS with ESS 3000. If you have a legacy environment (POWER8), it is recommended to migrate to IBM Spectrum Scale 5.1.x.x and use the POWER9 EMS as the single management server.
  • If you are adding ESS 3000 to a POWER8 EMS:
    • An additional connection for the container to the management VLAN must be added. A C10-T2 cable must be run to this VLAN.
    • A public/campus connection is required in C10-T3.
    • A management connection must be run from C10-T1 (This should be already in place if adding to an existing POWER8 EMS with legacy nodes).
    • Port 1 on each ESS 3000 canister must be connected to the management VLAN.
  • If you are using an ESS 3000 with a POWER9 EMS:
    • C11-T1 must be connected on the EMS to the management VLAN.
    • Port 1 on each ESS 3000 canister must be connected to the management VLAN.
    • C11-T2 must be connected on the EMS to the FSP VLAN.
    • HMC1 must be connected on the EMS to the FSP VLAN.
Note: It is mandatory that you connect C11-T3 to a campus connection or run an additional management connection. If you do not do this step, you will lose the connection to the EMS node when the container starts.

ESS 5000 or ESS 3200

POWER9 EMS support only

EMS must have the following connections:
  • C11-T1 to the management VLAN
  • C11-T2 to the FSP VLAN
  • C11-T3 to the campus network
  • HMC1 to the FSP VLAN
ESS 5000 nodes must have the following connections:
  • C11-T1 to the management VLAN
  • HMC1 to the FSP VLAN
ESS 3200 nodes must have the following connections:
  • Single management connection per canister:
    • Each connection is split between 2 MAC addresses:
      1. BMC
      2. Operating system
    • The BMC connection requires a VLAN tag to be set for proper communication with the EMS node.
ESS 3200 requirements
  • Management connections
    • Shared management port (visible to OS)
  • BMC connection
    • Shared management port (visible to BMC)
  • High-speed connections
    • InfiniBand or Ethernet
Management switch
  • Typically, a 48-port switch
  • Two VLANs required
    • Management VLAN (VLAN 102)
    • FSP/BMC VLAN (VLAN101)
  • ESS 3200 dedicated trunk ports
    • Routes BMC traffic to VLAN 101
Note: The VLANs shown here are default for the IBM Cumulus switch. The VLAN value can be modified according to your environment.
Figure 1. ESS 3200 container networking
ESS 3200 container networking
Figure 2. ESS 3200 network diagram
ESS 3200 network diagram
Figure 3. ESS 3200 Ethernet ports and switch
ESS 3200 Ethernet ports and switch
The ports highlighted in green are the ESS 3200 trunk ports. These are special ports that are for the ESS 3200 only. The reason for these ports is that each ESS 3200 canister has a single interface for both the BMC and the OS but unique MAC addresses. By using a VLAN tag, canister BMC MAC addresses are routed to the BMC/FSP/Service VLAN (Default is 101).
IBM racked orders have the switch preconfigured. Only the VLAN tag needs to be set. If you have an existing IBM Cumulus switch or customer supplied switch, it needs to be modified to accommodate the ESS 3200 trunk port requirement. For more information, see Switch VLAN configuration instructions.
Note: It is mandatory that you connect C11-T3 to a campus connection or run an additional management connection. If you do not do this step, you will lose the connection to the EMS node when the container starts.

ESS Legacy

POWER8 or POWER9 EMS supported

POWER8 EMS must have the following connections:
  • C10-T1 to the management VLAN
  • C10-T4 to the FSP/Service VLAN
  • C10-T2 to the management VLAN
  • C10-T3 optional campus connection
  • HMC1 to the FSP/Service VLAN
POWER9 EMS must have the following connections:
  • C11-T1 to the management VLAN
  • C11-T2 to the FSP VLAN
  • HMC1 to the FSP VLAN
  • C11-T3 to the campus or management network/VLAN
POWER8 nodes:
  • C12-T1 to the management VLAN
  • HMC1 to the FSP VLAN
Start of change

Code version

Start of changeEnd of change

Note:
  • The versions shown here might not be the GA version available on IBM FixCentral. It is recommended to go to IBM FixCentral and download the latest code.
  • ppc64le in the package name implies that each container runs on a POWER®-based EMS. For details about functions supported by respective containers, see Support matrix.
You can download the latest 6.1.x.x code (Start of change6.1.2.8End of change is the latest) from IBM Fix Central by using the following link.
There are (4) containers offered with (2) versions each (Data management + Data access). Example package names for each container are as follows:Start of change
ESS_DME_BASEIMAGE-6.1.2.8-x86_64-3200-Linux.tgz
ESS_DAE_BASEIMAGE-6.1.2.8-x86_64-3200-Linux.tgz
ESS_DME_BASEIMAGE-6.1.2.8-ppc64LE-5000-Linux.tgz
ESS_DAE_BASEIMAGE-6.1.2.8-ppc64LE-5000-Linux.tgz
ESS_DME_BASEIMAGE-6.1.2.8-x86_64-3000-Linux.tgz
ESS_DAE_BASEIMAGE-6.1.2.8-x86_64-3000-Linux.tgz
ESS_DME_BASEIMAGE-6.1.2.8-ppc64LE-Legacy-Linux.tgz
ESS_DAE_BASEIMAGE-6.1.2.8-ppc64LE-Legacy-Linux.tgz
End of change
Note: The container installs and runs on the EMS only. The EMS supported is Power-based only. Running container on a x86-based node is not supported as of now.
End of change

POWER8 considerations

If you are moving from an xCAT-based release (5.3.x) to a container-based release (6.1.x.x), the following considerations apply:
  • You must add an additional management network connection to C10-T2.
  • A public or additional management connection is mandatory in C10-T3.
  • Start of changeYou must stop and uninstall xCAT and all xCAT dependencies before installing the container. End of change
Start of change

Remote management considerations

Data center access has become more restrictive nowadays. Here are some considerations to enable remote support:
  • Always add a campus connection to the EMS (POWER8 and POWER9).
  • Consider adding campus connections to the HMC2 ports on all POWER servers (ESS Legacy, ESS 5000, POWER8 or POWER9 EMS). Consider cabling this port to a public network and setting a campus IP. This will allow remote recovery or debug of the EMS in case of an outage.
  • Consider adding campus connections to C11-T3 (POWER9 nodes) or C10-T3 (POWER8 nodes).
  • Consult with service about adding USB to Ethernet dongle to enable campus connections on the ESS 3200 system.
  • Add campus connection to a free port on each ESS 3000 canister. Also consider adding SMART PDUs on ESS 3000 frames to help remotely power cycle the system.
End of change

POWER8 + POWER9 considerations

  • If both POWER8 and POWER9 EMS nodes are in an environment, it is recommended that you use only the POWER9 EMS for management functions (containers, GUI, ESA, collector).
  • Only a single instance of all management services is recommended and solely on the POWER9 EMS.
  • POWER8 only needs to exist as a management node if you are mixing a non-container-based release (5.3.x) with a container-based release (6.x.x.x).
  • It is recommended that all nodes in the storage cluster contain the same ESS release and IBM Spectrum Scale version.
  • It is recommended that you upgrade to the latest level before adding a building block.
Note: If you are mixing ESS Legacy 5.3.x and ESS 3000 on a POWER8 EMS, the following considerations apply:
  • You cannot upgrade the EMS node from the ESS 3000 container.
  • ESS 3000 detects if xCAT is installed on the host EMS node. If xCAT is installed, it stops the upgrade.
  • You must upgrade the EMS node by using the legacy deployment procedure outlined in ESS 5.3.x Quick Deployment Guide.

Migrating from an ESS Legacy environment (xCAT-based 5.3.x) to an ESS Legacy container-based environment (6.1.x.x)

The following guidance is for customers migrating from an xCAT-based release to a container-based release for POWER8 offerings.
POWER9 EMS
You cannot run both POWER8 and POWER9 EMS nodes in the same environment for ESS Legacy. If you are moving a POWER9 EMS, migrate all services from the POWER8 EMS and uninstall xCAT. You can then re-use the POWER8 EMS for other purposes such as quorum node, client node, or spare EMS. The preference is to always use a POWER9 EMS if possible and you must not run multiple instances of GUI, performance monitoring collectors, etc. in the same cluster. For this requirement, there are exceptions for certain stretch cluster environments and if you are mixing ESS Legacy and container-based deployments such as ESS 5.3.7 on POWER8 and ESS 6.0.2.x on POWER9.
POWER8 EMS
If you are migrating from ESS 5.3.x to ESS 6.1.0.x on a POWER8 EMS, do the following steps.
  1. Stop and uninstall xCAT by doing the following steps on a POWER8 EMS, outside of the container.
    1. Stop xCAT.
      systemctl stop xcatd
    2. Uninstall xCAT.
      yum remove xCAT*
    3. Remove dependencies.
      yum remove dbus-devel dhcp bind java-1.8.0-openjdk
  2. Add a container connection to C10-T2.
  3. Start of changeAdd a campus connection to C10-T3, if it is not done already.End of change
  4. Update /etc/hosts with the desired container host name and IP address.

Other notes

  • The following tasks must be complete before starting a new installation (tasks done by manufacturing and the SSR):
    • SSR has ensured all hardware is clean, and IP addresses are set and pinging over the proper networks (through the code 20 operation).
    • /etc/hosts is blank.
    • The ESS tgz file (for the correct edition) is in the /home/deploy directory. If upgrade is needed, download from Fix Central and replace.
    • Network bridges are cleared.
    • Images and containers are removed.
    • SSH keys are cleaned up and regenerated.
    • All code levels are at the latest at time of manufacturing ship.
  • Customer must make sure that the high-speed connections are cabled and the switch is ready before starting.
  • All node names and IP addresses in this document are examples.
  • Changed root password should be same on each node, if possible. The default password is ibmesscluster. It is recommended to change the password after deployment is completed.
  • Each server's IPMI and ASMI passwords (POWER nodes only) are set to the server serial number. Consider changing these passwords when the deployment is complete.
  • Start of changeCheck whether the SSSD service is running on EMS and other nodes. Shut down the SSSD service on those nodes manually, before you upgrade the nodes.End of change

ESS best practices

  • ESS 6.x.x.x uses a new embedded license. It is important to know that installation of any Red Hat packages outside of the deployment upgrade flow is not supported. The container image provides everything required for a successful ESS deployment. If additional packages are needed, contact IBM for possible inclusion in future versions.
  • For ESS 3000, consider enabling TRIM support. This is outlined in detail in IBM Spectrum Scale RAID Administration. By default, ESS 3000 only allocates 80% of available space. Consult with IBM development, if going beyond 80% makes sense for your environment, that is if you are not concerned about the performance implications due to this change.
  • You must setup a campus or additional management connection before deploying the container.
  • If running with a POWER8 and a POWER9 EMS in the same environment, it is best to move all containers to the POWER9 EMS. If there is a legacy PPC64LE system in the environment, it is best to migrate all nodes to ESS 6.1.x.x and decommission the POWER8 EMS altogether. This way you do not need to run multiple ESS GUI instances.
  • If you have a POWER8 EMS, you must upgrade the EMS by using the legacy flow if there are xCAT based PPC64LE nodes in the environment (including protocol nodes). If there are just an ESS 3000 system and a POWER8 EMS, you can upgrade the EMS from the ESS 3000 container.
  • If you are migrating the legacy nodes to ESS 6.1.x.x on the POWER8 EMS, you must first uninstall xCAT and all dependencies. It is best to migrate over to the POWER9 EMS if applicable.
  • You must be at ESS 5.3.7 (Red Hat Enterprise Linux 7.7 / Python3) or later to run the ESS 3000 container on the POWER8 EMS.
  • You must run the essrun config load command against all the storage nodes (including EMS and protocol nodes) in the cluster before enabling admin mode central or deploying the protocol nodes by using the installation toolkit. For more information, see Deploying protocols.
  • If you are running a stretch cluster, you must ensure that each node has a unique hostid. The hostid might be non-unique if the same IP addresses and host names are being used on both sides of the stretch cluster. Run gnrhealthcheck before creating recovery groups when adding nodes in a stretch cluster environment. You can manually check the hostid on all nodes as follows:
    mmdsh -N { NodeClass | CommaSeparatedListofNodes } hostid

    If hostid on any node is not unique, you must fix by running genhostid. These steps must be done when creating a recovery group in a stretch cluster.

  • Consider placing your protocol nodes in file system maintenance mode before upgrades. This is not a requirement but you should strongly consider doing it. For more information, see File system maintenance mode.
  • Do not try to update the EMS node while you are logged in over the high-speed network. Update the EMS node only through the management or the campus connection.
  • After adding an I/O node to the cluster, run the gnrhealthcheck command to ensure that there are no issues before creating vdisk sets. For example, duplicate host IDs. Duplicate host IDs cause issues in the ESS environment.
  • Run the container from a direct SSH connection. Do not SSH from an I/O node or any node that might be rebooted by the container.
  • Do not log in and run the container over the high-speed network. You must log in through the campus connection.
  • Start of changeYou must stop Spectrum Scale tracing (mmtrace | mmtracectl) before starting the container or deploying any node. The container attempts to block if tracing is detected, it is recommended to manually inspect each ESS node before attempting to deploy.End of change

Support notes and rules

  • Multiple EMS nodes are not supported in the same cluster. If you are adding a POWER9 EMS to an existing cluster run by a POWER8 EMS, the POWER9 EMS must be the only one used for management functions such as GUI, performance monitoring collector, etc.
  • Multiple GUI instances are not supported in the same cluster.
  • One collector node must be run at a time in the cluster. This must be on the same node as the GUI.
  • You cannot mix majoresagent.pLinux-4.5 IBM Spectrum Scale versions in the storage cluster. All nodes must be updated to the latest level.
  • ESA must be running on the EMS.
  • Start of changeYou can run call home on the EMS.End of change
  • If possible, run the client nodes in a separate cluster than the storage.
  • The essrun (ESS deployment Ansible wrapper tool run within the container) tool does not use the GPFS admin network. It uses the management network only to communicate from the container to each of the nodes.
  • If POWER8 EMS only, consolidate potential xCAT and non-xCAT offerings to container versions.

    Example: If you have ESS 5.3.7.x (Legacy POWER8 offering on Scale 5.0.5.x) and ESS 3000 (Containerized support for ESS 3000 on Scale 5.x.x.x and above), convert the Legacy 5.3.7.x to 6.1.x.x so that only containers are running on POWER8 EMS.

    Note: This only applies to situations where there was already Scale 5.1.x.x+ in the environment.
    Note: There is no container offering for BE so environments with BE would have to remain at 5.0.5 release level (but the POWER8 EMS could still move to all container version).
  • If POWER8 EMS and POWER9 EMS are owned by the customer, it is recommended to consolidate to POWER9 EMS (all container versions).

    Example: If POWER8 EMS was running 5.1.x.x (ESS 3000, ESS Legacy or both) and customer has a POWER9 EMS (running ESS 5000 or ESS 3200) then should migrate the containers from POWER8 EMS to POWER9 and discard the POWER8 EMS (single management node).

  • If migrating from xCAT-based legacy offering to container based you must go from ESS 5.3.7.x.
  • Start of changeWhen you update ESS to 6.1.2.x for the first time, you must consider the implications of moving to MOFED 5.x. Review this flash carefully for more informationMellanox OFED 5.x considerations in IBM ESS V6.1.2.x.End of change
  • Spectrum Fusion, Spectrum Scale Container Native, and Spectrum Scale CSI all utilize the GUI rest-api server for provisioning of storage to container applications. Persistent Volume (PV) provisioning will halt when the ESS GUI is shut down and remain halted for the duration of the ESS upgrade, until the GUI is restarted. Ensure that the OpenShift and Kubernetes administrators are aware before proceeding.
Start of change

Client nodes

Client nodes need to be at MOFED 4.9.x or higher and converted to verbsRDMA core libs after the ESS cluster is moved to 6.1.2.x or higher. Moving to verbsRDMA core libs is especially important if verbsRDMA is in use in the storage cluster.

End of change
Start of change

Upgrade guidance

ESS upgrade chart
Note: Start of change
  • Upgrades to ESS 6.1.2.x follow the N-2 rule. You can upgrade from ESS 6.1.1.x, 6.1.2.x (that is, 6.1.1.2) or 6.1.0.x.
  • Start of changeYou can upgrade from ESS 6.1.2.5 and 6.1.2.6 to ESS 6.1.2.8 only. But you cannot upgrade from ESS 6.1.2.8 to ESS 6.1.3.x or higher versions.End of change
End of change
End of change
Start of change

Further legacy container migration guidance

You must migrate first to ESS 5.3.7.x before you upgrade to ESS 6.1.x.x (container version).

ESS 5.3.x.x upgrade guidance
    • You can upgrade to 5.3.7.x from 5.3.5.x (online) or 5.3.6.x (online).
    • For online upgrade you can jump one OS version and for offline upgrade you can jump two OS versions.

      Only exception is RHEL 7.7 to RHEL 7.9 upgrade. Because there is no RHEL 7.8.

      Online upgrade to RHEL 7.7 from RHEL 7.6 can be done.

      Upgrade to RHEL 7.7 from RHEL 7.5 must be done online.

ESS 6.1.x.x upgrade guidance
It is recommended to convert from ESS 5.3.7.x to ESS 6.1.2.x and follow the normal N-X rules. To convert to ESS 6.1.2.x, use the following table (based on the RHEL 7.9 kernel):
Table 1. RHEL kernels
ESS Kernel
6.1.2.8 3.10.0-1160.99.1.el7
6.1.2.7 3.10.0-1160.92.1.el7
6.1.2.6

3.10.0-1160.83.1.el7

4.18.0-372.41.1.el8_6

6.1.2.5 3.10.0-1160.76.1.el7
6.1.2.4 3.10.0-1160.49.1.el7
6.1.2.3 3.10.0-1160.62.1.el7
6.1.2.2 3.10.0-1160.71.1.el7
5.3.7.6 3.10.0-1160.62.1.el7
5.3.7.5 3.10.0-1160.59.1.el
5.3.7.4 3.10.0-1160.49.1.el7
5.3.7.3 3.10.0-1160.45.1.el7
5.3.7.2 3.10.0-1160.31.1
5.3.7.1 3.10.0-1160.24.1
5.3.7.0 3.10.0-1160.11.1.el7
An example of upgrade jump is as follows:
  • To upgrade to ESS 6.1.2.2, you can only upgrade from 5.3.7.4 or lower versions (that is, less than equal to 5.3.7.4).
  • To upgrade to ESS 6.1.2.3, you can only upgrade from 5.3.7.6 or lower versions.

It is not recommended to upgrade from ESS 5.3.7.x to ESS 6.1.1.2 anymore. Upgrade directly to ESS 6.1.2.x.

You may need to modify the container unblock jumps from a specific 5.3.7.x level. Issue to the following command to upgrade the ESS level in the container:
vim /opt/ibm/ess/deploy/ansible/vars.yml

Change (an example if you want to convert from ESS 5.3.7.1 or higher) LEGACY_SUPPORTED_VERSION: "5.3.7.3" to LEGACY_SUPPORTED_VERSION: "5.3.7.1".

End of change