Installation and upgrade related information and checklists

Review the following installation and upgrade related information before starting with the installation or the upgrade of Elastic Storage Server (ESS).

New features and enhancements

Release Changes
5.3.7.6
  • Support for IBM Spectrum Scale 5.0.5 PTF14
  • New Kernel 3.10.0-1160.62.1.el7.ppc64le
Start of change5.3.7.5End of change Start of change
  • Support for IBM Spectrum Scale 5.0.5 PTF13
  • No support for the BE architecture
  • New Kernel 3.10.0-1160.59.1.el7.ppc64le
  • New POWER firmware level SV860_240 (FW860.B0)
  • New Systemd 219-78.el7_9.5
  • New firmware rpm gpfs.ess.firmware-6.0.0-17.ppc64le.rpm
  • New Mellanox OFED MLNX_OFED_LINUX-4.9-4.1.7.2
  • New ESA level esagent.pLinux-4.5.-7.1.noarch.rpm
  • New IPR 19512c00
  • New 3200 support rpm (gpfs.gnr.support-ess3200-1.0.0-2.noarch.rpm)
  • New firmware SV860_240 (FW860.B0)
End of change
5.3.7.4
  • Support for IBM Spectrum Scale 5.0.5 PTF11
  • New Kernel 3.10.0-1160.49.1.el7.ppc64le
  • New POWER firmware level SV860_236 (FW860.A2)
  • New xCAT 2.16.3
  • New Mellanox OFED MLNX_OFED_LINUX-4.9-4.1.1.1
  • New ESA level esagent.pLinux-4.5.-6.1.noarch.rpm
5.3.7.3
  • Support for IBM Spectrum Scale 5.0.5 PTF10 efix1
  • New Kernel 3.10.0-1160.45.1.el7.ppc64le
  • New firmware rpm gpfs.ess.firmware-6.0.0-15.ppc64le
ESS 5.3.7.2
  • Support for IBM Spectrum Scale 5.0.5 PTF8
  • New Kernel version (3.10.0-1160.31.1)
  • New Mellanox OFED version (MLNX_OFED_LINUX-4.9-3.1.5.3)
  • New Power 8 FW SV860_231_165 (FW860.A0)
  • New HMC version V9R2M951
Start of changeESS 5.3.7.1End of change Start of change
  • Support for IBM Spectrum Scale 5.0.5 PTF7 efix1
  • New Kernel version (3.10.0-1160.24.1)
  • New Systemd version (219-78.el7_9.3)
  • Fixed issue on PPC64BE wherein gssinstallcheck might display garbage characters for HBA/Phy check.
  • Fixed issue when correcting the KVM memory issue.
  • Fixed issue with PPC64LE where the mpt3sas driver might not upgrade to 34.
  • Fixed issue wherein additional support RPMs were not installed on PPC64LE EMS node.
  • General fixes and improvements
End of change
Start of changeESS 5.3.7End of change Start of change
  • Support for IBM Spectrum Scale 5.0.5 PTF6
  • Support for Red Hat Enterprise Linux®: 7.9 (PPC64BE and PPC64LE)
  • New Kernel version (3.10.0-1160.11.1)
  • New POWER firmware version (FW860.90 (SV860_226))
  • New OFED version (MLNX_OFED_LINUX-4.9-2.2.5.1)
  • New ESA version (4.5.5-1)
  • New mpt3sas version (34.00.00.00)
  • New IPR version (19512b00)
End of change
ESS 5.3.6.2
  • Support for IBM Spectrum Scale 5.0.5 PTF4
  • New Kernel (3.10.0-1160.6.1)
  • New Systemd (219-78.el7_9.2)
  • New Network manager (1.18.8-2.el7_9)
ESS 5.3.6.1
  • Support for IBM Spectrum Scale 5.0.5 PTF3
  • New Kernel (3.10.0-1127.19.1.el7)
  • New Systemd (219-67.el7_7.10)
  • mpt3sas version (31.100.01.00)
  • New GNR firmware RPM
ESS 5.3.6
  • Support for IBM Spectrum Scale 5.0.5 PTF1 efix3
  • New Kernel (3.10.0-1062.26.1.el7)
  • New Systemd (219-67.el7_7.6)
  • New OFED (MLNX_OFED_LINUX-4.9-0.1.7.3)
  • New IPR (19512900)
  • New Power 8 System Firmware (FW860.81 (SV860_215))
  • New ESA (4.5.5-0)
  • New xCAT (2.15.1)
  • Deployment fixes and improvements
ESS 5.3.5.2
  • Support for IBM Spectrum Scale 5.0.4 PTF3 + efix 2
  • New Kernel (3.10.0-1062.12.1 )
  • Updated systemd (219-67.el7_7.4)
  • Updated network manager (1.18.0-5.el7_7.2)
  • Deployment bug fixes
ESS 5.3.5.1
  • Support for IBM Spectrum Scale 5.0.4 PTF2 + efix 5
  • Support for Red Hat Enterprise Linux 7.7
  • New Kernel on top of RHEL 7.7 (3.10.0-1062.9.1 )
  • Support for MLNX_OFED_LINUX-4.7-3.2.9.1
  • Support for IPR Firmware 19512600
  • Support for new Power 8 Firmware SV860_212 (FW860.80)
  • Support for new HMC version V9R1M940_SP0 (PPC64BE only)
  • Support for ESA agent 4.5.3-2
ESS 5.3.5
  • Support for IBM Spectrum® Scale 5.0.4 PTF1 + efix
  • Updated kernel, systemd, network manager
  • Support for ESA agent 4.5.2-1
  • Support for IPR Firmware 19512300
  • Support for new Power® 8 Firmware SV860_205 (FW860.70)
  • Support for new HMC version V9R1M920_SP0 (PPC64BE only)
  • Support for IBM Elastic Storage® System 3000 (ESS 3000)
  • Support for GL3C and GL3S
  • Support for MES GL2S -> GL3S -> GL4S
  • Support for Power 8 (PPC64LE) node call home
  • Enhanced essutils (formerly gssutils) support for SSRs
ESS 5.3.4.2
  • Support for IBM Spectrum Scale 5.0.3 PTF3
  • Support for Mellanox OFED 4.6.3
  • Support for Red Hat Enterprise Linux 7.6
  • Updated kernel, systemd, network manager
  • Support for xCAT 2.14.6
  • Support for ESA agent 4.5.1-1
  • Support for IPR Firmware 19512200
  • Support for new Power 8 Firmware SV860_205 / FW860.70
ESS 5.3.4.1
  • New Mellanox OFED (4.6-3) which reinstates support on Connect-IB adapters
  • New drive and enclosure firmware updates, which resolves certain stability and enclosure fan speed issues
  • New patch, which resolves PPC64LE serviceable events from disappearing from the OPAL log
  • Updated ESA RPM, which fixes a few Call Home issues
  • Updated IBM Spectrum Scale (5.0.3 PTF2 + efix4)
  • Updated Kernel version
  • Minor ESS deployment bug fixes and improvements
ESS 5.3.4
  • Support for new security features
    • Sudo
    • SELinux
    • Firewall
    • Admin mode central
    • Security certificate support in call home setup
  • Support for Red Hat Enterprise Linux 7.6
  • Support for IBM Spectrum Scale 5.0.3 PTF1
  • Support for new xCAT version (2.14.6)
  • Support for new IPR version (19512200)
  • Support for new systemd version (219-62.el7_6.6)
  • Support for new Network Manager version (1.12.0-10.el7_6)
  • Support for new MOFED version (4.6-1.0.1.2)
  • Support for new kernel version (3.10.0-957.12.2)
  • Support for new ESA version (4.5.1-0)
  • mmvdisk changed to default deployment method
  • Deployment support for Broadcom adapters
  • Support for new MES options (GLxC Support)
    • GL1C to GL2C
    • GL2C to GL4C
    • GL4C to GL5C
  • Deployment improvements and bug fixes

Component versions for this release

Note: Your version might be slightly different from the version indicated in this document. Refer to the release notes document that comes with the installation package for the exact version of the installation package and the component version.
The respective versions for the core components in this release of ESS are as follows:
  • Supported architecture: PPC64LE
  • Start of changeIBM Spectrum Scale: 5.0.5.14End of change
  • Start of changexCAT: 2.16.3End of change
  • HMC: V9R2M951
  • Start of changeSystem firmware: SV860_240 (FW860.B0) End of change
  • Start of changeRed Hat Enterprise Linux: 7.9 (PPC64LE)End of change
  • Start of changeKernel: 3.10.0-1160.62.1.el7.ppc64leEnd of change
  • Start of changeSystemd: 219-78.el7_9.5End of change
  • Network Manager: 1.18.8-2.el7_9
  • Start of changempt3sas: 34.00.00.00End of change
  • Start of changempt2sas = 20.00.04.00End of change
  • Start of changeIPR: 19512c00 End of change
  • Start of changeSAS adapter driver: 16.00.11.00End of change
  • Support RPM:
    • gpfs.gnr.support-ess3000-1.0.0-3.noarch.rpm
    • gpfs.gnr.support-essbase-1.0.0-3.noarch.rpm
    • gpfs.gnr.support-ess5000-1.0.0-3.noarch.rpm
    • gpfs.gnr.support-ess3200-1.0.0-2.noarch.rpm
  • Start of changeFirmware RPM: gpfs.ess.firmware-6.0.0-23.ppc64le.rpmEnd of change
  • Start of changeESA: 4.5.7-0End of change
  • Enclosure firmware:
    • PPC64LE:
      • 2U24 = 4230
      • 5U84 = 4087
      • 4U106 = 5266
  • Start of changeOFED: MLNX_OFED_LINUX-4.9-4.1.7.2
    Start of changeOFED firmware levels:
    • MT27500 = 10.16.1200
    • MT4099 = 2.42.5000
    • MT26448 = 2.9.1326
    • MT4103 = 2.42.5000
    • MT4113 = 10.16.1200
    • MT4115 = 12.28.2006
    • MT4117 = 14.31.1014
    • MT4119 = 16.31.1014
    • MT4120 = 16.31.1014
    • MT4121 = 16.31.1014
    • MT4122 = 16.31.1014
    End of change
    End of change

Supported editions on each architecture

The following ESS edition is supported on the available architecture.
PPC64LE
  • Data Access Edition
  • Data Management Edition

ESS best practices and support statements

  • Start of changeIf you are running a stretch cluster, you must ensure that each node has a unique hostid. The hostid might be non-unique if the same IP addresses and host names are being used on both sides of the stretch cluster. Run gnrhealthcheck before creating recovery groups when adding nodes in a stretch cluster environment. You can manually check the hostid on all nodes as follows:
    mmdsh -N { NodeClass | CommaSeparatedListofNodes } hostid

    If hostid on any node is not unique, you must fix by running genhostid. These steps must be done when creating a recovery group in a stretch cluster.

    End of change
  • versionlocks are not enabled in ESS. In the past, versionlocks were used to protect against unwarranted kernel and OFED updates. Although versionlocks no longer exist, the same rules apply regarding ESS packages verified by gssinstallcheck or gssinstall. You must get specific approval and guidance from the L2 Service to make any changes to an ESS configuration. The only exceptions are RHEL packages in addition to kernel, systemd, and network manager packages. RHEL packages other than the mentioned packages are customer responsibility and they might be updated for security purposes.
  • If you are upgrading to ESS 5.3.7.6, you must convert the environment to mmvdisk after the upgrade is completed.
  • It is advised that you set autoload to on to enable GPFS to recover automatically in case of a daemon problem. Deployment automatically enables this on new installations but you should disable autoload for an upgrade and re-enable it after an upgrade.
    To disable, issue the following command:
    mmchconfig autoload=no
    Once the maintenance operation or upgrade is complete, re-enable autoload.
    mmchconfig autoload=yes
  • Do not mount the file system on the ESS I/O server nodes.
  • It is advised that you disable automount for file systems when performing an upgrade to ESS 5.3.1 or later.
    mmchfs Device -A no

    Device is the device name of the file system.

    Automount should automatically be disabled when creating new file systems with gssgenvdisks.
    Remember: Mount the file system only on the EMS node where the GUI and the PM collector run.
  • Do not configure more than 5 failure groups in a single file system.
  • Consider moving all supported Infiniband devices to the Datagram mode (CONNECTED_MODE=no). For more information, see ESS networking considerations.
  • Running any additional service or protocols on any ESS node is not supported. This includes installing any additional RPMs, running any protocols (or any other type of service), or mounting the file system on any ESS I/O server node. This also applies to the EMS node, although you must mount the file system to support the IBM Spectrum Scale GUI.
  • RoCE (RDMA over Ethernet) is not supported in ESS.
  • Consider moving quorum, cluster, and file system management responsibilities from the ESS nodes to other server license nodes within the cluster.
  • It is not required, though highly recommended, that the code levels match during a building block addition. Be mindful of changing the release and file system format in mixed IBM Spectrum Scale environments.
  • You must take down the GPFS cluster to run firmware updates in parallel.
  • Do not independently update IBM Spectrum Scale (or any component) on any ESS node unless specifically advised from the L2 service. Normally this is only needed to resolve an issue. Under normal scenarios it is advised to only upgrade in our tested bundles.
  • It is acceptable for LBS or customers to update any security errata available from Red Hat Network (RHN). Only components checked and protected by ESS (for example, kernel, network manager, systemd) must not be modified unless advised by the IBM® service. For more information on applying security erratas see https://access.redhat.com/solutions/10021
  • Client node deployment is not supported from the ESS management node.
  • You must deploy or add building blocks from an EMS with the same architecture. There must be a dedicated EMS for the architecture (PPC64LE).
  • If running in a mixed architecture environment, the GUI and PM collector are recommended to run on the PPC64LE EMS node.
  • Modifying any ESS nodes as a proxy server is not supported.
  • Multiple building blocks are ideal as ESS now by default uses file system level metadata replication. If a single building block is used, by default gssgenvdisks uses one failure group and only IBM Spectrum Scale RAID level metadata replication.
  • It is recommended to use the highest available block size when creating vdisks or NSDs. The default block size is 16M (current maximum). If the customer primarily generates many tiny files (metadata heavy), consider splitting metadata and data NSDs and using smaller block sizes.
  • It is recommended that all nodes in a cluster run the same version of Mellanox OFED.
  • Automatic EMS failover is not supported. For help in setting up a redundant, standby EMS, contact the L2 service.
  • 4K MTU (InfiniBand) and 9000 MTU (Ethernet) are recommended. Changing to these MTU values requires associated switch-side changes.
  • Stretch clusters are supported in various configurations. Contact development or service for guidance.
  • If using a PPC64LE building block (8247), note that HMC is not used in that configuration. HMC is applicable for PPC64BE only.
  • Connect X-2 (ConnectX-EN) adapters are still supported by ESS.

Obtaining the required Red Hat Enterprise Linux and ESS code

Note: Contact IBM to obtain access to an online folder containing the required items for deployment.
The required Red Hat components and SHA256 are:
  • Start of changeRed Hat Enterprise Linux 7.9 ISO
    9893045dbb02ed9439bf571d24202935d37e74de74a07aa65a2827c3bc193335  rhels-7.9-server-ppc64le.iso
    
    End of change
  • Network manager version : 1.18.8-2.el7_9Start of change
    5e3b54a031f0bd9f04c89ebfbfe722c797ba7175c9deeaa6bf523a8192ea51e2  netmgr_5376_LE.tgz
    
    End of change
  • Start of changeSystemd version: 219-78.el7_9.5Start of change
    14f289572e7f7d35fd8452d4e6eff83918a8c260d9ad9b63d474aabe53b59883  systemd_5376_LE.tgz
    
    End of changeEnd of change
  • Start of changeKernel version: 3.10.0-1160.62.1
    af8eba6b1588e3dc1d2e1e424351fc102590f24a0fc7d396c9381eaf66f8d909   kernel_5376_LE.tgz
    
    End of change
  • Power 8 OPAL patch
    ea9c602234f446f009009eaba9634f40750da4071523d0e2c59dc646a35a2766  opal-patch-le.tar.gz
    

On ESS 5.3.7.6 systems shipped from manufacturing, these items can be found on the management server node in the /home/deploy directory.

Customers or business partners can download the required Red Hat components from Red Hat Network using the customer license. For more information, see:
The ESS software archive that is available in different versions for PPC64LE architecture.
Available PPC64LE packages:
ESS_DA_BASEIMAGE-5.3.7.6-ppc64le-Linux.tgz
ESS_DM_BASEIMAGE-5.3.7.6-ppc64le-Linux.tgz

ESS 5.3.7.6 can be downloaded from IBM FixCentral.

Once downloaded and placed in /home/deploy, untar and uncompress the package to view the contents. For example, for the Data Access edition PPC64LE package, use the following command:
tar -zxvf ESS_DA_BASEIMAGE-5.3.7.6-ppc64le-Linux.tgz
For example, from the BASEIMAGE tar file, files such as the following get extracted with the preceding command:
  • ESS_5.3.7.6_ppc64le_Release_note_Data_Access.txt: This file contains the release notes for the latest code.
  • gss_install-5.3.7.6_ppc64le_dataaccess_20220503T134712Z.tgz: This .tgz file contains the ESS code.
  • gss_install-5.3.7.6_ppc64le_dataaccess_20220503T134712Z.sha256sum: This .sha256 file to check the integrity of the tgz file.

Support for signed RPMs

ESS or IBM Spectrum Scale RPMs are signed by IBM.

The PGP key is located in /opt/ibm/gss/tools/conf
-rw-r-xr-x 1 root root 907 Dec 1 07:45 SpectrumScale_public_key.pgp
You can check if an ESS or IBM Spectrum Scale RPM is signed by IBM as follows.
  1. Import the PGP key.
    rpm --import  /opt/ibm/gss/tools/conf/SpectrumScale_public_key.pgp
  2. Verify the RPM.
    rpm -K RPMFile

Supported upgrade paths

Note:
A rule of thumb is that you can hop one OS level at a time. For example:
  • RHEL 7.5 -> RHEL 7.6 upgrade can be done in one hop
It is recommended that if you are doing an offline upgrade, it is safe to hop two OS releases (RHEL 7.4 -> RHEL 7.6). If you are doing an online upgrade, it is advised to do only one OS hop at a time (RHEL 7.5 -> RHEL 7.6). Review the GNR FAQ to see which ESS releases support the various OS levels.

Mixed environment recommendations

Running ESS building blocks of mixed levels is not recommended. If you choose to do so, following recommendations apply:
  • Nodes within a building block must be at the same levels.
  • Nodes between building blocks should not be greater than N-2 (OFED 4.4 and OFED 4.6, for example).

ESS 3000 considerations (POWER8 EMS)

  • If your system came racked with EMS but no ESS 3000 (Any supported legacy node and an EMS node are in the order):

    Use the installation flow in this document (gsschenv)

  • If any other configuration but no ESS 3000:

    Refer to the Legacy deployment instructions.

  • If your system comes in any configuration with ESS 3000, refer to the IBM ESS 3000 Version 6.0.1.x documentation.
  • Your EMS node must be at version 5.3.5 or later, with podman installed and C10-T2 connection, to support ESS 3000
  • You do not have to upgrade to ESS 5.3.6 to support ESS 3000.
  • If EMS + ESS 3000:

    The minimum configuration is EMS at version 5.3.5 with podman + C10-T2 connection + IBM Spectrum Scale 5.0.5.1 (Updated from container).

  • If EMS + ESS 3000 + ESS:

    It is advised to upgrade the EMS and ESS first to version 5.3.7.6. This is not a hard requirement, but the version must be 5.3.5 or later.

Security law changes

  • New systems and switches shipped from manufacturing now have either an expired password or one set to the serial number of the component.
  • You must take input from the customer before deployment starts and change the desired passwords.
  • The default root password for the OS is ibmesscluster. You are required to change it upon first login. This password must be set the same on each node.
  • The default ASMI passwords (login, IPMI, HMC, etc.) are set to the serial number of the server. IPMI must be the same on each node.
  • If the 1Gb Cumulus switch is shipped racked, the default password is the serial number (S11 number - label found on the back of the switch). If the switch is shipped unracked, you are required to set the password upon first login. The default password is CumulusLinux! but you will be prompted to change the password upon first login. If you have any issues logging in or you need help in setting up a VLAN with the switch, consult this documentation link.
  • You must set all required passwords before the deployment begins.

Support for hardware call home

For more information, see Configuring call home in ESS 5.3.x.

Pre-installation or pre-upgrade checklist

Before you arrive at a customer site or before upgrade, it is advised that you perform the following tasks:

Post-installation or post-upgrade checklist

After the installation or upgrade is completed, it is advised that you verify the following:

Other topics

For help with the following topics, and many others that are unlisted, contact L2 Service.
  • Restoring a management server
  • Part upgrades or replacements
  • VLAN reconfiguration on the 1Gb switch
  • Extending the 1Gb Cumulus management switch
  • Stretch cluster considerations

Sample installation and upgrade flow

New installations go through manufacturing CSC. The system is fully installed with ESS 5.3.7.6, tested, malfunctioning parts replaced, and required RHEL pieces shipped in /home/deploy.

Installation

  • SSR checkout complete
  • LBS arrival on site
  • Plug-n-Play mode demonstrated
  • Decisions made on file system names and sizes, block size, host names, IP addresses, and so on
  • Check high-speed switch settings and firmware
  • Deploy EMS and building block
  • Network bonds created
  • Cluster created
  • Recovery groups, NSDs, file system created
  • Stress test performed
  • Final checks performed
Proceed to install the protocol nodes, if applicable.
  • GUI setup (w/SNMP alerts if desired)
  • Call home setup
  • Nodes attached to RHN and security updates applied

Upgrade

  • Check high speed switch settings and firmware
  • Ensure that there are no hardware issues
  • Ensure client / protocol node compatibility
  • Ensure no heavy I/O operations are being performed
  • Upgrade ESS (rolling upgrade or with cluster down)
    • Always ensure you have quorum (if rolling upgrade)
    • Always carefully balance the recovery groups and scale management functions as you upgrade each node (if rolling upgrade)
  • Move the release level and the file system format, if applicable. Move the Recovery Group format to LATEST.
  • Final checks are performed
  • If applicable, upgrade the ESS protocol nodes
  • Ensure that call home and GUI are still working as expected
  • Use yum to upgrade any security related errata (yum -y security). Do not update any kernel, systemd, or network manager erratas.