ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide...

42
Elastic Storage Server Version 4.0 Quick Deployment Guide SC27-8580-01 IBM

Transcript of ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide...

Page 1: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Elastic Storage ServerVersion 4.0

Quick Deployment Guide

SC27-8580-01

IBM

Page 2: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM
Page 3: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Elastic Storage ServerVersion 4.0

Quick Deployment Guide

SC27-8580-01

IBM

Page 4: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

NoteBefore using this information and the product it supports, read the information in “Notices” on page 21.

This edition applies to version 4.0.x, 3.5, and 3.0 of the Elastic Storage Server (ESS) for Power®, and to allsubsequent releases and modifications until otherwise indicated in new editions.

IBM welcomes your comments; see the topic “How to submit your comments” on page viii. When you sendinformation to IBM, you grant IBM a nonexclusive right to use or distribute the information in any way it believesappropriate without incurring any obligation to you.

© Copyright IBM Corporation 2016.US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contractwith IBM Corp.

Page 5: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Contents

Tables . . . . . . . . . . . . . . . v

About this information . . . . . . . . viiWho should read this information. . . . . . . viiPrerequisite and related information . . . . . . viiConventions used in this information . . . . . viiiHow to submit your comments . . . . . . . viii

Deploying the Elastic Storage Server -for experienced users. . . . . . . . . 1Install the ESS system . . . . . . . . . . . 2Upgrade the ESS system . . . . . . . . . . 7

Appendix A. Known issues . . . . . . 13

Appendix B. Instructions for installingthe ESS Red Hat Linux Errata KernelUpdate . . . . . . . . . . . . . . . 19

Notices . . . . . . . . . . . . . . 21Trademarks . . . . . . . . . . . . . . 22

Glossary . . . . . . . . . . . . . . 25

© Copyright IBM Corp. 2016 iii

Page 6: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

iv ESS 4.0: Quick Deployment Guide

Page 7: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Tables

1. Conventions . . . . . . . . . . . . viii 2. Known issues in ESS 4.0.5. . . . . . . . 13

© Copyright IBM Corp. 2016 v

Page 8: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

vi ESS 4.0: Quick Deployment Guide

Page 9: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

About this information

This information guides you in quickly installing, or upgrading to, version 4.0.x of the Elastic StorageServer (ESS).

For detailed ESS installation and upgrade information, see Deploying the Elastic Storage Server.

Who should read this informationThis information is intended for experienced system installers and upgraders who are familiar with ESSsystems.

Prerequisite and related informationESS information

The ESS 4.0.x library consists of these information units:v Deploying the Elastic Storage Server, SC27-6659v Elastic Storage Server: Quick Deployment Guide, SC27-8580v IBM Spectrum Scale RAID: Administration, SC27-6658

For more information, see IBM® Knowledge Center:

http://www-01.ibm.com/support/knowledgecenter/SSYSP8_4.0.0/sts40_welcome.html

For the latest support information about IBM Spectrum Scale™ RAID, see the IBM Spectrum Scale RAIDFAQ in IBM Knowledge Center:

http://www.ibm.com/support/knowledgecenter/SSYSP8/sts_welcome.html

Related information

For information about:v IBM Spectrum Scale, see IBM Knowledge Center:

http://www.ibm.com/support/knowledgecenter/STXKQY/ibmspectrumscale_welcome.html

v IBM POWER8® servers, see IBM Knowledge Center:http://www.ibm.com/support/knowledgecenter/POWER8/p8hdx/POWER8welcome.htm

v The DCS3700 storage enclosure, see:– System Storage® DCS3700 Quick Start Guide, GA32-0960-03:

http://www.ibm.com/support/docview.wss?uid=ssg1S7004915

– IBM System Storage DCS3700 Storage Subsystem and DCS3700 Storage Subsystem with PerformanceModule Controllers: Installation, User's, and Maintenance Guide , GA32-0959-07:http://www.ibm.com/support/docview.wss?uid=ssg1S7004920

v The IBM Power Systems™ EXP24S I/O Drawer (FC 5887), see IBM Knowledge Center :http://www.ibm.com/support/knowledgecenter/8247-22L/p8ham/p8ham_5887_kickoff.htm

v Extreme Cluster/Cloud Administration Toolkit (xCAT), go to the xCAT website :http://sourceforge.net/p/xcat/wiki/Main_Page/

© Copyright IBM Corp. 2016 vii

Page 10: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Conventions used in this informationTable 1 describes the typographic conventions used in this information. UNIX file name conventions areused throughout this information.

Table 1. Conventions

Convention Usage

bold Bold words or characters represent system elements that you must use literally, such ascommands, flags, values, and selected menu options.

Depending on the context, bold typeface sometimes represents path names, directories, or filenames.

bold underlined bold underlined keywords are defaults. These take effect if you do not specify a differentkeyword.

constant width Examples and information that the system displays appear in constant-width typeface.

Depending on the context, constant-width typeface sometimes represents path names,directories, or file names.

italic Italic words or characters represent variable values that you must supply.

Italics are also used for information unit titles, for the first use of a glossary term, and forgeneral emphasis in text.

<key> Angle brackets (less-than and greater-than) enclose the name of a key on the keyboard. Forexample, <Enter> refers to the key on your terminal or workstation that is labeled with theword Enter.

\ In command examples, a backslash indicates that the command or coding example continueson the next line. For example:

mkcondition -r IBM.FileSystem -e "PercentTotUsed > 90" \-E "PercentTotUsed < 85" -m p "FileSystem space used"

{item} Braces enclose a list from which you must choose an item in format and syntax descriptions.

[item] Brackets enclose optional items in format and syntax descriptions.

<Ctrl-x> The notation <Ctrl-x> indicates a control character sequence. For example, <Ctrl-c> meansthat you hold down the control key while pressing <c>.

item... Ellipses indicate that you can repeat the preceding item one or more times.

| In synopsis statements, vertical lines separate a list of choices. In other words, a vertical linemeans Or.

In the left margin of the document, vertical lines indicate technical changes to theinformation.

How to submit your commentsYour feedback is important in helping us to produce accurate, high-quality information. You can addcomments about this information in IBM Knowledge Center:

http://www.ibm.com/support/knowledgecenter/SSYSP8/sts_welcome.html

To contact the IBM Spectrum Scale development organization, send your comments to the followingemail address:

[email protected]

viii ESS 4.0: Quick Deployment Guide

Page 11: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Deploying the Elastic Storage Server - for experienced users

This topic includes a concise set of deployment instructions for those who are familiar with ElasticStorage Server (ESS) systems.

In these instructions:v All version numbers shown are examples. The version depends on the release and edition that is being

deployed.v Node names ems1, gssio1, and gssio2 are examples. Each environment could have its own unique

naming conventions.

Follow these high-level steps:1. Complete the prerequisite tasks.2. Unpack the ESS install/upgrade software from FixCentral at http://www-933.ibm.com/support/

fixcentral/swg/selectFixes?parent=Software%2Bdefined%2Bstorage&product=ibm/StorageSoftware/IBM+Spectrum+Scale+RAID&release=All&platform=All&function=all

3. Obtain the required Kernel Errata: http://www-01.ibm.com/support/docview.wss?uid=ssg1S1005719.4. Complete one of the following tasks:

a. Install the ESS system.b. Upgrade the ESS system.

Complete the prerequisite tasks

Complete these tasks before proceeding:1. Ensure nodes are properly prepared for deployment.v EMS and IO Server node network requirements are met with correct /etc/hosts entries in EMS

nodev HMC is properly configured for the EMS and IO Server nodes and partition names are correctly setv Nodes are powered up

2. Obtain a Red Hat Enterprise Linux 7.1 ISO image file (e.g., rhel-server-7.1-ppc64-dvd.iso) or DVD for64-bit IBM Power Systems architecture. The ISO or DVD is used to upgrade EMS node as well asupgrade or deploy IO Server nodes.

3. Obtain the ESS software archive from FixCentral at http://www-933.ibm.com/support/fixcentral/swg/selectFixes?parent=Software%2Bdefined%2Bstorage&product=ibm/StorageSoftware/IBM+Spectrum+Scale+RAID&release=All&platform=All&function=all.

4. Review the list of known issues for the ESS version you are installing.See Appendix A, “Known issues,” on page 13 for more information.

Install the management server software1. Unpack the ESS software archive:

tar -zxvf gss_install-4.0.5_ppc64_advanced_20160729T004528Z.tgz

2. Check the MD5 checksum:

md5sum -c gss_install-4.0.5_ppc64_advanced_20160729T004528Z.md5

3. Make sure the /opt/ibm/gss/install directory is clean:

/bin/sh gss_install-4.0.5_ppc64_advanced_20160729T004528Z --remove

4. Extract the ESS packages and accept the license as shown below. By default it will be extracted to:

© Copyright IBM Corporation © IBM 2016 1

|

||

||

Page 12: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

/opt/ibm/gss/install directory/bin/sh gss_install-4.0.5_ppc64_advanced_20160729T004528Z –text-only

5. For install and deployment go to the section titled Install the ESS system on page 2. To upgrade anexisting ESS system, go to the section titled Upgrade the ESS system on page 6.

Install the ESS systemFollow these steps to perform a new installation of the ESS software on a management server node andI/O server nodes. Node hostnames ems1, gssio1, and gssio2 are examples. Each environment could haveits own unique naming conventions. For an xCAT command such as updatenode, use an xCAT hostname.For the IBM Spectrum Scale commands (those start with mm), use an IBM Spectrum Scale hostname. Forexample, ems1 is an xCAT hostname (typically a hostname associated with the management interface)and ems1-hs is the corresponding IBM Spectrum Scale hostname (typically a hostname associated withthe high speed interface).1. Copy the gssdeploy script and customize it for your environment by editing it:

cp /opt/ibm/gss/install/samples/gssdeploy /var/tmpchmod +x /var/tmp/gssdeploy

/var/tmp is a sample directory name. You can specify a different directory name. The gssdeployscript uses a directory called /tmp/gssdeploy, so do not copy the script to /tmp. Copy the iso (if iso isused) into the directory defined in the customized section of the script. By default it is set to/opt/ibm/gss/iso.

2. Clean the current xCAT installation and associated configuration to remove any preexisting xCATconfiguration:/var/tmp/gssdeploy -c

3. Update the ESS repositories on the management server node:cd /opt/ibm/gss/installinstaller/gssinstall -m manifest -u

4. Run the gssdeploy script:

/var/tmp/gssdeploy -x

If any step fails, address the issue by looking at the error message. Depending on the nature of theproblem, you may need to restart from the last completed step, or you may need to start from thebeginning. If you are starting from the beginning, restart from step 2 (gssdeploy -c) of this section.

5. Log out and then log back in to acquire the environment updates.6. Set up the Kernel Errata repository. Obtain the required Kernel Errata stated in http://www-

01.ibm.com/support/docview.wss?uid=ssg1S1005719 and complete steps 1 to 5 stated in the SectionInstall the Kernel Update. Also see Appendix B, “Instructions for installing the ESS Red Hat LinuxErrata Kernel Update,” on page 19

7. Update the management server node. Here ems1 is the xCAT hostname. This step updates the nodesystem profile, prepares OFED and gplbin rpm and installs the required rpms.

updatenode ems1 -P gss_updatenode

Use systemctl reboot to reboot the management server node and run this step again as shownbelow. This additional step is required to account for changes in kernel, removal of drivers, etc.updatenode ems1 -P gss_updatenode

8. Update OFED on the management server node:

updatenode ems1 -P gss_ofed

9. Update the IP RAID Adapter firmware on the management server node:updatenode ems1 -P gss_ipraid

10. Use systemctl reboot to reboot the management server node.

2 ESS 4.0: Quick Deployment Guide

||

|||

||||

||

|

Page 13: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Deploy the I/O server nodes1. Before initiating the deployment of the I/O server nodes, verify that the attached storage enclosures

are powered off.2. Deploy on the I/O server nodes using the customized deploy script:

./gssdeploy -d

3. Run:

nodestat gss_ppc64

The installation is complete when nodestat displays sshd (it may take approximately 30 minutes) forall I/O server nodes. Here gss_ppc64 is the xCAT nodegroup containing I/O Server nodes. To followthe progress of an I/O server install, you can tail the console log by doingtailf /var/log/consoles/gssio1

where gssio1 is an example default I/O node name.4. At the end of the deployment wait approximately five minutes and reboot the node:

xdsh gss_ppc64 systemctl reboot

5. Update the IP RAID Adapter firmware on the I/O server nodes:updatenode gss_ppc64 -P gss_ipraid

6. Once rebooted, verify the installation by running gssinstallcheck:gssinstallcheck -G ems1,gss_ppc64

Check for any error with the following:a. Installed packagesb. Linux kernel releasec. OFED leveld. IPR SAS FWe. ipraid RAID levelf. ipraid RAID statusg. IPR SAS queue depthh. System firmwarei. System profile settingj. Host adapter driverk. Storage Adapters with 64-bit DMA setting

Note: This item (Storage Adapters) will only be shown for I/O Server nodes.l. Network Adapters with 64-bit DMA setting

Ignore other errors that may be flagged by the gssinstallcheck script. They will go away after theremaining installation steps are completed.

Check the system hardware

After the I/O server nodes have been installed successfully, power on the attached enclosures. Waitapproximately five to 10 minutes from power on for discovery to complete before moving on to the nextstep. Here is the list of key log files that should be reviewed for possible problem resolution duringdeployment.v By default /var/log/message log from all IO Server nodes are directed to the message log in the EMS

node.v gssdeploy log is located at /var/log/gssv xCAT log is located at /var/log/xcat

Deploying the Elastic Storage Server - for experienced users 3

|

|

Page 14: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

v Console outputs from the IO Server node during deployment are located at /var/log/consoles1. Run gss_iprraid:

updatenode gss_ppc64 -P gss_iprraid

2. Run gssstoragequickcheck:

gssstoragequickcheck -G gss_ppc64

3. Run gss_sashba script:

updatenode gss_ppc64 -P gss_sashba

4. Run gssfindmissingdisks:

gssfindmissingdisks -G gss_ppc64

If gssfindmissingdisks displays an error, run mmgetpdisktopology and pipe it to topsummary oneach I/O server node to obtain more information about the error:

mmgetpdisktopology | topsummary

5. Run gsscheckdisks:GSSENV=INSTALL gsscheckdisks -G gss_ppc64 --encl all --iotest a --write-enable

Attention: When run with --iotest w (write) or --iotest a (all), gsscheckdisks will perform write I/Oto the disks attached through the JBOD. This will overwrite the disks and will result in the loss of anyconfiguration or user data stored on the attached disks. gsscheckdisks should be run only during theinstallation of a building block to validate that read and write operations can be performed to theattached drives without any error. The GSSENV environment variable must be set to INSTALL toindicate that gsscheckdisks is being run during installation.

6. Check for any hardware serviceable events and address them as needed. To view the serviceableevents, issue the following command:gssinstallcheck -N ems1,gss_ppc64 --srv-events

Note that during the initial deployment of the nodes, SRC BA15D001 may be logged as a serviceableevent by Partition Firmware. This is normal and should be cleared after the initial deployment. Formore information, see Appendix A, “Known issues,” on page 13.

Set up the high-speed network

Set up the high-speed network that will be used for the cluster data communication. Update your/etc/hosts file with high-speed network entries showing the high-speed IP address and correspondinghost name. Copy the modified /etc/hosts to I/O Server nodes of the cluster. With the Ethernethigh-speed network, you can use the gssgennetworks script to create a bonded Ethernet interface overactive (up) high-speed network interfaces. See “ESS 4.0.5 issues” on page 131. Update /etc/hosts file with high-speed hostname entries in the EMS node and copy the modified

/etc/hosts file to the I/O Server nodes as follows:xdcp gss_ppc64 /etc/hosts /etc/hosts

2. To see the current set of active (up) interfaces on all nodes, run:gssgennetworks -N ems1,gss_ppc64 --suffix=-hs

where ems1 is the name of the EMS node and gss_ppc64 is the I/O server node group and -hs is thenodename suffix of the high-speed host name.

3. To create a bonded interface, in all nodes run:gssgennetworks -N ems1,gss_ppc64 --suffix=-hs --create-bond

4 ESS 4.0: Quick Deployment Guide

|

|

||

|

|||

|

|

|

Page 15: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

The script sets miimon to 100, the bonding mode to 802.3ad (LACP), and xmit_hash_policy tolayer3+4. In some network configurations xmit_hash_policy of layer2+3 may provide a better loadbalancing of the traffic over the slave interfaces of the bond. You can change the xmit_hash_policy asfollows once the bond is created.nmcli c mod bond-bond0 +bond.option xmit_hash_policy=layer2+3

where bond-bond0 is the bonded interface.The other bond options are left with the default values, including lacp_rate (the default is slow). Forproper network operation, the Ethernet switch settings in the networking infrastructure must matchthe I/O server node interface bond settings. If the Ethernet switch and network infrastructure cannotsupport bonding mode 802.3ad (LACP), another bonding mode can be selected. For more information,see the gssgennetworks man page .See Deploying the Elastic Storage Server for:v more information about gssgennetworks

v information about creating a bonded InfiniBand network.

Note: If using bonded IP over IB do the following:Remove or comment out any CONNECTED_MODE=yes statement from the correspondingslave-bond interface scripts located in /etc/sysconfig/network-scripts directory of the EMS and IOServer nodes. These scripts are created as part of the IP over IB bond creation. An example of theslave-bond interface with the modification is shown below.TYPE=Infiniband <= change from Ethernet to InfinibandNAME=bond-slave-ib0 <= bond-slave-ib0 is the slave connectionUUID=86c0af63-4b6c-475c-a724-0fb074dc9092DEVICE=ib0 <= slave interfaceONBOOT=yesMASTER=bond0 <= master bond interfaceSLAVE=yes#CONNECTED_MODE=yes <= do not add this line, comment out if this statement existsNM_CONTROLLED=yes <= add this line

Run:nmcli c reload bond-bond0Here bond-bond0 is the connection name of the bond.

4. Once the high-speed network is set up, stress test the high-speed network as follows:GSSENV=TEST gssnettest -N ems1,gss_ppc64 --suffix=-hs

where ems1 is the EMS node name and gss_ppc64 is the I/O server node group and -hs is thenodename suffix of the high-speed host name. This test should only be run in a test environment as itcan highly stress the high-speed network.

Create the cluster, recovery groups, and file system1. Create the GPFS™ cluster:

gssgencluster -C test01 -G gss_ppc64 --suffix=-hs --accept-license

In this example, test01 is used as the cluster name and -hs is used as the suffix of the hostname.2. Create the recovery groups:

gssgenclusterrgs -G gss_ppc64 --suffix=-hs

3. Create the vdisks, NSDs, and file system:gssgenvdisks --create-vdisk --create-nsds --create-filesystem --contact-node gssio1

4. Add the management server node to the cluster:

gssaddnode -N ems1 --cluster-node gssio1 --suffix=-hs --accept-license --no-fw-update

Deploying the Elastic Storage Server - for experienced users 5

|||

|

|

|||

|

|

|||

Page 16: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

In this example, the management server hostname is ems1 with a suffix of -hs (ems1-hs) in thehigh-speed network. The --no-fw-update option is used because the management server node doesnot contain a SAS adapter or attached drives.

Check the installed software and system health1. Run gssinstallcheck on the management server:

gssinstallcheck -N ems1

2. Run gssinstallcheck on the I/O server nodes:gssinstallcheck -G gss_ppc64

3. Shut down GPFS in all nodes and reboot all nodes.a. Shut down GPFS all nodes:

mmshutdown -a

b. Reboot all server nodes:xdsh gss_ppc64 "systemctl reboot"

c. Reboot EMS node:systemctl reboot

4. After reboots, run:gssinstallcheck -G gss_ppc64 --phy-mapping

Ensure that the phy mapping check is OK.5. Restart GPFS in all nodes and wait for all nodes to become active:

mmstartup -a

6. Mount the filesystem and perform a stress test. For example, run:

mmmount gpfs0 -agssstress /gpfs/gpfs0 gssio1 gssio2

In this example, gssstress is invoked on the management server node. It is run on I/O server nodesgssio1 and gssio2 with /gpfs/gpfs0 as the target path. By default gssstress runs for 20 iterations andcan be adjusted using the -i option (type gssstress and press Enter to see the available options).During the IO stress test, check for network error by running from another console:gssinstallcheck -N ems1,gss_ppc64 -–net-errors

7. Perform a health check. Run:

gnrhealthcheck

Address any issues that are identified.8. Check for any open hardware serviceable events and address them as needed. The serviceable events

can be viewed as follows:gssinstallcheck -N ems1,gss_ppc64 --srv-events

Note that during initial deployment of the nodes, SRC BA15D001 may be logged as serviceable eventby Partition Firmware. This is normal and should be cleared after the initial deployment. For moreinformation, see Appendix A, “Known issues,” on page 13.

Install the ESS GUI1. Generate performance collector in the EMS node. The EMS node must be part of the ESS cluster and

node name must be the node name used in the cluster (e.g., esm1-hs). Run:

mmperfmon config generate --collectors ems1-hs

2. Set up the nodes in the ems nodeclass and gss_ppc64 nodeclass for performance monitoring. Run:mmchnode --perfmon -N ems,gss_ppc64

6 ESS 4.0: Quick Deployment Guide

||

|

||

|

|||

Page 17: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

3. Capacity and filesetquota monitoring is not enabled in the GUI by default.a. To enable capacity and filesetquota run the following command:

mmperfmon config update GPFSDiskCap.restrict=<EMS nodename> GPFSDiskCap.period=<period in sec>

Here the EMS node name must be the name shown in the mmlscluster output.b. Verify that the GPFSDiskCap.period is set correctly in the /opt/IBM/zimon/ZIMonSensors.cfg. If

it is not set correctly, verify that the EMS node is correctly provided. The following example showswhat it should look like when period is set to 86400 sec (one day) and the task only runs in theems-hs (mmlscluster output showing ems1-hs.gpfs.net) node.{

name = "GPFSDiskCap"period = 86400restroct = "ems1-hs.gpfs.net"

}

Note: To enable quota the filesystem quota checking must be enabled. Refer mmchfs -Q andmmcheckquota commands in the IBM Spectrum Scale Administration and Programming Reference Guide.

4. Start sensors in the EMS node and I/O server nodes:systemctl start pmsensorsxdsh gss_ppc64 "systemctl start pmsensors"

5. Start performance collector in the EMS node:systemctl start pmcollector

6. Enable and start gpfsgui:systemctl enable gpfsgui.servicesystemctl start gpfsgui

7. To launch the ESS GUI in a browser, go to: https://EssGuiNode where ESSGuiNode is the hostnameor IP address of the EMS node for GUI access. To log in, type admin in the User Name field and yourpassword in the Password field on the login page. The default password for admin is admin001. Walkthrough each panel and complete the GUI Setup Wizard.

This completes the Install task of the ESS system.

Upgrade the ESS system

Perform a hardware and software healthcheck and address any issues before starting the upgrade of thesystem. During the upgrade process if a step fails, it must be addressed before moving to the next step.Follow these steps to perform an upgrade of the ESS system.

Prepare the system for upgrade1. Perform a health check. Run:

gnrhealthcheck

Address any issues that are identified.2. Wait for any of these commands that are performing file system maintenance tasks to complete:

mmadddiskmmapplypolicymmcheckquotammdeldiskmmfsckmmlssnapshotmmrestorefsmmrestripefile

Deploying the Elastic Storage Server - for experienced users 7

||

|||

Page 18: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

mmrestripefsmmrpldisk

3. It is recommended that you stop the creation and deletion of snapshots using mmcrsnapshot andmmdelsnapshot during the upgrade window.

4. Check for any hardware serviceable events from the HMC. If you are running ESS 4.0.3 or newer youcan obtain serviceable events using gssinstallcheck as follows:gssinstallcheck -N ems1,gss_ppc64 --srv-events

Address any hardware issues identified in the serviceable events.5. Check the status of the local boot drive shown in the ipraid RAID status line output of the

gssinstallcheck:gssinstallcheck -N ems1,gss_ppc64

If the status does not show that it is optimized, get it serviced. The local drive state can also beobtained by running iprconfig from the command line and selecting option 1. Check for RAID 10Array Status. It should state Optimized.

Upgrading from versions earlier than ESS 3.0

ESS versions 2.5.x should first be upgraded to ESS 3.5.2 prior to upgrading to ESS 4.0. Contact CustomerSupport for further details.

Upgrading from ESS 3.x.y and ESS 4.0.x

Perform the following steps if you are upgrading from ESS 3.x.y and ESS 4.0.x:1. Update ESS repositories on the management server node:

cd /opt/ibm/gss/installinstaller/gssinstall -m manifest -u

2. Go to the next section titled Update the management server node.

Update the management server node1. On the management server node, stop GUI services:

systemctl stop gpfsgui

2. Save collector configuration files in the EMS node for later usage:cp /opt/IBM/zimon/ZIMonCollector.cfg /tmp

3. Save sensor configuration files in EMS and IO Server nodes (ssh to IO Server nodes and run thecommand locally) for later usage:cp /opt/IBM/zimon/ZIMonSensors.cfg /tmp

4. Shut down IBM Spectrum Scale on the management server node while making sure quorum is stillmaintained. Run:mmshutdown

5. If you have not installed the Kernel Errata stated in the http://www-01.ibm.com/support/docview.wss?uid=ssg1S1005719, obtain the required Kernel Errata stated in this link. Complete steps1 to 5 in the section stated in the Section Install the Kernel Update. Also see Appendix B,“Instructions for installing the ESS Red Hat Linux Errata Kernel Update,” on page 19. Otherwise,skip this step.

6. Update the management server node:updatenode ems1 -P gss_updatenode

Use systemctl reboot to reboot the management server node and complete this step again asfollows:updatenode ems1 -P gss_updatenode

8 ESS 4.0: Quick Deployment Guide

||

|

|

||

|

|||

|

Page 19: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

7. If upgrading from ESS 3.x.y perform the following steps to upgrade IBM Spectrum Scale RAIDconfiguration parameters.Skip these steps if upgrading from ESS 4.0.x./opt/ibm/gss/tools/samples/gssupg400.sh -b

/opt/ibm/gss/tools/samples/gssupg400.sh -c

/opt/ibm/gss/tools/samples/gssupg400.sh -p

8. Update OFED on the management server node:

updatenode ems1 -P gss_ofed

9. Update IP RAID Adapter firmware on the management server node:updatenode ems1 -P gss_ipraid

10. If using bonded IP over IB, remove or comment out CONNECTED_MODE=yes statement from thecorresponding slave-bond interface scripts located in the /etc/sysconfig/network-scripts directoryof the ems1 node. An example of the slave-bond interface with the modification is shown below.

TYPE=InfinibandNAME=bond-slave-ib0UUID=86c0af63-4b6c-475c-a724-0fb074dc9092DEVICE=ib0ONBOOT=yesMASTER=bond0SLAVE=yes#CONNECTED_MODE=yes <= this line commented outNM_CONTROLLED=yes

11. Use systemctl reboot to reboot the management server node.12. Start IBM Spectrum Scale on the management server node:

mmstartup

13. Verify that IBM Spectrum Scale is in the active state:mmgetstate

Do not proceed if the system is not active.

Update the I/O server nodes

Repeat the following steps for each I/O server node, one node at a time.1. Move the cluster and file system manager role to another node if the current node is a cluster

manager or file system manager.a. To find the cluster and file system managers, run:

mmlsmgr

b. To change the file system manager, run:

mmchmgr gpfs0 gssio2-hs

In this example, gssio2-hs is the new file system manager of file system gpfs0.c. To change the cluster manager, run:

mmchmgr -c gssio2-hs

In this example, gssio2-hs is the new cluster manager.2. Move the recovery group in the current I/O server node to the peer I/O server node in the same

building block.a. To list the recovery groups, run:

mmlsrecoverygroup

b. To list the active server, primary server, and secondary server, run:

Deploying the Elastic Storage Server - for experienced users 9

|

|

|

|

Page 20: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

mmlsrecoverygroup rg_gssio1-hs -L | grep active -A2

c. To move the recovery group from the current active I/O server node (rg_gssio1-hs) to the peerI/O server node (gssio2-hs) in the same building block, run:

mmchrecoverygroup rg_gssio1-hs --servers gssio2-hs,gssio1-hs

3. After confirming that the recovery group has been successfully moved to the peer I/O server node,shut down IBM Spectrum Scale on the current I/O server node while maintaining quorum:

mmshutdown -N CurrentIoServer-hs

4. Run updatenode:

updatenode CurrentIoServer -P gss_updatenode

Reboot the I/O server node and complete this step again if you are instructed to do so in theupdatenode output. Reboot the I/O server node as follows :xdsh CurrentIoServer "systemctl reboot"

5. Update OFED.

updatenode CurrentIoServer -P gss_ofed

6. Update IP RAID FW in the I/O Server node that is being upgraded.updatenode CurrentIoServer -P gss_ipraid

7. If using bonded IP over IB, remove or comment out CONNECTED_MODE=yes statement from thecorresponding slave-bond interface scripts located in the /etc/sysconfig/network-scripts directoryof the CurrentIOServer node. An example of the slave-bond interface with the modification is shownbelow.

TYPE=InfinibandNAME=bond-slave-ib0UUID=86c0af63-4b6c-475c-a724-0fb074dc9092DEVICE=ib0ONBOOT=yesMASTER=bond0SLAVE=yes#CONNECTED_MODE=yes <= this line commented outNM_CONTROLLED=yes

8. Reboot the I/O server node as follows:xdsh CurrentIoServer "systemctl reboot"

9. Update the SAS host adapter firmware on CurrentIoServer:

CurrentIoServer$ mmchfirmware --type host-adapter

Here CurrentIOServer is an IO Server node and the command is run on the IO Server node.10. Update the node configuration:

/opt/ibm/gss/tools/samples/gssupg400.sh -s CurrentIoServer-hs

This command is run from the EMS node.11. Run phy check and ensure that the phy mapping is OK:

gssinstallcheck -N CurrentIoServer --phy-mapping

12. Start IBM Spectrum Scale on the I/O server node. Run:mmstartup -N CurrentIoServer-hs

Once Spectrum Scale daemon is successfully started, move back the cluster manager and the filesystem manager if required. Move back the recovery group that was moved to the peer I/O Servernode of the same building block in Step 2 above.

13. Wait until the management server node shows that it is active, using the following command:mmgetstate

10 ESS 4.0: Quick Deployment Guide

|

|

|

|

Page 21: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

14. Repeat steps 1 through 13 for the peer I/O Server node of the same building block.15. Repeat all steps in this section for each additional building block.

Update the enclosure and drive firmware1. To update the storage enclosure firmware, run the following command from one I/O Server node of

each building block:CurrentIoServer$ mmchfirmware --type storage-enclosure

2. To update the drive firmware, run the following command from one I/O Server node of eachbuilding block:CurrentIoServer$ mmchfirmware --type drive

The drive update can take some time to complete. You can update the drives more quickly by takingthe system offline (shutting down IBM Spectrum Scale) and using the --fast-offline option.

Check the installed software and system health1. Run gssinstallcheck on the management server:

gssinstallcheck -N ems1

2. Run gssinstallcheck on the I/O server nodes:gssinstallcheck -G gss_ppc64

3. Perform a health check. Run:

gnrhealthcheck

4. Check for any hardware serviceable events and address them as needed. To view the serviceableevents, issue the following command:gssinstallcheck -N ems1,gss_ppc64 --srv-events

Note that during the initial deployment of the nodes, SRC BA15D001 may be logged as a serviceableevent by Partition Firmware. This is normal and should be cleared after the initial deployment. Formore information, see Appendix A, “Known issues,” on page 13.

Upgrading GUI

Upgrading from ESS 3.0.x

Perform the following steps to upgrade from ESS 3.0.x:1. Restore the sensor configuration file (that was saved in step 2 of the Update the management server

node section) in each node (ssh to EMS and IO Server nodes and run the command locally):cp /tmp/ZIMonSensors.cfg /opt/IBM/zimon

2. Restore the collector configuration file that was saved in step 2 of the Update management servernode section:cp /tmp/ZIMonCollector.cfg /opt/IBM/zimon

3. Start the performance data collector service in the EMS node and sensor for performance datacollection in all nodes:systemctl start pmcollectorsystemctl start pmsensorsxdsh gss_ppc64 "systemctl restart pmsensors"

4. Enable and start the GUI:systemctl enable gpfsgui.servicesystemctl start gpfsgui

Upgrading from ESS 3.5.x or ESS 4.0.x

Perform the following steps to upgrade from ESS 3.5.x:

Deploying the Elastic Storage Server - for experienced users 11

||

|

|||

Page 22: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

1. Generate performance collector in the EMS node. The EMS node must be part of the ESS cluster andthe node name must be the node name used in the cluster (e.g., ems1-hs). Run:mmperfmon config generate --collectors ems1-hs

2. Set up the nodes in the ems nodeclass and gss_ppc64 nodeclass for performance monitoring. Run:mmchnode --perfmon -N ems,gss_ppc64

3. Capacity and filesetquota monitoring is not enabled in the GUI by default.a. To enable capacity and filesetquota run the following command:

mmperfmon config update GPFSDiskCap.restrict=<EMS nodename> GPFSDiskCap.period=<period in sec>

Here the EMS node name must be the name shown in the mmlscluster output.b. Verify that the GPFSDiskCap.period is set correctly in the /opt/IBM/zimon/ZIMonSensors.cfg. If

it is not set correctly, verify that the EMS node is corectly provided. The following example showswhat it should look like when period is set to 86400 sec (one day) and the task only runs in theems1-hs (mmlscluster output showing ems1-hs.gpfs.net) node.

Note: Note: To enable quota the filesystem quota checking must be enabled. Refer mmchfs -Q andmmcheckquota commands in the IBM Spectrum Scale Administration and Programming ReferenceGuide.{

name = "GPFSDiskCap"period = 86400restroct = "ems1-hs.gpfs.net"

}

Note: To enable quota, the filesystem quota checking must be enabled. Refer to the mmchfs -Qand mmcheckquota commands in the IBM Spectrum Scale Administration and ProgrammingReference Guide.

4. Start sensors in the EMS node and I/O server nodes:systemctl start pmsensorssystemctl start pmcollectorxdsh gss_ppc64 "systemctl start pmsensors"

5. Enable and start the GUI:systemctl enable gpfsgui.servicesystemctl start gpfsgui

This completes the upgrade task of the ESS system.

12 ESS 4.0: Quick Deployment Guide

Page 23: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Appendix A. Known issues

This topic includes known issues for ESS.

ESS 4.0.5 issues

Table 2 includes information about known issues in ESS 4.0.5 and how to resolve these issues. Dependingon which fix level you are installing, these might or might not apply to you.

Table 2. Known issues in ESS 4.0.5

IssueEnvironmentaffected Description Resolution or action

1. After upgradingto to ESS 4.0.5(from pre-ESS3.0.0), the Java™

1.6 rpm still existswhich may failsome securityscans.

Cluster softwareupgrade

Type: Upgrade

IBM SpectrumScaleVersion:Advanced orStandard

Affected nodes:I/O IO + EMS

As of 3.0.0 a separate java rpm is nolonger required for GUI operation(java is packaged within the GUIrpm). The java rpm remains on thesystem and may cause some securityscans to fail. The workaround is tosimply remove the rpm via yum.

The rpm in question is:

java-1.6.0-ibm-1.6.0.16.7-1jpp.ppc64

To work around this issue, do thefollowing:

1. Run:

xdsh ems1,gss_ppc64"yum -y removejava-1.6*"

2. Confirm the removal by runningthis command:

xdsh ems1,gss_ppc64"rpm -qa | grep-i java-1.6"

No output should be returned.

© Copyright IBM Corporation © IBM 2016 13

Page 24: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Table 2. Known issues in ESS 4.0.5 (continued)

IssueEnvironmentaffected Description Resolution or action

2. gssinstallcheckmay flag an errorregarding pagepool size inmulti-buildingblock situations ifthe physicalmemory sizesdiffer.

Softwarevalidation

Type: Install orUpgrade

IBM SpectrumScale Version:Advanced orStandard

Affected nodes:I/O

gssinstallcheck is a new toolintroduced in ESS 3.5, that helpsvalidate software, firmware, andconfiguration settings.

If adding (or installing) buildingblocks of a different memoryfootprint installcheck will flag this asan error. Best practice states thatyour IO servers should all have thesame memory footprint, thuspagepool value. Page pool iscurrently set at ~60% of physicalmemory per IO node.

Example from gssinstallcheck:

[ERROR] pagepool:found 142807662592expected range147028338278 -179529339371

1. Confirm each IO node'sindividual memory footprint.From the EMS, run the followingcommand against your IO xCATgroup:

xdsh gss_ppc64 "cat/proc/meminfo |grep MemTotal"

Note: This value is in KB

If the physical memory variesbetween servers and/or buildingblocks, consider adding mermoryand re-calculating pagepool toensure consistency.

2. Validate the pagepool settings inSpectrum Scale:

mmlsconfig | grep -A1 pagepool

Note: This value is in MB.

If the pagepool value setting isnot roughly ~60% of physicalmemory, then you shouldconsider recalculating and settingan updated value.

Please refer to the Spectrum Scaledocumentation on how to updatethe pagepool value.

http://www-01.ibm.com/support/knowledgecenter/SSFKCN/gpfs_welcome.html

3. New Disk PrepScript fails whencopying to IOnodes. Issue willbe called out ingssinstallcheck.This issue is seenwhen upgradingfrom ESS 3.0.Xonly

Softwarevalidation

Type: Upgrade

Version: Advancedor Standard

Affected Nodes:IO

When a user upgrades from ESS3.0.X, the command gssUpg400.sh -pfails to copy the New Disk PrepScript to the IO nodes.

Users will see “End of Task” vs“Copying prepare new disk script".

Further validation of the issue can befound when runninggssinstallcheck. A user would see:

[ERROR] New diskprep script not found.

To work around this issue, run thefollowing from the EMS:

/usr/lpp/mmfs/bin/mmdsh -N gss_ppc64 "cp/usr/lpp/mmfs/samples/vdisk/tspreparenewpdiskforuse/usr/lpp/mmfs/bin/tspreparenewpdiskforuse

To validate, do the following:

xdsh gss_ppc64 "ls -l/usr/lpp/mmfs/bin/tspreparenewpdiskforuse"

Finally, re-run gssinstallcheck tovalidate that the error has beenresolved.

14 ESS 4.0: Quick Deployment Guide

Page 25: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Table 2. Known issues in ESS 4.0.5 (continued)

IssueEnvironmentaffected Description Resolution or action

4. Whenupgrading to ESS4.0.5 the gss_ofedscript may fail fornodes upgradedto Red Hat 7.2.

Cluster Network

Type: Upgrade

Version: Advancedor Standard

Affected Nodes:IO+ EMS

Users who connect nodes to the RedHat Network should only applyspecific security updates. If allnon-essential updates are applied(not advised), systems may not fullyupgrade to Red Hat 7.2. Whenupgrading to ESS 4.0.5, users wouldencounter the following issue whenrunning the gss_ofed script:

gssio1: gss_ofed [DEBUG]:Detected rhel7u2 ppc64.Disabling installing32bit rpms...

gssio1: gss_ofed [DEBUG]:Error: The currentMLNX_OFED_LINUXis intended for rhel7.1

gssio1: gss_ofed [ERROR]:Mellanox install failedRC: 172

The workaround for this issue is todowngrade the redhat-releaseversion back to 7.1. For each nodeexperiencing this problem run thefollowing command:

yum -y downgraderedhat-release-serverunbound-libs

Then validate the level again:cat/etc/redhat-release

Red Hat Enterprise Linux Serverrelease 7.1 (Maipo)You may nowre-run the gss_ofed script to properlyupdate the Mellanox firmware anddriver levels.

Appendix A. Known issues 15

Page 26: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Table 2. Known issues in ESS 4.0.5 (continued)

IssueEnvironmentaffected Description Resolution or action

5. gssgennetworksscript requireshigh-speedhostnames to bederived from IOServer (xCAT)hostnames usingsuffix, prefix, orboth.

High-speednetworkgeneration

Type: Install

Version: Advancedor Standard

Gssgennetworks requires the targethostname provided in -N or -Goption are reachable to create thehigh-speed network on the targetnode.If the xCAT node name doesnot contain the same base name asthe

high-speed name you may beaffected by this issue.

A typical deployment scenario is:

gssio1 // xCAT name

gssio1-hs // high-speed

An Issue scenario is:

gssio1 // xCAT name

foo1abc-hs // high-speed name

Create entries in the /etc/hosts withnode names that are reachable overthe management network such thatthe high-speed hostnames can bederived from it using somecombination of suffix and/or prefix.

For example if the high-speedhostnames are foo1abc-hs,goo1abc-hs :

a. Add foo1 and goo1 to the/etc/hosts using managementnetwork address (reachable) in theEMS node only.

b. Use:

gssgennetworks -N foo1,goo1 –suffixabc-hs --create-bond

c. Remove the entries foo1 and goo1from /etc/hosts file on the EMSnode once the high-speed networksare created.

Example of how to fix (/etc/hosts):

// Before<IP><Long Name><Short Name>192.168.40.21gssio1.gpfs.net

gssio1192.168.40.22gssio2.gpfs.netgssio2

X.X.X.Xfoo1abc-hs.gpfs.netfoo1abc-hs

X.X.X.Ygoo1abc-hs.gpfs.netgoo1abc-hs

// Fix192.168.40.21gssio1.gpfs.netgssio1 foo1

192.168.40.22gssio2.gpfs.netgssio2 goo1

X.X.X.Xfoo1abc-hs.gpfs.netfoo1abc-hs

X.X.X.Ygoo1abc-hs.gpfs.netgoo1abc-hs

gssgennetworks -N foo1,goo1 --suffix=abc-hs--create-bond

16 ESS 4.0: Quick Deployment Guide

Page 27: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Table 2. Known issues in ESS 4.0.5 (continued)

IssueEnvironmentaffected Description Resolution or action

6. During the GUIwizard setup, youmay see anerroneous errorduring thefirmwareverification.

GUI

Type: Install

Version:

Advanced orStandard

Affected Nodes:EMS (GUI host)

Upon a new installation of an ESS4.0.5 system users may receive anerror during the GUI wizard setupstating that the ESM firmware isunknown (Firmware levels are notup to date).

Verify firmware is only run if theuser checks the box during thewizard setup. This error is deemedinvalid and should not prevent thesuccessful completion of the ESSinstall or GUI wizard.

There is no resolution at this time.Please take the following stepsthough to assure a healthy system:

v Make sure gnrhealthcheck is cleanof any problems.

v Make sure gssinstallcheck is runand all firmware levels are up todate and valid.

After the GUI wizard is completemake sure there are no eventsrequiring action.

7. During the GUIwizard setup, youmay see a failureduring the ‘VerifyInstallation’system setupscreen regardingPerformanceMonitoring.

GUI

Type: Install

Version:

Advanced orStandard

Affected Nodes:EMS (GUI host)

The performance monitoring checkof ‘Verify Installation’ may fail in theGUI wizard. You will see an errorlike:

“Did not get performance data fromany node...”

Or “Could not get performancecounters from node ‘X’”

The workaround is simply to pressthe ‘Verify Installation again’ button.You may receive the error 2 or 3times before a successful greencheckmark is displayed next toPerformance Monitoring. If after 3times the Performance Monitoringcheck continues to fail, please checkthat your pmsensors and pmcollectorare correctly configured and started.

8. During initialdeployment of thenodes, SRCBA15D001 islogged byPartitionFirmware

PlatformFirmware

Type: Install

Version:

Advanced orStandard

Affected Nodes:IO + EMS

During deployment of EMS and IOServer nodes, Partition Firmwaremay log Unrecoverable Error, Loss ofFunction and SRC BA15D001.

This error is triggered duringscanning of network ports that arenot populated. The scanning is doneto determine xCAT managementnetwork connection topologybetween EMS and IO Server nodeswhere some network ports canremain unpopulated in a validconfiguration. Therefore, this errorcan be ignored when generatedduring deployment.

Appendix A. Known issues 17

|||||||

||

|

|

||

||

||||

|||||||||||

Page 28: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

18 ESS 4.0: Quick Deployment Guide

Page 29: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Appendix B. Instructions for installing the ESS Red Hat LinuxErrata Kernel Update

This topic provides instructions for installing the Red Hat Linux Errata Kernel Update for ESS.

Perform the following steps to prepare for installation of the ESS Red Hat Linux Errata Kernel Update.1. Obtain the required Kernel Errata here: http://www-01.ibm.com/support/

docview.wss?uid=ssg1S1005719.The following packages are provided in the RHBA-2016-1183_for_ESS.tgz:

dracut-033-241.el7_1.5.ppc64.rpm

dracut-config-rescue-033-241.el7_1.5.ppc64.rpm

dracut-network-033-241.el7_1.5.ppc64.rpm

kernel-3.10.0-229.34.1.el7.ppc64.rpm

kernel-abi-whitelists-3.10.0-229.34.1.el7.noarch.rpm

kernel-bootwrapper-3.10.0-229.34.1.el7.ppc64.rpm

kernel-debug-devel-3.10.0-229.34.1.el7.ppc64.rpm

kernel-devel-3.10.0-229.34.1.el7.ppc64.rpm

kernel-doc-3.10.0-229.34.1.el7.noarch.rpm

kernel-headers-3.10.0-229.34.1.el7.ppc64.rpm

kernel-tools-3.10.0-229.34.1.el7.ppc64.rpm

kernel-tools-libs-3.10.0-229.34.1.el7.ppc64.rpm

perf-3.10.0-229.34.1.el7.ppc64.rpm

2. Copy rpms and setup repository:a. Unpack the Errata Kernel archive on the management server node:

$ cd /var/tmp$ tar -zxvf RHBA-2016-1183_for_ESS.tgz

b. Set up the management server node Kernel repository:$ /opt/ibm/gss/xcat/bin/gssxcatconfig -k

c. Remove old RPMs from the kernel repository directory:$ cd /install/gss/otherpkgs/rhels7/ppc64/kernel/$ rm -f *.rpm

d. Copy the Errata Kernel RPMs into the repository directory:$ cd /install/gss/otherpkgs/rhels7/ppc64/kernel/$ cp /var/tmp/RHBA-2016-1183_for_ESS/*.rpm .

e. Update the Kernel repository information:$ cd /install/gss/otherpkgs/rhels7/ppc64/kernel/$ createrepo .

3. Return to the corresponding install or upgrade task.

© Copyright IBM Corp. 2016 19

|

||

|

|

|

|

|

|

|

|

|

|

|

|

||

Page 30: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

20 ESS 4.0: Quick Deployment Guide

Page 31: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Notices

This information was developed for products and services offered in the U.S.A.

IBM may not offer the products, services, or features discussed in this document in other countries.Consult your local IBM representative for information on the products and services currently available inyour area. Any reference to an IBM product, program, or service is not intended to state or imply thatonly that IBM product, program, or service may be used. Any functionally equivalent product, program,or service that does not infringe any IBM intellectual property right may be used instead. However, it isthe user's responsibility to evaluate and verify the operation of any non-IBM product, program, orservice.

IBM may have patents or pending patent applications covering subject matter described in thisdocument. The furnishing of this document does not grant you any license to these patents. You can sendlicense inquiries, in writing, to:

IBM Director of LicensingIBM CorporationNorth Castle DriveArmonk, NY 10504-1785U.S.A.

For license inquiries regarding double-byte (DBCS) information, contact the IBM Intellectual PropertyDepartment in your country or send inquiries, in writing, to:

Intellectual Property LicensingLegal and Intellectual Property LawIBM Japan Ltd.19-21, Nihonbashi-Hakozakicho, Chuo-kuTokyo 103-8510, Japan

The following paragraph does not apply to the United Kingdom or any other country where suchprovisions are inconsistent with local law:

INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS"WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOTLIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY ORFITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or impliedwarranties in certain transactions, therefore, this statement may not apply to you.

This information could include technical inaccuracies or typographical errors. Changes are periodicallymade to the information herein; these changes will be incorporated in new editions of the publication.IBM may make improvements and/or changes in the product(s) and/or the program(s) described in thispublication at any time without notice.

Any references in this information to non-IBM Web sites are provided for convenience only and do not inany manner serve as an endorsement of those Web sites. The materials at those Web sites are not part ofthe materials for this IBM product and use of those Web sites is at your own risk.

IBM may use or distribute any of the information you supply in any way it believes appropriate withoutincurring any obligation to you.

© Copyright IBM Corp. 2016 21

Page 32: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Licensees of this program who wish to have information about it for the purpose of enabling: (i) theexchange of information between independently created programs and other programs (including thisone) and (ii) the mutual use of the information which has been exchanged, should contact:

IBM CorporationDept. 30ZA/Building 707Mail Station P3002455 South Road,Poughkeepsie, NY 12601-5400U.S.A.

Such information may be available, subject to appropriate terms and conditions, including in some cases,payment or a fee.

The licensed program described in this document and all licensed material available for it are providedby IBM under terms of the IBM Customer Agreement, IBM International Program License Agreement orany equivalent agreement between us.

Any performance data contained herein was determined in a controlled environment. Therefore, theresults obtained in other operating environments may vary significantly. Some measurements may havebeen made on development-level systems and there is no guarantee that these measurements will be thesame on generally available systems. Furthermore, some measurements may have been estimated throughextrapolation. Actual results may vary. Users of this document should verify the applicable data for theirspecific environment.

Information concerning non-IBM products was obtained from the suppliers of those products, theirpublished announcements or other publicly available sources. IBM has not tested those products andcannot confirm the accuracy of performance, compatibility or any other claims related to non-IBMproducts. Questions on the capabilities of non-IBM products should be addressed to the suppliers ofthose products.

This information contains examples of data and reports used in daily business operations. To illustratethem as completely as possible, the examples include the names of individuals, companies, brands, andproducts. All of these names are fictitious and any similarity to the names and addresses used by anactual business enterprise is entirely coincidental.

COPYRIGHT LICENSE:

This information contains sample application programs in source language, which illustrate programmingtechniques on various operating platforms. You may copy, modify, and distribute these sample programsin any form without payment to IBM, for the purposes of developing, using, marketing or distributingapplication programs conforming to the application programming interface for the operating platform forwhich the sample programs are written. These examples have not been thoroughly tested under allconditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of theseprograms. The sample programs are provided "AS IS", without warranty of any kind. IBM shall not beliable for any damages arising out of your use of the sample programs.

If you are viewing this information softcopy, the photographs and color illustrations may not appear.

TrademarksIBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International BusinessMachines Corp., registered in many jurisdictions worldwide. Other product and service names might betrademarks of IBM or other companies. A current list of IBM trademarks is available on the Web at“Copyright and trademark information” at www.ibm.com/legal/copytrade.shtml.

22 ESS 4.0: Quick Deployment Guide

Page 33: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Intel is a trademark of Intel Corporation or its subsidiaries in the United States and other countries.

Java and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/orits affiliates.

Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both.

Microsoft, Windows, and Windows NT are trademarks of Microsoft Corporation in the United States,other countries, or both.

UNIX is a registered trademark of The Open Group in the United States and other countries.

Notices 23

Page 34: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

24 ESS 4.0: Quick Deployment Guide

Page 35: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

Glossary

This glossary provides terms and definitions forthe ESS solution.

The following cross-references are used in thisglossary:v See refers you from a non-preferred term to the

preferred term or from an abbreviation to thespelled-out form.

v See also refers you to a related or contrastingterm.

For other terms and definitions, see the IBMTerminology website (opens in new window):

http://www.ibm.com/software/globalization/terminology

B

building blockA pair of servers with shared diskenclosures attached.

BOOTPSee Bootstrap Protocol (BOOTP).

Bootstrap Protocol (BOOTP)A computer networking protocol thst isused in IP networks to automaticallyassign an IP address to network devicesfrom a configuration server.

C

CEC See central processor complex (CPC).

central electronic complex (CEC)See central processor complex (CPC).

central processor complex (CPC) A physical collection of hardware thatconsists of channels, timers, main storage,and one or more central processors.

clusterA loosely-coupled collection ofindependent systems, or nodes, organizedinto a network for the purpose of sharingresources and communicating with eachother. See also GPFS cluster.

cluster managerThe node that monitors node status usingdisk leases, detects failures, drivesrecovery, and selects file system

managers. The cluster manager is thenode with the lowest node numberamong the quorum nodes that areoperating at a particular time.

compute nodeA node with a mounted GPFS file systemthat is used specifically to run a customerjob. ESS disks are not directly visible fromand are not managed by this type ofnode.

CPC See central processor complex (CPC).

D

DA See declustered array (DA).

datagramA basic transfer unit associated with apacket-switched network.

DCM See drawer control module (DCM).

declustered array (DA)A disjoint subset of the pdisks in arecovery group.

dependent filesetA fileset that shares the inode space of anexisting independent fileset.

DFM See direct FSP management (DFM).

DHCP See Dynamic Host Configuration Protocol(DHCP).

direct FSP management (DFM)The ability of the xCAT software tocommunicate directly with the PowerSystems server's service processor withoutthe use of the HMC for management.

drawer control module (DCM)Essentially, a SAS expander on a storageenclosure drawer.

Dynamic Host Configuration Protocol (DHCP)A standardized network protocol that isused on IP networks to dynamicallydistribute such network configurationparameters as IP addresses for interfacesand services.

E

Elastic Storage Server (ESS)A high-performance, GPFS NSD solution

© Copyright IBM Corporation © IBM 2016 25

Page 36: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

made up of one or more building blocksthat runs on IBM Power Systems servers.The ESS software runs on ESS nodes -management server nodes and I/O servernodes.

encryption keyA mathematical value that allowscomponents to verify that they are incommunication with the expected server.Encryption keys are based on a public orprivate key pair that is created during theinstallation process. See also file encryptionkey (FEK), master encryption key (MEK).

ESS See Elastic Storage Server (ESS).

environmental service module (ESM)Essentially, a SAS expander that attachesto the storage enclosure drives. In thecase of multiple drawers in a storageenclosure, the ESM attaches to drawercontrol modules.

ESM See environmental service module (ESM).

Extreme Cluster/Cloud Administration Toolkit(xCAT)

Scalable, open-source cluster managementsoftware. The management infrastructureof ESS is deployed by xCAT.

F

failbackCluster recovery from failover followingrepair. See also failover.

failover(1) The assumption of file system dutiesby another node when a node fails. (2)The process of transferring all control ofthe ESS to a single cluster in the ESSwhen the other clusters in the ESS fails.See also cluster. (3) The routing of alltransactions to a second controller whenthe first controller fails. See also cluster.

failure groupA collection of disks that share commonaccess paths or adapter connection, andcould all become unavailable through asingle hardware failure.

FEK See file encryption key (FEK).

file encryption key (FEK)A key used to encrypt sectors of anindividual file. See also encryption key.

file systemThe methods and data structures used tocontrol how data is stored and retrieved.

file system descriptorA data structure containing keyinformation about a file system. Thisinformation includes the disks assigned tothe file system (stripe group), the currentstate of the file system, and pointers tokey files such as quota files and log files.

file system descriptor quorumThe number of disks needed in order towrite the file system descriptor correctly.

file system managerThe provider of services for all the nodesusing a single file system. A file systemmanager processes changes to the state ordescription of the file system, controls theregions of disks that are allocated to eachnode, and controls token managementand quota management.

fileset A hierarchical grouping of files managedas a unit for balancing workload across acluster. See also dependent fileset,independent fileset.

fileset snapshotA snapshot of an independent fileset plusall dependent filesets.

flexible service processor (FSP)Firmware that provices diagnosis,initialization, configuration, runtime errordetection, and correction. Connects to theHMC.

FQDNSee fully-qualified domain name (FQDN).

FSP See flexible service processor (FSP).

fully-qualified domain name (FQDN)The complete domain name for a specificcomputer, or host, on the Internet. TheFQDN consists of two parts: the hostnameand the domain name.

G

GPFS clusterA cluster of nodes defined as beingavailable for use by GPFS file systems.

GPFS portability layerThe interface module that each

26 ESS 4.0: Quick Deployment Guide

Page 37: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

installation must build for its specifichardware platform and Linuxdistribution.

GPFS Storage Server (GSS)A high-performance, GPFS NSD solutionmade up of one or more building blocksthat runs on System x servers.

GSS See GPFS Storage Server (GSS).

H

Hardware Management Console (HMC)Standard interface for configuring andoperating partitioned (LPAR) and SMPsystems.

HMC See Hardware Management Console (HMC).

I

IBM Security Key Lifecycle Manager (ISKLM)For GPFS encryption, the ISKLM is usedas an RKM server to store MEKs.

independent filesetA fileset that has its own inode space.

indirect blockA block that contains pointers to otherblocks.

inode The internal structure that describes theindividual files in the file system. There isone inode for each file.

inode spaceA collection of inode number rangesreserved for an independent fileset, whichenables more efficient per-filesetfunctions.

Internet Protocol (IP)The primary communication protocol forrelaying datagrams across networkboundaries. Its routing function enablesinternetworking and essentiallyestablishes the Internet.

I/O server nodeAn ESS node that is attached to the ESSstorage enclosures. It is the NSD serverfor the GPFS cluster.

IP See Internet Protocol (IP).

IP over InfiniBand (IPoIB)Provides an IP network emulation layeron top of InfiniBand RDMA networks,which allows existing applications to runover InfiniBand networks unmodified.

IPoIB See IP over InfiniBand (IPoIB).

ISKLMSee IBM Security Key Lifecycle Manager(ISKLM).

J

JBOD arrayThe total collection of disks andenclosures over which a recovery grouppair is defined.

K

kernel The part of an operating system thatcontains programs for such tasks asinput/output, management and control ofhardware, and the scheduling of usertasks.

L

LACP See Link Aggregation Control Protocol(LACP).

Link Aggregation Control Protocol (LACP)Provides a way to control the bundling ofseveral physical ports together to form asingle logical channel.

logical partition (LPAR)A subset of a server's hardware resourcesvirtualized as a separate computer, eachwith its own operating system. See alsonode.

LPAR See logical partition (LPAR).

M

management networkA network that is primarily responsiblefor booting and installing the designatedserver and compute nodes from themanagement server.

management server (MS)An ESS node that hosts the ESS GUI andxCAT and is not connected to storage. Itcan be part of a GPFS cluster. From asystem management perspective, it is thecentral coordinator of the cluster. It alsoserves as a client node in an ESS buildingblock.

master encryption key (MEK)A key that is used to encrypt other keys.See also encryption key.

Glossary 27

Page 38: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

maximum transmission unit (MTU)The largest packet or frame, specified inoctets (eight-bit bytes), that can be sent ina packet- or frame-based network, such asthe Internet. The TCP uses the MTU todetermine the maximum size of eachpacket in any transmission.

MEK See master encryption key (MEK).

metadataA data structure that contains accessinformation about file data. Suchstructures include inodes, indirect blocks,and directories. These data structures arenot accessible to user applications.

MS See management server (MS).

MTU See maximum transmission unit (MTU).

N

Network File System (NFS)A protocol (developed by SunMicrosystems, Incorporated) that allowsany host in a network to gain access toanother host or netgroup and their filedirectories.

Network Shared Disk (NSD)A component for cluster-wide disknaming and access.

NSD volume IDA unique 16-digit hexadecimal numberthat is used to identify and access allNSDs.

node An individual operating-system imagewithin a cluster. Depending on the way inwhich the computer system is partitioned,it can contain one or more nodes. In aPower Systems environment, synonymouswith logical partition.

node descriptorA definition that indicates how IBMSpectrum Scale uses a node. Possiblefunctions include: manager node, clientnode, quorum node, and non-quorumnode.

node numberA number that is generated andmaintained by IBM Spectrum Scale as thecluster is created, and as nodes are addedto or deleted from the cluster.

node quorumThe minimum number of nodes that mustbe running in order for the daemon tostart.

node quorum with tiebreaker disksA form of quorum that allows IBMSpectrum Scale to run with as little as onequorum node available, as long as there isaccess to a majority of the quorum disks.

non-quorum nodeA node in a cluster that is not counted forthe purposes of quorum determination.

O

OFED See OpenFabrics Enterprise Distribution(OFED).

OpenFabrics Enterprise Distribution (OFED)An open-source software stack includessoftware drivers, core kernel code,middleware, and user-level interfaces.

P

pdisk A physical disk.

PortFastA Cisco network function that can beconfigured to resolve any problems thatcould be caused by the amount of timeSTP takes to transition ports to theForwarding state.

R

RAID See redundant array of independent disks(RAID).

RDMASee remote direct memory access (RDMA).

redundant array of independent disks (RAID)A collection of two or more disk physicaldrives that present to the host an imageof one or more logical disk drives. In theevent of a single physical device failure,the data can be read or regenerated fromthe other disk drives in the array due todata redundancy.

recoveryThe process of restoring access to filesystem data when a failure has occurred.Recovery can involve reconstructing dataor providing alternative routing through adifferent server.

28 ESS 4.0: Quick Deployment Guide

Page 39: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

recovery group (RG)A collection of disks that is set up by IBMSpectrum Scale RAID, in which each diskis connected physically to two servers: aprimary server and a backup server.

remote direct memory access (RDMA)A direct memory access from the memoryof one computer into that of anotherwithout involving either one's operatingsystem. This permits high-throughput,low-latency networking, which isespecially useful in massively-parallelcomputer clusters.

RGD See recovery group data (RGD).

remote key management server (RKM server)A server that is used to store masterencryption keys.

RG See recovery group (RG).

recovery group data (RGD)Data that is associated with a recoverygroup.

RKM serverSee remote key management server (RKMserver).

S

SAS See Serial Attached SCSI (SAS).

secure shell (SSH) A cryptographic (encrypted) networkprotocol for initiating text-based shellsessions securely on remote computers.

Serial Attached SCSI (SAS)A point-to-point serial protocol thatmoves data to and from such computerstorage devices as hard drives and tapedrives.

service network A private network that is dedicated tomanaging POWER8 servers. Provides

Ethernet-based connectivity among theFSP, CPC, HMC, and management server.

SMP See symmetric multiprocessing (SMP).

Spanning Tree Protocol (STP)A network protocol that ensures aloop-free topology for any bridgedEthernet local-area network. The basicfunction of STP is to prevent bridge loopsand the broadcast radiation that resultsfrom them.

SSH See secure shell (SSH).

STP See Spanning Tree Protocol (STP).

symmetric multiprocessing (SMP)A computer architecture that provides fastperformance by making multipleprocessors available to completeindividual processes simultaneously.

T

TCP See Transmission Control Protocol (TCP).

Transmission Control Protocol (TCP)A core protocol of the Internet ProtocolSuite that provides reliable, ordered, anderror-checked delivery of a stream ofoctets between applications running onhosts communicating over an IP network.

V

VCD See vdisk configuration data (VCD).

vdisk A virtual disk.

vdisk configuration data (VCD)Configuration data that is associated witha virtual disk.

X

xCAT See Extreme Cluster/Cloud AdministrationToolkit.

Glossary 29

Page 40: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

30 ESS 4.0: Quick Deployment Guide

Page 41: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM
Page 42: ESS 4.0: Quick Deployment Guide - IBM Stora ge Ser ver V ersion 4.0 Quick Deplo yment Guide SC27-8580-01 IBM

IBM®

Printed in USA

SC27-8580-01