Redp 5497
Redp 5497
                                      Redpaper
International Technical Support Organization
July 2018
                                               REDP-5497-00
 Note: Before using this information and the product it supports, read the information in “Notices” on
 page vii.
This edition applies to IBM Power Systems S922, S914, and S924, machine types and model numbers
9009-22A, 9009-41A, and 9009-42A.
                 Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
                 Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii
                 Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
                 Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
                 Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .x
                 Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
                 Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
iv   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.7 External IO subsystems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .             111
   2.7.1 Peripheral Component Interconnect Express Gen3 I/O expansion drawer . . . .                                                 111
   2.7.2 PCIe Gen3 I/O expansion drawer optical cabling . . . . . . . . . . . . . . . . . . . . . . . .                              112
   2.7.3 PCIe Gen3 I/O expansion drawer system power control network cabling . . . . .                                               114
2.8 External disk subsystems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .             114
   2.8.1 EXP24SX SAS Storage Enclosure and EXP12SX SAS Storage Enclosure. . . .                                                      114
   2.8.2 IBM Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .         116
2.9 Operating system support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .             117
   2.9.1 AIX operating system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .              118
   2.9.2 IBM i . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .   119
   2.9.3 Linux operating system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .              119
   2.9.4 Virtual I/O Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .          120
2.10 POWER9 reliability, availability, and serviceability capabilities. . . . . . . . . . . . . . . . .                              120
2.11 Manageability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .       122
   2.11.1 Service user interfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .              122
   2.11.2 IBM Power Systems Firmware maintenance . . . . . . . . . . . . . . . . . . . . . . . . . .                                 129
   2.11.3 Concurrent firmware maintenance improvements . . . . . . . . . . . . . . . . . . . . . .                                   132
   2.11.4 Electronic Services and Electronic Service Agent . . . . . . . . . . . . . . . . . . . . . .                               133
                                                                                                                      Contents         v
vi   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Notices
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
   AIX®                                 Power Architecture®                   Redbooks®
   C3®                                  POWER6™                               Redbooks (logo)     ®
   Easy Tier®                           POWER7®                               RS/6000™
   IBM®                                 POWER8®                               Storwize®
   IBM Spectrum®                        POWER9™                               System Storage™
   Interconnect®                        PowerHA®                              XIV®
   Micro-Partitioning®                  PowerPC®
   POWER®                               PowerVM®
Intel, Intel Xeon, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks
of Intel Corporation or its subsidiaries in the United States and other countries.
The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.
LTO, Ultrium, the LTO Logo and the Ultrium logo are trademarks of HP, IBM Corp. and Quantum in the U.S.
and other countries.
Microsoft, and the Windows logo are trademarks of Microsoft Corporation in the United States, other
countries, or both.
Fedora, Red Hat, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the United
States and other countries.
UNIX is a registered trademark of The Open Group in the United States and other countries.
VMware, and the VMware logo are registered trademarks or trademarks of VMware, Inc. or its subsidiaries in
the United States and/or other jurisdictions.
Other company, product, or service names may be trademarks or service marks of others.
viii    IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Preface
                 This IBM® Redpaper publication is a comprehensive guide that covers the IBM Power
                 System S922 (9009-22A), IBM Power System S914 (9009-41A), and IBM Power System
                 S924 (9009-42A) servers that support IBM AIX®, IBM i, and Linux operating systems. The
                 objective of this paper is to introduce the major innovative Power S914, Power S922, and
                 Power 924 offerings and their relevant functions:
                  The new IBM POWER9™ processor, which is available at frequencies of 2.3 - 3.8 GHz,
                   2.8 - 3.8 GHz, 2.9 - 3.8 GHz, 3.4 - 3.9 GHz, 3.45 - 3.9 GHz, 3.5 - 3.9 GHz, and
                   3.8 - 4.0 GHz.
                  Significantly strengthened cores and larger caches.
                  Two integrated memory controllers that double the memory footprint of IBM POWER8®
                   servers.
                  Integrated I/O subsystem and hot-pluggable Peripheral Component Interconnect®
                   Express (PCIe) Gen4 and Gen3 I/O slots.
                  I/O drawer expansion options offer greater flexibility.
                  Support for Coherent Accelerator Processor Interface (CAPI) 2.0.
                  New IBM EnergyScale technology offers new variable processor frequency modes that
                   provide a significant performance boost beyond the static nominal frequency.
                 This publication is for professionals who want to acquire a better understanding of IBM Power
                 Systems products. The intended audience includes the following roles:
                    Clients
                    Sales and marketing professionals
                    Technical support professionals
                    IBM Business Partners
                    Independent software vendors (ISVs)
                 This paper expands the current set of IBM Power Systems documentation by providing a
                 desktop reference that offers a detailed technical description of the Power S914, Power S922,
                 and Power S924 systems.
                 This paper does not replace the current marketing materials and configuration tools. It is
                 intended as an extra source of information that, together with existing sources, can be used to
                 enhance your knowledge of IBM server solutions.
Authors
                 This paper was produced by a team of specialists from around the world working at the
                 International Technical Support Organization, Austin Center.
                 Young Hoon Cho is a Power Systems Top Gun with the post-sales Technical Support Team
                 for IBM in Korea. He has over 10 years of experience working on IBM RS/6000™,
                 IBM System p, and Power Systems products. He provides second-line technical support to
                 Field Engineers working on Power Systems and system management.
Thanks to the following individuals for their contribution and support of this publication:
              Ron Arroyo, Tonny JH Bastiaans, Matthew Butterbaugh, Jason Clegg, Nigel Griffiths,
              Daniel Henderson, Jeanine Hinck, Rex Kuo, Ray Laning, Chris Mann, Benjamin Mashak,
              Stephen Mroz, Thoi Nguyen, Kanisha Patel, William Starke, Jeff Stuecheli, Justin Thaler,
              Brian W Thompto, Julie Villarreal
              IBM
              John Banchy
              SIRIUS Computer Solutions
              Neil Palmer
              DPS Inc.
x   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
        Find out more about the residency program, browse the residency index, and apply online at:
        ibm.com/redbooks/residencies.html
Comments welcome
        Your comments are important to us!
        We want our papers to be as helpful as possible. Send us your comments about this paper or
        other IBM Redbooks publications in one of the following ways:
         Use the online Contact us review Redbooks form found at:
           ibm.com/redbooks
         Send your comments in an email to:
           redbooks@us.ibm.com
         Mail your comments to:
           IBM Corporation, International Technical Support Organization
           Dept. HYTD Mail Station P099
           2455 South Road
           Poughkeepsie, NY 12601-5400
                                                                                      Preface    xi
xii   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                                                                                                    1
                 The Power S922 server supports two processor sockets, offering 10-core or 20-core typical
                 2.9 - 3.8 GHz (maximum), 8-core or 16-core typical 3.4 - 3.9 GHz (maximum), or 4-core
                 typical 2.8 - 3.8 GHz (maximum) POWER9 configurations in a 19-inch rack-mount, 2U (EIA
                 units) drawer configuration. All the cores are active. The Power S922 server supports a
                 maximum of 32 DDR4 DIMM slots. Memory features that are supported are 8 GB, 16 GB, 32
                 GB, 64 GB, and 128 GB, and run at different speeds of 2133, 2400, and 2666 Mbps, offering
                 a maximum system memory of 4096 GB.
                 The Power S914 server supports one-processor sockets, offering 4-core typical 2.3 - 3.8 GHz
                 (maximum), 6-core typical 2.3 - 3.8 GHz (maximum), or 8-core typical 2.8 - 3.8 GHz
                 (maximum) POWER9 processor-based configurations in a 19-inch rack-mount, 4U (EIA units)
                 drawer or desk-side configuration. All the cores are active. The Power S914 server supports a
                 maximum of 16 DDR4 DIMM slots. Memory features that are supported are 8 GB, 16 GB, 32
                 GB, and 64 GB, and run at speeds of 2133, 2400, and 2666 Mbps, offering a maximum
                 system memory of 1024 GB.
                 The Power S924 server supports two processor sockets, offering 8-core or 16-core typical
                 3.8 - 4.0 GHz (maximum), 10-core or 20-core typical 3.5 - 3.9 GHz (maximum), 11-core or
                 22-core typical 3.45 - 3.9 GHz, or 24-core typical 3.4 - 3.9 GHz (maximum) configurations in a
                 19-inch rack-mount, 4U (EIA units) drawer configuration. All the cores are active. The Power
                 S924 server supports a maximum of 32 DDR4 DIMM slots. Memory features that are
                 supported are 8 GB, 16 GB, 32 GB, 64 GB, and 128 GB, and run at different speeds of 2133,
                 2400, and 2666 Mbps, offering a maximum system memory of 4096 GB.
                 The Power S922 server supports a maximum of 32 DDR4 Registered DIMM (RDIMM) slots. If
                 only one processor socket is populated, then only 16 RDIMMs can be used. The memory
                 features that are supported are 16 GB, 32 GB, 64 GB, and 128 GB, and run at speeds of
                 1600, 2133, and 2666 MHz, offering a maximum system memory of 2 TB if one socket is
                 single-chip populated and 4 TB with both sockets populated.
                 The IBM Active Memory Expansion feature enables memory expansion by using
                 compression and decompression of memory content, which can effectively expand the
                 maximum memory capacity if extra server workload capacity and performance are available.
                 Each of these backplane options uses leading-edge, integrated SAS RAID controller
                 technology that is designed and patented by IBM.
                 The NVMe option offers fast boot times and is ideally suited for the rootvg of Virtual I/O Server
                 (VIOS) partitions.
2    IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.1.2 Power S914 server
          The Power S914 (9009-41A) server is a powerful one-socket server that ships with up to eight
          activated cores. It has the I/O configuration flexibility to meet today’s growth and tomorrow’s
          processing needs. A one-socket system with a 4-core or 6-core POWER9 processor is
          available in either rack (19-inch rack-mount 4U (EIA units)) or tower configurations. The
          8-core higher performance system is available only in a rack configuration.
          The Power S914 server supports a maximum of 16 DDR4 error-correcting code (ECC)
          RDIMM slots. The memory features that are supported are 16 GB, 32 GB, 64 GB, and
          128 GB, and run at speeds of 2133 - 2666 MHz, offering a maximum system memory of 1 TB.
          If you use the 4-core processor #EP10, the system is limited to four memory DDR4 ECC
          RDIMMS supporting only the 8_GB RDIMMS (# EM60), 16 GB RDIMMs (#EM62), or the
          32 GB RDIMMS (#EM63), which offers a maximum of 64 GB of RAM per system.
          The IBM Active Memory Expansion feature enables memory expansion by using
          compression and decompression of memory content, which can effectively expand the
          maximum memory capacity if extra server workload capacity and performance are available.
          Each of these backplane options uses leading-edge, integrated SAS RAID controller
          technology that is designed and patented by IBM.
The NVMe option offers fast boot times and is ideally suited for the rootvg of VIOS partitions.
          For more information about the NVMe technology, see “PCIe3 NVMe carrier card with two
          M.2 module slots (#EC59)” on page 102.
              The Power S924 server supports a maximum of 32 DDR4 RDIMM slots per processor socket
              that is populated. The memory features that are supported are 16 GB, 32 GB, 64 GB, and
              128 GB, offering a maximum system memory of 4 TB.
              The Active Memory Expansion feature enables memory expansion by using compression and
              decompression of memory content, which can effectively expand the maximum memory
              capacity if more server workload capacity and performance are available.
4   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                 Each of the three backplane options uses leading-edge, integrated SAS RAID controller
                 technology that is designed and patented by IBM.
              The LCD panel is mandatory in the Power S914 Tower. It is also required if the server runs
              IBM i. In the Power S914, Power S922, and Power S924 rack-mounted servers, it is optional,
              but if a rack contains any of these servers, one of the them must have an LCD panel.
              The LCD panel can be moved (by using the correct procedure) from one server to another
              server, which enables appropriate service to be carried out.
              Table 1-1 Electrical characteristics for Power S922, Power S914, and Power S924 servers
                Electrical              Properties
                characteristics
                                        Power S922 server        Power S914 server       Power S924 server
                Operating voltage       1400 W power supply:     900 W power supply:     1400 W power supply:
                                        200 - 240 V AC           100 - 127 V AC or       200 - 240 V AC
                                                                 200 - 240 V AC
                                                                 1400 W power supply:
                                                                 200 - 240 V AC
Power consumption 1880 watts (maximum) 1600 watts (maximum) 2750 watts (maximum)
                Power-source loading    1.94 kVa (maximum        1.65 kVa (maximum       2.835 kVa (maximum
                                        configuration)           configuration)          configuration)
                Note: The maximum measured value is the worst-case power consumption that is
                expected from a fully populated server under an intensive workload. The maximum
                measured value also accounts for component tolerance and non-ideal operating
                conditions. Power consumption and heat load vary greatly by server configuration and
                utilization. The IBM Systems Energy Estimator should be used to obtain a heat output
                estimate that is based on a specific configuration.
6   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Table 1-2 lists the environment requirements for the servers.
Table 1-2 Environment requirements for Power S922, Power S914, and Power S924 servers
 Environment               Recommended              Allowable operating       Non-operating
                           operating
 Note: IBM does not recommend operation above 27C, however, one can expect full
 performance up to 35C for these systems. Above 35C, the system is capable of operating,
 but possible reductions in performance may occur to preserve the integrity of the system
 components. Above 40C there may be reliability concerns for components within the
 system.
Table 1-3 Noise emissions for Power S922, Power S914, and Power S924 servers
 Product             Declared A-weighted sound power          Declared A-weighted sound
                     level, LWAd (B)                          pressure level, LpAm (dB)
 Tip:
  Declared level LWad is the upper-limit A-weighted sound power level. Declared level
   LpAm is the mean A-weighted emission sound pressure level that is measured at the
   1-meter bystander positions.
  All measurements are made in conformance with ISO 7779 and declared in
   conformance with ISO 9296.
  10 dB (decibel) equals 1 B (bel).
Figure 1-6 show the front view of the Power S922 server.
8   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
           Table 1-6 shows the physical dimensions of the rack-mounted Power S914 and Power S924
           chassis. The server is available only in a rack-mounted form factor and takes 4U (4 EIA units)
           of rack space.
           Table 1-6 Physical dimensions of the rack-mounted Power S914 and Power S924 chassis
            Dimension                           Power S914 server            Power S924 server
                                                (9009-41A)                   (9009-42A)
Figure 1-7 shows the front view of the Power S924 server.
10   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
            High-performance Mbps DDR4 ECC memory:
              – 8 GB, 16 GB, 32 GB, or 64 GB memory.
              – Up to 1024 GB of DDR4 memory with one POWER processor.
            Storage feature:
              – Twelve SFF-3 bays/RDX bays. Optionally, split the SFF-3 bays and add a second
                integrated SAS controller without cache.
              – Eighteen SFF-3 bays/Dual IOA with Write Cache and External SAS port.
              – Twelve SFF-3 bays/RDX Bay/Dual IOA with Write Cache and External SAS port.
                Optionally, attach an EXP12SX/EXP24SX SAS HDD/SSD Expansion Drawer to the
                dual IOA.
            Up to two PCIe3 NVMe carrier cards with two M.2 module slots (with up to four
             Mainstream 400 GB SSD NVMe M.2 modules). One PCIe3 NVMe carrier card can be
             ordered only with a storage backplane. If a PCIe3 NVMe carrier card is ordered with a
             storage backplane, then the optional split feature is not supported.
            PCIe slots with single processor:
              – One x16 Gen4 full-height, half-length (CAPI).
              – One x8 Gen4 full-height, half-length (with x16 connector) (CAPI).
              – Two x8 Gen3 full-height, half-length (with x16 connectors).
              – Four x8 Gen3 full-height, half-length. (One of these slots is used for the required base
                LAN adapter.)
            Integrated:
              – Service processor.
              – EnergyScale technology.
              – Hot-swap and redundant cooling.
              – One front USB 3.0 port.
              – Two rear USB 3.0 ports.
              – Two HMC 1 GbE RJ45 ports.
              – One system port with RJ45 connector.
              – Four hot-plug, redundant power supplies.
              – 19-inch rack-mounting hardware (4U).
12   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.4.4 Minimum features
           The minimum Power S922 or Power S914 initial order must include a processor module, two
           8 GB DIMMs, two power supplies and power cords, an operating system indicator, a cover set
           indicator, and a Language Group Specify. Also, it must include one of the following storage
           options and one of the following network options:
            Storage options:
              – For boot from NVMe: One NVMe carrier and one NVMe M.2 Module.
              – For boot from a local SFF-3 HDD/SDD: One storage backplane and one SFF-3 HDD or
                SDD.
              – For boot from SAN: An internal HDD or SSD and RAID card are not required if Boot
                from SAN (#0837) is selected. A Fibre Channel adapter must be ordered if #0837 is
                selected.
            Network options:
              – One PCIe2 4-port 1 Gb Ethernet adapter.
              – One of the supported 10 Gb Ethernet adapters.
           The minimum Power S924 initial order must include a processor module, two 8 GB DIMMs,
           four power supplies and power cords, an operating system indicator, a cover set indicator, and
           a Language Group Specify. Also, it must include one of the following storage options and one
           of the following network options:
            Storage options:
              – For boot from NVMe: One NVMe carrier and one NVMe M.2 Module.
              – For boot from a local SFF-3 HDD/SDD: One storage backplane and one SFF-3 HDD or
                SDD.
              – For boot from SAN: An internal HDD or SSD and RAID card are not required if Boot
                from SAN (#0837) is selected. A Fibre Channel adapter must be ordered if #0837 is
                selected.
            Network options:
              – One PCIe2 4-port 1 Gb Ethernet adapter.
              – One of the supported 10 Gb Ethernet adapters.
Table 1-7 summarizes the processor features that are available for the Power S922 server.
14   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Power S914 processor modules
A maximum of one processor with four processor cores (#EP10), one processor with six
processor cores (#EP11), or one processor with eight processor cores (#EP12) is allowed. All
processor cores must be activated. The following list defines the allowed quantities of
processor activation entitlements:
 One 4-core, typical 2.3 - 3.8 GHz (maximum) processor (#EP10) requires that four
  processor activation codes be ordered. A maximum of four processor activations (#EP40)
  is allowed.
 One 6-core, typical 2.3 - 3.8 GHz (maximum) processor (#EP11) requires that six
  processor activation codes be ordered. A maximum of six processor activation FCs
  (#EP41) is allowed.
 One 8-core, typical 2.8 - 3.8 GHz (maximum) processor (#EP12) requires that eight
  processor activation codes be ordered. A maximum of eight processor activation FCs
  (#EP42) is allowed.
Table 1-8 summarizes the processor features that are available for the Power S914 server.
16   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
          Table 1-9 summarizes the processor features that are available for the Power S924 server.
          Table 1-10 lists the memory features that are available for the Power S922, Power S914, and
          Power S924 servers.
          Table 1-10 Summary of memory features for Power S922, Power S914, and Power S924 servers
           Feature code            DIMM capacity            Minimum quantity        Maximum quantity
                                                                                    S922, and S924/S914
#EM60 8 GB 0 32/16a
#EM62 16 GB 0 32/16
#EM63 32 GB 0 32/16
#EM64 64 GB 0 32/16
              The x16 slots can provide up to twice the bandwidth of x8 slots because they offer twice as
              many PCIe lanes. PCIe Gen4 slots can support up to twice the bandwidth of a PCIe Gen3
              slot, and PCIe Gen3 slots can support up to twice the bandwidth of a PCIe Gen2 slot,
              assuming an equivalent number of PCIe lanes.
              At least one PCIe Ethernet adapter is required on the server by IBM to ensure proper
              manufacture, test, and support of the server. One of the x8 PCIe slots is used for this required
              adapter.
              These servers are smarter about energy efficiency when cooling the PCIe adapter
              environment. They sense which IBM PCIe adapters are installed in their PCIe slots and, if an
              adapter requires higher levels of cooling, they automatically speed up fans to increase airflow
              across the PCIe adapters. Faster fans increase the sound level of the server.
              The Storage Backplane option (#EJ1F) provides eight SFF-3 bays and one SAS controller
              with zero write cache.
18   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
By optionally adding the Split Backplane (#EJ1H), a second integrated SAS controller with no
write cache is provided, and the eight SSF-3 bays are logically divided into two sets of four
bays. Each SAS controller independently runs one of the four-bay sets of drives.
Four backplane options are available for the Power S914 and Power S924 servers:
 Base Storage Backplane 12 SFF-3 bays/RDX bay (#EJ1C)
 6 +6 SFF-3 bays split backplane for #EJ1C (#EJ1E)
 Expanded function Storage Backplane 18 SFF-3 bays/Dual IOA with Write Cache and
  optional external SAS port (#EJ1D)
 Expanded function Storage Backplane 12 SFF-3 bays/RDX bay/Dual IOA with Write
  Cache and optional external SAS port (#EJ1M)
The Base Storage Backplane option (#EJ1C) provides 12 SFF-3 bays and one SAS controller
with zero write cache.
By optionally adding the Split Backplane (#EJ1E), a second integrated SAS controller with no
write cache is provided, and the 12 SSF-3 bays are logically divided into two sets of six bays.
Each SAS controller independently runs one of the six-bay sets of drives.
The backplane options provide SFF-3 SAS bays in the system unit. These 2.5-inch or SFF
SAS bays can contain SAS drives (HDD or SSD) mounted on a Gen3 tray or carrier. Thus,
the drives are designated SFF-3. SFF-1 or SFF-2 drives do not fit in an SFF-3 bay. All SFF-3
bays support concurrent maintenance or hot-plug capability.
These backplane options use leading-edge, integrated SAS RAID controller technology that
is designed and patented by IBM. A custom-designed PowerPC® based ASIC chip is the
basis of these SAS RAID controllers, and provides RAID 5 and RAID 6 performance levels,
especially for SSD. Internally, SAS ports are implemented and provide plenty of bandwidth.
The integrated SAS controllers are placed in dedicated slots and do not reduce the number of
available PCIe slots.
This backplane option supports HDDs or SSDs or a mixture of HDDs and SSDs in the SFF-3
bays. Mixing HDDs and SSDs applies even within a single set of six bays of the split
backplane option.
 Note: If mixing HDDs and SSDs, they must be in separate arrays (unless you use the
 IBM Easy Tier® function).
This backplane option can offer different drive protection options: RAID 0, RAID 5, RAID 6, or
RAID 10. RAID 5 requires a minimum of three drives of the same capacity. RAID 6 requires a
minimum of four drives of the same capacity. RAID 10 requires a minimum of two drives.
Hot-spare capability is supported by RAID 5, RAID 6, or RAID 10.
This backplane option is supported by AIX and Linux, VIOS, and IBM i. It is highly
recommended but not required that the drives be protected. With IBM i, all drives are required
to be protected by either RAID or mirroring.
Unlike the hot-plug PCIe slots and SAS bays, concurrent maintenance is not available for the
integrated SAS controllers. Scheduled downtime is required if a service action is required for
these integrated resources.
Table 1-11 shows the available disk drive FCs that can be installed in the Power S922 server.
Table 1-11 Disk drive feature code description for the Power S922 server
 Feature      CCIN      Description                                               Maximum     OS support
 code
 ESNM         5B43      300 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672         AIX and Linux
                        (AIX/Linux)
1953 19B1 300 GB 15k RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
 ESNK         5B41      300 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       8           AIX and Linux
                        (AIX/Linux)
ESDB 59E0 300 GB 15K RPM SAS SFF-3 Disk Drive (AIX/Linux) 8 AIX and Linux
ES94 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ES90 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESGV 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESB0 5B19 387 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESB2 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESB8 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESBA 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESGT 5B19 387 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 8 AIX and Linux
1964 19B3 600 GB 10k RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
ESEV 59D2 600 GB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESD5 59D0 600 GB 10K RPM SAS SFF-3 Disk Drive (AIX/Linux) 8 AIX and Linux
ESF5 59D3 600 GB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 8 AIX and Linux
 ESNR         5B47      600 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672         AIX and Linux
                        (AIX/Linux)
 ESNP         5B45      600 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       8           AIX and Linux
                        (AIX/Linux)
ESNA 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESNC 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESGZ 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESB4 5B1A 775 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESB6 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBE 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESBG 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESGX 5B1A 775 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 8 AIX and Linux
20      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Feature   CCIN   Description                                                   Maximum       OS support
code
ESJ0 5B29 931 GB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ8 5B2B 931 GB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESF3 59DA 1.2 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESF9 59DB 1.2 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 8 AIX and Linux
ESNE 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBJ 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESBL 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESNG 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESFT 59DD 1.8 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESFV 59DE 1.8 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 8 AIX and Linux
ESJ2 5B21 1.86 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJA 5B20 1.86 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ESJ4 5B2D 3.72 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJC 5B2C 3.72 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ES62      5B1D   3.86-4.0 TB 7200 RPM 4K SAS LFF-1 Nearline Disk Drive         336           AIX and Linux
                 (AIX/Linux)
ESJ6 5B2F 7.45 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJE 5B2E 7.45 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 8 AIX and Linux
ES64      5B1F   7.72-8.0 TB 7200 RPM 4K SAS LFF-1 Nearline Disk Drive         336           AIX and Linux
                 (AIX/Linux)
EQ62 5B1D Quantity 150 of #ES62 3.86-4.0 TB 7200 rpm 4k LFF-1 Disk 2
EQ64 5B1F Quantity 150 of #ES64 7.72-8.0 TB 7200 rpm 4k LFF-1 Disk 2
Table 1-12 shows the available disk drive FCs that can be installed in the Power S914 server.
Table 1-12 Disk drive feature code description for the Power S914 server
 Feature       CCIN      Description                                              Maximum     OS support
 code
1948 19B1 283 GB 15k RPM SAS SFF-2 Disk Drive (IBM i) 672 IBM i
 ESNJ          5B41      283 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive      18          IBM i
                         (IBM i)
ESDA 59E0 283 GB 15K RPM SAS SFF-3 Disk Drive (IBM i) 18 IBM i
 ESNM          5B43      300 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive      672         AIX and Linux
                         (AIX/Linux)
1953 19B1 300 GB 15k RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
 ESNK          5B41      300 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive      18          AIX and Linux
                         (AIX/Linux)
ESDB 59E0 300 GB 15K RPM SAS SFF-3 Disk Drive (AIX/Linux) 18 AIX and Linux
ES94 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ES95 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ES90 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ES91 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESGV 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESG6 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for IBM i 336 IBM i
ESGT 5B19 387 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB0 5B19 387 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB2 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
22      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Feature   CCIN   Description                                               Maximum       OS support
code
ESB8 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB9 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBA 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBB 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
1962 19B3 571 GB 10k RPM SAS SFF-2 Disk Drive (IBM i) 672 IBM i
ESEU 59D2 571 GB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4224 672 IBM i
ESD4 59D0 571 GB 10K RPM SAS SFF-3 Disk Drive (IBM i) 18 IBM i
ESF4 59D3 571 GB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4224 18 IBM i
ESNQ      5B47   571 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672           IBM i
                 (IBM i)
ESDN 59CF 571 GB 15K RPM SAS SFF-2 Disk Drive - 528 Block (IBM i) 672 IBM i
ESNN      5B45   571 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       18            IBM i
                 (IBM i)
1964 19B3 600 GB 10k RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
ESEV 59D2 600 GB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESD5 59D0 600 GB 10K RPM SAS SFF-3 Disk Drive (AIX/Linux) 18 AIX and Linux
ESF5 59D3 600 GB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 18 AIX and Linux
ESNR      5B47   600 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672           AIX and Linux
                 (AIX/Linux)
ESDP      59CF   600 GB 15K RPM SAS SFF-2 Disk Drive - 5xx Block           672           AIX and Linux
                 (AIX/Linux)
ESNP      5B45   600 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       18            AIX and Linux
                 (AIX/Linux)
ESNA 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESNB 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESNC 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESND 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESGZ 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESGG 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for IBM i 336 IBM i
ESGX 5B1A 775 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB4 5B1A 775 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB6 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBE 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESBF 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBG 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBH 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJ0 5B29 931 GB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ1 5B29 931 GB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJ8 5B2B 931 GB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJ9 5B2B 931 GB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESD2 59CD 1.1 TB 10K RPM SAS SFF-2 Disk Drive (IBM i) 672 IBM i
ESF2 59DA 1.1 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4224 672 IBM i
ESF8 59DB 1.1 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4224 18 IBM i
ESD3 59CD 1.2 TB 10K RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
ESF3 59DA 1.2 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESF9 59DB 1.2 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 18 AIX and Linux
ESNE 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESNF 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESNG 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESNH 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBJ 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESBK 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBL 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBM 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESFS 59DD 1.7 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4224 672 IBM i
ESFU 59DE 1.7 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4224 18 IBM i
ESFT 59DD 1.8 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESFV 59DE 1.8 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 18 AIX and Linux
ESJ2 5B21 1.86 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ3 5B21 1.86 TB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJA 5B20 1.86 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJB 5B20 1.86 TB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESJ4 5B2D 3.72 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ5 5B2D 3.72 TB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJC 5B2C 3.72 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJD 5B2C 3.72 TB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
 ES62          5B1D      3.86-4.0 TB 7200 RPM 4K SAS LFF-1 Nearline Disk Drive   336       AIX and Linux
                         (AIX/Linux)
24      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Feature   CCIN   Description                                                Maximum       OS support
code
ESJ6 5B2F 7.45 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ7 5B2F 7.45 TB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJE 5B2E 7.45 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJF 5B2E 7.45 TB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ES64      5B1F   7.72-8.0 TB 7200 RPM 4K SAS LFF-1 Nearline Disk Drive      336           AIX and Linux
                 (AIX/Linux)
EQ62 5B1D Quantity 150 of #ES62 3.86-4.0 TB 7200 rpm 4k LFF-1 Disk 2
EQ64 5B1F Quantity 150 of #ES64 7.72-8.0 TB 7200 rpm 4k LFF-1 Disk 2
EQDN      59CF   Quantity 150 of #ESDN (571 GB 15K RPM SAS SFF-2 for IBM    4
                 i)
EQDP      59CF   Quantity 150 of #ESDP (600 GB 15K RPM SAS SFF-2 for        4
                 AIX/LINUX)
26      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                 Table 1-13 shows the available disk drive FCs that can be installed in the Power S924 server.
Table 1-13 Disk drive feature code description for the Power S924 server
 Feature       CCIN      Description                                               Maximum       OS support
 code
 ESNL          5B43      283 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672           IBM i
                         (IBM i)
1948 19B1 283 GB 15k RPM SAS SFF-2 Disk Drive (IBM i) 672 IBM i
 ESNJ          5B41      283 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       18            IBM i
                         (IBM i)
ESDA 59E0 283 GB 15K RPM SAS SFF-3 Disk Drive (IBM i) 18 IBM i
 ESNM          5B43      300 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672           AIX and Linux
                         (AIX/Linux)
1953 19B1 300 GB 15k RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
 ESNK          5B41      300 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       18            AIX and Linux
                         (AIX/Linux)
ESDB 59E0 300 GB 15K RPM SAS SFF-3 Disk Drive (AIX/Linux) 18 AIX and Linux
ES94 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ES95 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ES90 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ES91 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESGV 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESG6 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for IBM i 336 IBM i
ESGT 5B19 387 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB0 5B19 387 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB2 5B16 387 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESB8 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB9 5B13 387 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBA 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBB 5B10 387 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
1962 19B3 571 GB 10k RPM SAS SFF-2 Disk Drive (IBM i) 672 IBM i
ESEU 59D2 571 GB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4224 672 IBM i
ESD4 59D0 571 GB 10K RPM SAS SFF-3 Disk Drive (IBM i) 18 IBM i
ESF4 59D3 571 GB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4224 18 IBM i
 ESNQ          5B47      571 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive       672           IBM i
                         (IBM i)
ESDN 59CF 571 GB 15K RPM SAS SFF-2 Disk Drive - 528 Block (IBM i) 672 IBM i
 ESNN          5B45      571 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive       18            IBM i
                         (IBM i)
1964 19B3 600 GB 10k RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
ESEV 59D2 600 GB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESD5 59D0 600 GB 10K RPM SAS SFF-3 Disk Drive (AIX/Linux) 18 AIX and Linux
ESF5 59D3 600 GB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 18 AIX and Linux
 ESNR          5B47      600 GB 15K RPM SAS SFF-2 4k Block Cached Disk Drive     672       AIX and Linux
                         (AIX/Linux)
 ESDP          59CF      600 GB 15K RPM SAS SFF-2 Disk Drive - 5xx Block         672       AIX and Linux
                         (AIX/Linux)
 ESNP          5B45      600 GB 15K RPM SAS SFF-3 4k Block Cached Disk Drive     18        AIX and Linux
                         (AIX/Linux)
ESNA 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESNB 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESNC 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESND 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESGZ 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESGG 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for IBM i 336 IBM i
ESGX 5B1A 775 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB4 5B1A 775 GB Enterprise SAS 5xx SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESB6 5B17 775 GB Enterprise SAS 5xx SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBE 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESBF 5B14 775 GB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBG 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBH 5B11 775 GB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJ0 5B29 931 GB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ1 5B29 931 GB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJ8 5B2B 931 GB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJ9 5B2B 931 GB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESD2 59CD 1.1 TB 10K RPM SAS SFF-2 Disk Drive (IBM i) 672 IBM i
ESF2 59DA 1.1 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4224 672 IBM i
ESF8 59DB 1.1 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4224 18 IBM i
ESD3 59CD 1.2 TB 10K RPM SAS SFF-2 Disk Drive (AIX/Linux) 672 AIX and Linux
ESF3 59DA 1.2 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESF9 59DB 1.2 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 18 AIX and Linux
ESNE 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
28      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Feature   CCIN   Description                                                 Maximum       OS support
code
ESNF 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESNG 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESNH 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBJ 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESBK 5B15 1.55 TB Enterprise SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESBL 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESBM 5B12 1.55 TB Enterprise SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESFS 59DD 1.7 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4224 672 IBM i
ESFU 59DE 1.7 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4224 18 IBM i
ESFT 59DD 1.8 TB 10K RPM SAS SFF-2 Disk Drive 4K Block - 4096 672 AIX and Linux
ESFV 59DE 1.8 TB 10K RPM SAS SFF-3 Disk Drive 4K Block - 4096 18 AIX and Linux
ESJ2 5B21 1.86 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ3 5B21 1.86 TB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJA 5B20 1.86 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJB 5B20 1.86 TB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ESJ4 5B2D 3.72 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ5 5B2D 3.72 TB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJC 5B2C 3.72 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJD 5B2C 3.72 TB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ES62      5B1D   3.86-4.0 TB 7200 RPM 4K SAS LFF-1 Nearline Disk Drive       336           AIX and Linux
                 (AIX/Linux)
ESJ6 5B2F 7.45 TB Mainstream SAS 4k SFF-2 SSD for AIX/Linux 336 AIX and Linux
ESJ7 5B2F 7.45 TB Mainstream SAS 4k SFF-2 SSD for IBM i 336 IBM i
ESJE 5B2E 7.45 TB Mainstream SAS 4k SFF-3 SSD for AIX/Linux 18 AIX and Linux
ESJF 5B2E 7.45 TB Mainstream SAS 4k SFF-3 SSD for IBM i 18 IBM i
ES64      5B1F   7.72-8.0 TB 7200 RPM 4K SAS LFF-1 Nearline Disk Drive       336           AIX and Linux
                 (AIX/Linux)
EQ62 5B1D Quantity 150 of #ES62 3.86-4.0 TB 7200 rpm 4k LFF-1 Disk 2
EQ64 5B1F Quantity 150 of #ES64 7.72-8.0 TB 7200 rpm 4k LFF-1 Disk 2
 EQDN          59CF      Quantity 150 of #ESDN (571 GB 15K RPM SAS SFF-2 for     4
                         IBM i)
 EQDP          59CF      Quantity 150 of #ESDP (600 GB 15K RPM SAS SFF-2 for     4
                         AIX/LINUX)
30      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Feature   CCIN    Description                                                Maximum       OS support
code
           The RDX docking station #EUA4 accommodates RDX removable disk cartridges of any
           capacity. The disk is in a protective rugged cartridge enclosure that plugs into the docking
           station. The docking station holds one removable rugged disk drive/cartridge at a time. The
           rugged removable disk cartridge and docking station perform saves, restores, and backups
           similar to a tape drive. This docking station can be an excellent entry capacity/performance
           option.
           The Standalone USB DVD drive (#EUA5) is an optional, stand-alone external USB-DVD
           device. It requires high current at 5V and must use the front USB 3.0 port.
1.6 I/O drawers for Power S922, Power S914, and Power S924
servers
           If more Gen3 PCIe slots beyond the system node slots are required, PCIe Gen3 I/O drawers
           can be attached to the Power S922, Power S914, and Power S924 servers.
            Note: The PCIe expansion drawer (#EMX0) and EXP24SX /EXP12SX SAS Storage
            Enclosures (#ESLS/ELLS or #ESLL/ELLL) do not apply to the 4-core configuration Power
            S914 server.
           The 7226-1U3 offers a 1U rack-mountable dual bay enclosure with storage device options of
           LTO5, 6, 7, and 8 tape drives with both SAS and Fibre Channel interface. The 7226 also
           offers DVD-RAM SAS and USB drive features and RDX 500 GB, 1 TB, and 2 TB drive
           options. Up to two drives (or four DVD-RAM) can be installed in any combination in the 7226
           enclosure.
            Note: The tower system configuration of Power S914 does not allow attachment of any I/O
            expansion drawer.
              The PCIe Gen3 I/O Expansion Drawer has two redundant, hot-plug power supplies. Each
              power supply has its own separately ordered power cord. The two power cords plug into a
              power supply conduit that connects to the power supply. The single-phase AC power supply is
              rated at 1030 W and can use 100 - 120V or 200 - 240V. If using 100 - 120V, then the
              maximum is 950 W. It is a preferred practice that the power supply connects to a power
              distribution unit (PDU) in the rack. Power Systems PDUs are designed for a 200 - 240V
              electrical source.
              A blind swap cassette (BSC) is used to house the full-height adapters that go into these slots.
              The BSC is the same BSC that is used with the previous generation server's 12X attached I/O
              drawers (#5802, #5803, #5877, and #5873). The drawer is shipped with a full set of BSCs,
              even if the BSCs are empty.
              Concurrent repair and add/removal of PCIe adapters are done by HMC-guided menus or by
              operating system support utilities.
32   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.6.2 I/O drawers and usable PCI slots
           Figure 1-9 shows the rear view of the PCIe Gen3 I/O expansion drawer that is equipped with
           two PCIe3 6-slot fan-out modules with the location codes for the PCIe adapter slots.
Figure 1-9 Rear view of a PCIe Gen3 I/O expansion drawer with PCIe slots location codes
              Table 1-14 PCIe slot locations for the PCIe Gen3 I/O expansion drawer with two fan-out modules
               Slot                             Location code                    Description
              In the table:
               All slots support full-length, regular-height adapters or short (LP) adapters with a
                regular-height tailstock in single-wide, Gen3, BSC.
               Slots C1 and C4 in each PCIe3 6-slot fan-out module are x16 PCIe3 buses, and slots C2,
                C3®, C5, and C6 are x8 PCIe buses.
               All slots support enhanced error handling (EEH).
               All PCIe slots are hot-swappable and support concurrent maintenance.
              Table 1-15 summarizes the maximum number of I/O drawers that are supported and the total
              number of PCI slots that are available.
              Table 1-15 Maximum number of I/O drawers that are supported and total number of PCI slots
               Server                    Maximum number of       Maximum number of       Maximum PCIe slots
                                         I/O Exp Drawers         I/O fan-out modules
34   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.6.3 EXP24SX (#ESLS) and EXP12SX (#ESLL) SAS Storage Enclosures
          If you need more disks than are available with the internal disk bays, you can attach more
          external disk subsystems, such as an EXP24SX SAS Storage Enclosure (#ESLS) or
          EXP12SX SAS Storage Enclosure (#ESLL).
          The EXP24SX is a storage expansion enclosure with twenty-four 2.5-inch SFF SAS bays. It
          supports up to 24 hot-plug HDDs or SSDs in only 2 EIA of space in a 19-inch rack. The
          EXP24SX SFF bays use SFF Gen2 (SFF-2) carriers or trays.
          The EXP12SX is a storage expansion enclosure with twelve 3.5-inch large form factor (LFF)
          SAS bays. It supports up to 12 hot-plug HDDs in only 2 EIA of space in a 19-inch rack. The
          EXP12SX SFF bays use LFF Gen1 (LFF-1) carriers/trays. The 4 KB sector drives (#4096 or
          #4224) are supported. SSDs are not supported.
          With AIX, Linux, and VIOS, the EXP24SX or the EXP12SX can be ordered with four sets of
          six bays (mode 4), two sets of 12 bays (mode 2), or one set of 24-four bays (mode 1). With
          IBM i, only one set of 24 bays (mode 1) is supported. It is possible to change the mode setting
          in the field by using software commands along with a specifically documented procedure.
           Important: When changing modes, a skilled, technically qualified person should follow the
           special documented procedures. Improperly changing modes can potentially destroy
           existing RAID sets, prevent access to existing data, or allow other partitions to access
           another partition's existing data.
          Four mini-SAS HD ports on the EXP24SX or EXP12SX are attached to PCIe Gen3 SAS
          adapters or attached to an integrated SAS controller in the Power S914, Power S922, or
          Power S924 servers.
          The attachment between the EXP24SX or EXP12SX and the PCIe3 SAS adapters or
          integrated SAS controllers is through SAS YO12 or X12 cables. All ends of the YO12 and X12
          cables have mini-SAS HD narrow connectors.
The EXP24SX or EXP12SX includes redundant AC power supplies and two power cords.
              The IBM Power S914 Solution Edition supports 4-core (#4928) and 6-core (#4927)
              configurations. For a list of participating ISVs, a registration form, and more details, see IBM i
              Solution Editions.
              Here are the requirements to be eligible to purchase an IBM Solution Edition for Power S914:
               The server must be ordered with a Solution Edition FC.
               The server must be a new purchase, not a miscellaneous execution system (MES)
                upgrade.
               The offering must include new or upgraded software licenses, or software maintenance
                from the ISV for the qualifying IBM server. Services and training for the qualifying server
                can also be provided. Contact the individual ISV for details.
               Proof of purchase (for example, a copy of the invoice) of the solution with a participating
                ISV must be provided to IBM on request. The proof must be dated within 90 days before or
                after the date of order of the qualifying server.
               The combined value (software, maintenance, services, and training) of the ISV purchase
                must be US $6,500 or greater. Contact the individual ISV for details.
               Not eligible as an IBM Capacity BackUp (CBU).
36   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.8 IBM Capacity BackUp
          The CBU designation enables you to temporarily transfer IBM i processor license entitlements
          and IBM i user license entitlements that are purchased for a primary machine to a secondary
          CBU-designated system for HADR operations. Temporarily transferring these resources
          instead of purchasing them for your secondary system can result in significant savings.
          Processor activations cannot be transferred.
          If your primary or CBU machine is sold or discontinued from use, any temporary entitlement
          transfers must be returned to the machine on which they were originally acquired. For CBU
          registration, terms, and conditions, and more information, see IBM Power Systems Capacity
          BackUp.
          After a new CBU system is registered as a pair with the proposed primary system and the
          configuration is approved, you can temporarily move your optional IBM i processor license
          entitlement and IBM i user license entitlements from the primary system to the CBU system
          when the primary system is down or while the primary system processors are inactive. The
          CBU system can then support failover and role-swapping for a full range of test, disaster
          recovery, and high availability scenarios. Temporary entitlement transfer means that the
          entitlement is a property that is transferred from the primary system to the CBU system,
          which may remain in use on the CBU system while the registered primary and CBU systems
          are deployed for the high availability or disaster recovery operation. The intent of the CBU
          offering is to enable regular role-swap operations.
          The primary systems for a Power S922 (9009-22A) CBU server with its IBM i P10 software
          tier can be a POWER9 or POWER8 server with a P10 or P20 software tie, as shown in the
          following list:
             Power S824 (8286-42A) server
             Power S814 6-core or 8-core (8286-41A) server
             Power S822 (8284-22A) server
             Power S924 (9009-42A) server
             Power S914 6-core or 8-core (9009-41A) server
             Power S922 (9009-22A) server
          The primary machine must be in the same enterprise as the CBU system. The IBM i Solution
          Editions are not eligible for CBU status.
              Before you can temporarily transfer IBM i user entitlements, you must have more than the
              minimum number of IBM i user entitlements on a primary system. You can then transfer any
              IBM i user entitlements above the minimum, assuming the total IBM i users on the primary
              system do not require the IBM i entitlement that you want to transfer during the time of the
              transfer. The Power S824 and Power S924 servers do not have user entitlements that can be
              transferred, and only processor license entitlements can be transferred. The minimum
              number of IBM i users on the POWER9 and POWER8 with IBM i user entitlements are:
                 Power S814 6-core or 8-core (8286-41A) server: Ten users
                 Power S822 (8284-22A) server: Ten users
                 Power S914 6-core or 8-core (9009-41A) server: Ten users
                 Power S922 (9009-22A) server: Ten users
              For example, if you have a Power S914 6-core server as your primary system with two IBM i
              processor license entitlements (one above the minimum) and 40 IBM i user entitlements (30
              above the minimum), you can temporarily transfer up to one IBM i entitlement and up to 30
              user entitlements. During this temporary transfer, the CBU system's internal records of its
              total number of IBM i processor and user license entitlements is not updated, and you might
              see IBM i license noncompliance warning messages from the CBU system.
              After a new CBU system is registered along with the proposed primary system and the
              configuration is approved, you can temporarily move your optional IBM i processor license
              entitlement and IBM i user license entitlements from the primary system to the CBU system
              when the primary system is down or while the primary system processors are inactive. The
              CBU system can then support failover and role-swapping for a full range of test, disaster
              recovery, and high availability scenarios. Temporary entitlement transfer means that the
              entitlement is a property that is transferred from the primary system to the CBU system,
              which may remain in use on the CBU system while the registered primary and CBU systems
              are in deployment for the high availability or disaster recovery operation. The intent of the
              CBU offering is to enable regular role-swap operations. The primary machine must be in the
              same enterprise as the CBU system. The IBM i Solution Editions are not eligible for CBU
              status.
38   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
The Power S914 server is available with six or eight cores in the P10 software tier and four
cores in the P05 software tier.
Before you can temporarily transfer IBM i user entitlements, you must have more than the
minimum number of IBM i user entitlements on a primary server. You can then transfer any
IBM i user entitlements above the minimum, assuming that the total IBM i users on the
primary system do not require the IBM i entitlement that you want to transfer during the time
of the transfer. The Power S924 and S824 servers do not have IBM i user entitlements to
transfer, only processor entitlements. For a P10 primary, the minimum number of IBM i user
entitlements on the eligible P10 POWER9 and POWER8 servers are:
   Power S814 6-core or 8-core (8286-41A) server: Ten users
   Power S822 (8284-22A) server: Ten users
   Power S914 6-core or 8-core (9009-41A) server: Ten users
   Power S922 (9009-22A) server: Ten users
Before you can temporarily transfer IBM i user entitlements, you must have more than the
minimum number of IBM i user entitlements on a primary server. You can then transfer any
IBM i user entitlements above the minimum, assuming that the total IBM i users on the
primary system do not require the IBM i entitlement that you want to transfer during the time
of the transfer. The minimum number of IBM i user entitlements on the P05 or P10 POWER9
and POWER8 with IBM i user entitlements are:
   Power S814 4-core (8286-41A) server: Five users
   Power S814 6-core or 8-core (8286-41A) server: Ten users
   Power S822 (8284-22A) server: Ten users
   Power S914 4-core (9009-41A) server: Five users
   Power S914 6-core or 8-core (9009-41A) server: Ten users
   Power S922 (9009-22A) server: Ten users
              After a new CBU system is registered along with the proposed primary system and the
              configuration is approved, you can temporarily move your optional IBM i processor license
              entitlement and IBM i Enterprise Enablement (#5250) entitlements from the primary system
              to the CBU system when the primary system is down or while the primary system processors
              are inactive. The CBU system can then support failover and role-swapping for a full range of
              test, disaster recovery, and high availability scenarios. Temporary entitlement transfer means
              that the entitlement is a property that is transferred from the primary system to the CBU
              system, which may remain in use on the CBU system while the registered primary and CBU
              systems are in deployment for the high availability or disaster recovery operation. The intent
              of the CBU offering is to enable regular role-swap operations.
              Before you can temporarily transfer Enterprise Enablement entitlements (#5250), you must
              have more than one Enterprise Enablement entitlement (#5250) on the primary server and at
              least one Enterprise Enablement entitlement (#5250) on the CBU system. You can then
              transfer the entitlements that are not required on the primary server during the time of transfer
              and that are above the minimum of one entitlement. The minimum number of permanent
              entitlements on the CBU is one; however, you are required to license all permanent
              workloads, such as replication workloads. If, for example, the replication workload uses four
              processor cores at peak workload, then you are required to permanently license four cores on
              the CBU.
              For example, if you have a 12-core Power S824 server as your primary system with six IBM i
              processor license entitlements (five above the minimum) and two Enterprise Enablement
              entitlements (#5250) (one above the minimum), you can temporarily transfer up to five IBM i
              entitlements and one Enterprise Enablement entitlement (#5250). During the temporary
              transfer, the CBU system's internal records of its total number of IBM i processor entitlements
              are not updated, and you might see IBM i license noncompliance warning messages from the
              CBU system.
40   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
        The CBU specify (#0444) is available only as part of a new server purchase. Certain system
        prerequisites must be met, and system registration and approval are required before the CBU
        specify feature can be applied on a new server. Standard IBM i terms and conditions do not
        allow either IBM i processor license entitlements or Enterprise Enablement entitlements
        (#5250) to be transferred permanently or temporarily. These entitlements remain with the
        machine for which they were ordered. When you register the association between your
        primary and on-order CBU systems, you must agree to certain terms and conditions
        regarding the temporary transfer.
        The servers with P20 or higher software tiers do not have user entitlements that can be
        transferred, and only processor license entitlements can be transferred.
        Here are the eligible primary servers for a Power S924 CBU:
             Power S824 (8286-42A) server
             Power S924 (9009-42A) server
             Power E870 (9119-MME) server
             Power E880 (9119-MHE) server
             Power E870C (9080-MME) server
             Power E880C (9080-MHE) server
            Order information: The racking approach for the initial order must be either a 7014-T00,
            7014-T42, or 7965-94Y. If an extra rack is required for I/O expansion drawers as an MES to
            an existing system, either an #0551, #0553, or #ER05 rack must be ordered.
            You must leave 2U of space at either the bottom or top of the rack, depending on the
            client's cabling preferences, to allow for cabling to exit the rack.
        If a system will be installed in a rack or cabinet that is not from IBM, ensure that the rack
        meets the requirements that are described in 1.9.10, “Original equipment manufacturer racks”
        on page 52.
            Responsibility: The client is responsible for ensuring that the installation of the drawer in
            the preferred rack or cabinet results in a configuration that is stable, serviceable, safe, and
            compatible with the drawer requirements for power, cooling, cable management, weight,
            and rail security.
               Important: If more weight is added to the top of the rack, for example, adding #6117, the
               490 kg (1080 lb) must be reduced by the weight of the addition. As an example, #6117
               weighs approximately 45 kg (100 lb) so the new maximum weight of drawers that the rack
               can support in a zone 4 earthquake environment is 445 kg (980 lb). In the zone 4
               earthquake environment, the rack must be configured starting with the heavier drawers at
               the bottom of the rack.
42   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.9.2 IBM 7014 Model T42 rack
                 The 2.0-meter (79.3-inch) Model T42 addresses the client requirement for a tall enclosure to
                 house the maximum amount of equipment in the smallest possible floor space. The following
                 features are for the model T42 rack (which differ from the model T00):
                  The T42 rack has 42U (EIA units) of usable space (6U of extra space).
                  The model T42 supports AC power only.
                  Weights are as follows:
                   – T42 base empty rack: 261 kg (575 lb)
                   – T42 full rack: 930 kg (2045 lb)
The available door options for the Model T42 rack are shown in Figure 1-12.
                   High end: For the High-End Appearance Front Door (#6250), use the High-End
                   Appearance Side Covers (#6238) to make the rack appear as though it is a high-end
                   server (but in a 19-inch rack format instead of a 24-inch rack).
               The #ERG7 provides an attractive black full height rack door. The door is steel, with a
                perforated flat front surface. The perforation pattern extends from the bottom to the top of
                the door to enhance ventilation and provide some visibility into the rack. The non-acoustic
                door has a depth of about 134 mm (5.3 in.).
              For more information about planning for the installation of the IBM Rear Door Heat
              Exchanger, see IBM Knowledge Center.
              The IBM 42U Slim Rack has a lockable perforated front steel door, providing ventilation,
              physical security, and visibility of indicator lights in the installed equipment within. In the rear,
              either a lockable perforated rear steel door (#EC02) or a lockable Rear Door Heat Exchanger
              (RDHX)(1164-95X) is used. Lockable optional side panels (#EC03) increase the rack's
              aesthetics, help control airflow through the rack, and provide physical security. Multiple 42U
              Slim Racks can be bolted together to create a rack suite (#EC04).
              Up to six optional 1U PDUs can be placed vertically in the sides of the rack. More PDUs can
              be located horizontally, but they each use 1U of space in this position.
44   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.9.5 2.0 Meter Rack (#0553)
           The 2.0 Meter Rack (#0553) is a 42 EIA unit rack. The rack that is delivered as #0553 is the
           same rack that is delivered when you order the 7014-T42 rack. The included features might
           vary. Certain features that are delivered as part of the 7014-T42 must be ordered separately
           with #0553.
           PDUs on the rack are optional. Each #7196 and #7189 PDU uses one of six vertical mounting
           bays. Each PDU beyond four uses 1U of rack space.
           If ordering Power Systems equipment in an MES order, use the equivalent rack #ER05
           instead of 7965-94Y so IBM Manufacturing can ship the hardware in the rack.
           The PDUs have 12 client-usable IEC 320-C13 outlets. There are six groups of two outlets that
           are fed by six circuit breakers. Each outlet is rated up to 10 amps, but each group of two
           outlets is fed from one 15 amp circuit breaker.
           High-function PDUs provide more electrical power per PDU and offer better “PDU footprint”
           efficiency. In addition, they are intelligent PDUs that provide insight to actual power usage by
           receptacle and also provide remote power on/off capability for easier support by individual
           receptacle. The new PDUs can be ordered as #EPTJ, #EPTL, #EPTN, and #EPTQ.
Table 1-17 lists the feature codes for the high-function PDUs announced in October 2019.
46   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Four PDUs can be mounted vertically in the back of the T00 and T42 racks. Figure 1-13
shows the placement of the four vertically mounted PDUs. In the rear of the rack, two more
PDUs can be installed horizontally in the T00 rack and three in the T42 rack. The four vertical
mounting locations are filled first in the T00 and T42 racks. Mounting PDUs horizontally uses
1U per PDU and reduces the space that is available for other racked components. When
mounting PDUs horizontally, the preferred approach is to use fillers in the EIA units that are
occupied by these PDUs to facilitate proper air-flow and ventilation in the rack.
1 2
The PDU receives power through a UTG0247 power-line connector. Each PDU requires one
PDU-to-wall power cord. Various power cord features are available for various countries and
applications by varying the PDU-to-wall power cord, which must be ordered separately. Each
power cord provides the unique design characteristics for the specific power requirements. To
match new power requirements and save previous investments, these power cords can be
requested with an initial order of the rack or with a later upgrade of the rack features.
Table 1-18 Wall power cord options for the power distribution unit and iPDU features
 Feature      Wall plug           Rated voltage       Phase      Rated amperage        Geography
 code                             (V AC)
#6654 NEMA L6-30 200-208, 240 1 24 amps US, Canada, LA, Japan
 #6492        IEC 309, 2P+G,      200-208, 240        1          48 amps               US, Canada, LA, Japan
              60A
                    Notes: Ensure that the appropriate power cord feature is configured to support the power
                    that is being supplied. Based on the power cord that is used, the PDU can supply
                    4.8 - 19.2 kVA. The power of all the drawers plugged into the PDU must not exceed the
                    power cord limitation.
                 To better enable electrical redundancy, each server has two power supplies that must be
                 connected to separate PDUs, which are not included in the base order.
                 For maximum availability, a preferred approach is to connect power cords from the same
                 system to two separate PDUs in the rack, and to connect each PDU to independent power
                 sources.
                 For detailed power requirements and power cord details about the 7014 and 7965-94Y racks,
                 see the IBM Knowledge Center at:
https://www.ibm.com/support/knowledgecenter/POWER9/p9hdx/POWER9welcome.htm
48     IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.9.8 Rack-mounting rules
           Consider the following primary rules when you mount the system into a rack:
            The system is placed at any location in the rack. For rack stability, start filling a rack from
             the bottom.
            Any remaining space in the rack can be used to install other systems or peripheral
             devices, if the maximum permissible weight of the rack is not exceeded and the installation
             rules for these devices are followed.
            Before placing the system into the service position, be sure to follow the rack
             manufacturer’s safety instructions regarding rack stability.
               Order information: The racking approach for the initial order must be either a 7014-T00,
               7014-T42, or 7965-94Y. If an extra rack is required for I/O expansion drawers as an
               Manufacturing Equipment Specification (MES) to an existing system, either an #0551,
               #0553, or #ER05 rack must be ordered.
               You must leave 2U of space at either the bottom or top of the rack, depending on the
               client's cabling preferences, to allow for cabling to exit the rack.
           The IBM System Storage 7226 Multi-Media Enclosure supports LTO Ultrium and DAT160
           Tape technology, DVD-RAM, and RDX removable storage requirements on the following IBM
           systems:
               IBM POWER6™ processor-based systems
               IBM POWER7® processor-based systems
               IBM POWER8 processor-based systems
               IBM POWER9 processor-based systems
           The IBM System Storage 7226 Multi-Media Enclosure offers an expansive list of drive feature
           options, as shown in Table 1-19.
#5763 DVD Front USB Port Sled with DVD-RAM USB Drive Available
              Removable RDX drives are in a rugged cartridge that inserts in to an RDX removable (USB)
              disk docking station (#1103 or #EU03). RDX drives are compatible with docking stations,
              which are installed internally in IBM POWER6, IBM POWER6+, POWER7,
              IBM POWER7+, POWER8, and POWER9 processor-based servers, where applicable.
              Media that is used in the 7226 DAT160 SAS and USB tape drive features are compatible with
              DAT160 tape drives that are installed internally in IBM POWER6, POWER6+, POWER7,
              POWER7+, POWER8, and POWER9 processor-based servers.
50   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Media that is used in LTO Ultrium 5 Half-High 1.5 TB tape drives are compatible with Half
High LTO5 tape drives that are installed in the IBM TS2250 and TS2350 external tape drives,
IBM LTO5 tape libraries, and half-high LTO5 tape drives that are installed internally in IBM
POWER6, POWER6+, POWER7, POWER7+, POWER8, and POWER9 processor-based
servers.
Figure 1-14 shows the IBM System Storage 7226 Multi-Media Enclosure.
The IBM System Storage 7226 Multi-Media Enclosure offers customer-replaceable unit
(CRU) maintenance service to help make the installation or replacement of new drives
efficient. Other 7226 components are also designed for CRU maintenance.
The IBM System Storage 7226 Multi-Media Enclosure is compatible with most IBM POWER6,
POWER6+, POWER7, POWER7+, POWER8, and POWER9 processor-based systems that
offer current level AIX, IBM i, and Linux operating systems.
 Unsupported: IBM i does not support 7226 USB tape devices but does support 7226 USB
 DVD devices.
For a complete list of host software versions and release levels that support the IBM System
Storage 7226 Multi-Media Enclosure, see System Storage Interoperation Center (SSIC).
 Note: Any of the existing 7216-1U2, 7216-1U3, and 7214-1U2 multimedia drawers are
 also supported.
The Model TF4 is a follow-on product to the Model TF3 and offers the following features:
 A slim, sleek, and lightweight monitor design that occupies only 1U (1.75 in.) in a 19-inch
  standard rack.
 A 18.5-inch (409.8 mm x 230.4 mm) flat panel TFT monitor with truly accurate images and
  virtually no distortion.
 The ability to mount the IBM Travel Keyboard in the 7316-TF4 rack keyboard tray.
https://www.ibm.com/support/knowledgecenter/POWER9/p9hdx/POWER9welcome.htm
Rear Door
Figure 1-15 Top view of the rack specification dimensions (not specific to IBM)
52   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
            The vertical distance between the mounting holes must consist of sets of three holes
             spaced (from bottom to top) 15.9 mm (0.625 in.), 15.9 mm (0.625 in.), and 12.67 mm (0.5
             in.) on-center, making each three-hole set of vertical hole spacing 44.45 mm (1.75 in.)
             apart on center. Rail-mounting holes must be 7.1 mm ± 0.1 mm (0.28 in. ± 0.004 in.) in
             diameter. Figure 1-16 shows the top front specification dimensions.
15.9mm 15.9mm
                                                    12.7mm                       12.7mm
                                                    15.9mm                       15.9mm
                                                    15.9mm                       15.9mm
                                                    12.7mm                       12.7mm
              The HMC management features help improve server usage, simplify systems management,
              and accelerate provisioning of server resources by using PowerVM virtualization technology.
              The HMC is available as a hardware appliance or as a vHMC. The Power S922, Power S914,
              and Power S924 servers support several service environments, including attachment to one
              or more HMCs or vHMCs. This is the default configuration for servers supporting multiple
              logical partitions with dedicated resource or virtual I/O.
              Hardware support for customer replaceable units (CRUs) come standard with the HMC. In
              addition, users can upgrade this support level to IBM onsite support to be consistent with
              other Power Systems servers.
               Note:
                An HMC or vHMC is required for the Power S922, Power S914, and Power S924
                 servers.
                Integrated Virtual Management (IVM) is no longer supported.
For more information about vHMC, see Virtual HMC Appliance (vHMC) Overview.
Figure 1-17 shows the HMC model selections and tier updates.
Old HMC virtual appliance HMC virtual appliance for x86 systems
54   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
1.10.3 Hardware Management Console code level
          The HMC code must be running at Version 9 Release 1 modification 920 (V9R1M920) or later
          when you use the HMC with the Power S922, Power S914, and Power S924 servers.
          If you are attaching an HMC to a new server or adding a function to an existing server that
          requires a firmware update, the HMC machine code might need to be updated to support the
          firmware level of the server. In a dual-HMC configuration, both HMCs must be at the same
          version and release of the HMC code.
          To determine the HMC machine code level that is required for the firmware level on any
          server, go to Fix Level Recommendation Tool (FLRT) on or after the planned availability date
          for this product.
FLRT identifies the correct HMC machine code for the selected system firmware level.
           Note:
            Access to firmware and machine code updates is conditional on entitlement and license
             validation in accordance with IBM policy and practice. IBM might verify entitlement
             through customer number, serial number electronic restrictions, or any other means or
             methods that are employed by IBM at its discretion.
            HMC V9 supports only the Enhanced+ version of the GUI. The Classic version is no
             longer available.
            HMC V9R1.911.0 added support for managing IBM OpenPOWER systems. The same
             HMC that is used to manage flexible service processor (FSP)-based enterprise
             systems can manage the baseboard management controller (BMC) based Power
             Systems AC and Power Systems LC servers. This support provides a consistent and
             consolidated hardware management solution.
            HMC V9 supports connections to servers that are based on IBM servers that are based
             on POWER9, POWER8, and POWER7 processors. There is no support in this release
             for servers that are based on POWER6 processors or earlier.
          You may use either architecture to manage the servers. You also may use one Intel-based
          HMC and one POWER8 based HMC if the software is at the same level.
          It is a preferred practice to use the new POWER8 processor-based consoles for server
          management.
          Intel-based HMCs
          HMCs that are based on Intel processors that support V9 code are:
           7042-CR9
           7042-CR8
           7042-CR7
              All future HMC development will be done for the POWER8 processor-based 7063-CR1 model
              and its successors.
                  Note: System administrators can remotely start or stop a 7063-CR1 HMC by using
                  ipmitool or WebUI.
              For the HMC to communicate properly with the managed server, eth0 of the HMC must be
              connected to either the HMC1 or HMC2 ports of the managed server, although other network
              configurations are possible. You may attach a second HMC to the remaining HMC port of the
              server for redundancy. The two HMC ports must be addressed by two separate subnets.
56   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
           Figure 1-18 shows a simple network configuration to enable the connection from the HMC to
           the server and to allow for dynamic LPAR operations. For more information about HMC and
           the possible network connections, see IBM Power Systems HMC Implementation and Usage
           Guide, SG24-7491.
Management LAN
Figure 1-18 Network connections from the HMC to service processor and LPARs
           By default, the service processor HMC ports are configured for dynamic IP address
           allocation. The HMC can be configured as a DHCP server, providing an IP address at the
           time that the managed server is powered on. In this case, the FSP is allocated an IP address
           from a set of address ranges that is predefined in the HMC software.
           If the service processor of the managed server does not receive a DHCP reply before
           timeout, predefined IP addresses are set up on both ports. Static IP address allocation is also
           an option and can be configured by using the Advanced System Management Interface
           (ASMI) menus.
            Notes: The two service processor HMC ports have the following features:
             1 Gbps connection speed.
             Visible only to the service processor. They can be used to attach the server to an HMC
              or to access the ASMI options from a client directly from a client web browser.
             Use the following network configuration if no IP addresses are set:
               – Service processor eth0 (HMC1 port): 169.254.2.147 with netmask 255.255.255.0
               – Service processor eth1 (HMC2 port): 169.254.3.147 with netmask 255.255.255.0
                Figure 1-19 shows one possible highly available HMC configuration that manages two
                servers. Each HMC is connected to one FSP port of each managed server.
                For simplicity, only the hardware management networks (LAN1 and LAN2) are highly
                available. However, the open network (LAN3) can be made highly available by using a similar
                concept and adding a second network between the partitions and HMCs.
                For more information about redundant HMCs, see IBM Power Systems HMC Implementation
                and Usage Guide, SG24-7491.
58    IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                                                                                                  2
                 The speeds that are shown are at an individual component level. Multiple components and
                 application implementation are key to achieving the best performance.
                 Always do the performance sizing at the application workload environment level and evaluate
                 performance by using real-world performance measurements and production workloads.
                                                                                                       340 GB/s
                                                                                                    memory bandwidth
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                                                                                                                                                                     POWER9
                                DDR4 DIMM
                                                                                                                                                                                      SCM 1
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                                                                                                                                                                                                                                                                                                HDD/SDD 1
                                                                                                                                                                                            16Gbps
                               DDR4 DIMM                                                                                                                                                                                                                                                                        HDD/SDD 2
                               DDR4 DIMM
                               DDR4 DIMM                                                                                                                                                                                                                                                                        HDD/SDD 3
                               DDR4 DIMM
                                                                                                                                                                                                                                                                                                                HDD/SDD 4
                               DDR4 DIMM
                               DDR4 DIMM                                                                                                                                                                                                                                                                        HDD/SDD 5
                               DDR4 DIMM
                               DDR4 DIMM                                                                                                                                                                                                                                                                        HDD/SDD 6
                                                                                                                                                                                     POWER9
                               DDR4 DIMM                                                                                                                                                                                                                                                    4+4 dual SAS
                                                                                                                                                                                      SCM 0                                                                                                                     HDD/SDD 7
                               DDR4 DIMM                                                                                                                                                                                                                                                      as internal
                               DDR4 DIMM
                                                                                                                                                                                                                                                                                               storage          HDD/SDD 8
x8
                                                                                                                                                                                                                                                                              x8
                                                                                                                                                          x8                         x16                                                                                                  1st SAS               2nd SAS
                                                                                                                                                          16 GB/s
16 GB/s
                                                                                                                                                                                                                                                                               x4
                                                                                                                                                                                                                                                                                                      NVMe      NVMe SSD
                                                                                                                      PEX                                                                                                            PEX                                       x4
                                                                                                                                                                                                                                                                                                      slot1
                                                                                                                                                                                                                                                                                                                NVMe SSD
                                                                                                                                                                                                                                                                         4x USB 3.0
                                                                                                                                                                                                                                   PCIe Gen3 x8 C11
PCIe Gen3 x8 C6
PCIe Gen3 x8 C7
60   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                             170 GB/s                                                                                                                                                                                                    HDD/SDD 1
                          memory bandwidth
          DDR4 DIMM                                                                                                                                                                                                                      HDD/SDD 2
          DDR4 DIMM
          DDR4 DIMM                                                                                                                                                                                                                      HDD/SDD 3
          DDR4 DIMM
                                                                                                                                                                                                                                         HDD/SDD 4
          DDR4 DIMM
          DDR4 DIMM                                                                                                                                                                                                                      HDD/SDD 5
          DDR4 DIMM
          DDR4 DIMM                                                                                                                                                                                                                      HDD/SDD 6
                                                                                                            POWER9
          DDR4 DIMM                                                                                                                                                                                               6+6 Dual SAS
                                                                                                              SCM                                                                                                                        HDD/SDD 7
          DDR4 DIMM                                                                                                                                                                                                 as internal
          DDR4 DIMM
                                                                                                                                                                                                                     storage             HDD/SDD 8
                                                                                                                                                                                                            x8
                                                                                 x8                         x16                                                                                                  1st SAS                 2nd SAS
                                                                                  16 GB/s
16 GB/s
                                                                                                                                                                                                            x4   NVMe                     NVMe
                                             PEX                                                                                                              PEX
                                                                                                                                                                                                                 slot1                    slot2
x4
PCIe Gen3 x8 C6
PCIe Gen3 x8 C7
4 x USB 3.0
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                DDR4 DIMM
                                                                                                                                                                                                                                                                                                                 HDD/SDD 1
                                                                                                                                                                                             16Gbps
                               DDR4 DIMM                                                                                                                                                                                                                                                                         HDD/SDD 2
                               DDR4 DIMM
                               DDR4 DIMM                                                                                                                                                                                                                                                                         HDD/SDD 3
                               DDR4 DIMM
                                                                                                                                                                                                                                                                                                                 HDD/SDD 4
                               DDR4 DIMM
                               DDR4 DIMM                                                                                                                                                                                                                                                                         HDD/SDD 5
                               DDR4 DIMM
                               DDR4 DIMM                                                                                                                                                                                                                                                                         HDD/SDD 6
                                                                                                                                                                                      POWER9
                               DDR4 DIMM                                                                                                                                                                                                                                                   6+6 Dual SAS
                                                                                                                                                                                       SCM 0                                                                                                                     HDD/SDD 7
                               DDR4 DIMM                                                                                                                                                                                                                                                     as internal
                               DDR4 DIMM
                                                                                                                                                                                                                                                                                              storage            HDD/SDD 8
                                                                                                                                                                                                                                                                             x8
                                                                                                                                                           x8                         x16                                                                                                1st SAS                 2nd SAS
                                                                                                                                                           16 GB/s
16 GB/s
                                                                                                                                                                                                                                                                                                                  NVMe SSD
                                                                                                                                                                                                                                                                             x4                      NVMe
                                                                                                                       PEX                                                                                                              PEX
                                                                                                                                                                                                                                                                                                     slot1
                                                                                                                                                                                                                                                                                                                  NVMe SSD
PCIe Gen3 x8 C5
PCIe Gen3 x8 C6
PCIe Gen3 x8 C7
62   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.1 The IBM POWER9 processor
          This section introduces the latest processor in the IBM Power Systems product family, and
          describes its main characteristics and features in general.
          The servers are offered with various numbers of cores that are activated, and a selection of
          clock frequencies, so IBM can make offerings at several price points, and customers can
          select a particular server (or servers) to fit their budget and performance requirements.
          The POWER9 processor is single-chip modules (SCMs) manufactured on the IBM 14-nm
          FinFET Silicon-On-Insulator (SOI) architecture. Each module is 68.5 mm x 68.5 mm and
          contains 8 billion transistors.
          As shown in Figure 2-4, the chip contains 24 cores, two memory controllers, Peripheral
          Component Interconnect Express (PCIe) Gen4 I/O controllers, and an interconnection system
          that connects all components within the chip at 7 TBps. Each core has 512 KB of L2 cache,
          and 10 MB of L3 embedded DRAM (eDRAM). The interconnect also extends through module
          and system board technology to other POWER9 processors in addition to DDR4 memory and
          various I/O devices.
          POWER9 processor has eight memory channels, and each channel supports up to two DDR4
          DIMM slots. The Power S914 server can support up to 1 TB of memory, and the Power S922
          and Power S924 servers in a two-SCM configuration can support up to 4 TB of memory.
               Note: The total values represent the maximum of 12 cores for the POWER9 architecture.
               Servers that are discussed in this paper have a maximum of 16 or 20 cores per module.
64   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
            Enhanced branch prediction that uses both local and global prediction tables with a
             selector table to choose the best predictor
            Improved out-of-order execution
            Two symmetric fixed-point execution units
            Two symmetric load/store units and two load units, all four of which can also run simple
             fixed-point instructions
            An integrated, multi-pipeline vector-scalar floating point unit for running both scalar and
             SIMD-type instructions, including the Vector Multimedia eXtension (VMX) instruction set
             and the improved Vector Scalar eXtension (VSX) instruction set, which is capable of up to
             16 floating point operations per cycle (eight double precision or 16 single precision)
            In-core AES encryption capability
            Hardware data prefetching with 16 independent data streams and software control
            Hardware decimal floating point (DFP) capability
           More information about Power ISA Version 3.0, see OpenPOWER: IBM Power ISA Version
           3.0B.
           Figure 2-5 shows a picture of the POWER9 core, with some of the functional units
           highlighted.
              Table 2-2 shows a comparison between the different POWER processors in terms of SMT
              capabilities that are supported by each processor architecture.
Table 2-3 shows the processor feature codes (FCs) for the Power S922 server.
              Table 2-3 Processor feature codes specification for the Power S922 server
               Number of cores                  Frequency                        Feature code
              The Power S914 (9009-41A) server is the entry server that supports a one-processor socket
              with up to eight cores.
Table 2-4 shows the processor FCs for the Power S914 server.
              Table 2-4 Processor feature codes specification for the Power S914 server
               Number of cores                  Frequency                        Feature code
66   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
          The Power S924 (9009-42A) server is a powerful 2-socket server with up to 24 cores. A
          system can be ordered with a single processor and a second processor can be added as an
          MES upgrade.
Table 2-5 shows the processor FCs for the Power S924 server.
          Table 2-5 Processor feature codes specification for the Power S924 server
           Number of cores                  Frequency                           Feature code
The speed of the memory depends on the DIMM size and placement.
8 GB 2666 2133
16 GB 2666 2133
32 GB 2400 2133
64 GB 2400 2133
          The Power S914 server can support up to 1 TB of memory. The Power S922 and Power S924
          servers with a two-SCM configuration can operate up to 4 TB of memory.
              The on-chip L3 cache is organized into separate areas with differing latency characteristics.
              Each processor core is associated with a fast 10 MB local region of L3 cache (FLR-L3), but
              also has access to other L3 cache regions as a shared L3 cache. Additionally, each core can
              negotiate to use the FLR-L3 cache that is associated with another core, depending on the
              reference patterns. Data can also be cloned and stored in more than one core’s FLR-L3
              cache, again depending on the reference patterns. This intelligent cache management
              enables the POWER9 processor to optimize the access to L3 cache lines and minimize
              overall cache latencies.
68   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.1.8 Hardware transactional memory
           Transactional memory is an alternative to lock-based synchronization. It attempts to simplify
           parallel programming by grouping read and write operations and running them like a single
           operation. Transactional memory is like database transactions where all shared memory
           accesses and their effects are either committed together or discarded as a group. All threads
           can enter the critical region simultaneously. If there are conflicts in accessing the shared
           memory data, threads try accessing the shared memory data again or are stopped without
           updating the shared memory data. Therefore, transactional memory is also called a lock-free
           synchronization. Transactional memory can be a competitive alternative to lock-based
           synchronization.
           Applications can have customized functions in Field Programmable Gate Arrays (FPGAs) and
           queue work requests directly into shared memory queues to the FPGA. Applications can also
           have customized functions by using the same effective addresses (pointers) they use for any
           threads running on a host processor. From a practical perspective, CAPI enables a
           specialized hardware accelerator to be seen as an extra processor in the system with access
           to the main system memory and coherent communication with other processors in the
           system.
                                               Accelerator
                                                                                     CAPI
                                 Processor
                                                                      Processor         .
Memory Memory
              The benefits of using CAPI include the ability to access shared memory blocks directly from
              the accelerator, perform memory transfers directly between the accelerator and processor
              cache, and reduce the code path length between the adapter and the processors. This
              reduction in the code path length might occur because the adapter is not operating as a
              traditional I/O device, and there is no device driver layer to perform processing. CAPI also
              presents a simpler programming model.
              The accelerator adapter implements the POWER Service Layer (PSL), which provides
              address translation and system memory cache for the accelerator functions. The custom
              processors on the system board, consisting of an FPGA or an ASIC, use this layer to access
              shared memory regions, and cache areas as though they were a processor in the system.
              This ability enhances the performance of the data access for the device and simplifies the
              programming effort to use the device. Instead of treating the hardware accelerator as an I/O
              device, it is treated as a processor, which eliminates the requirement of a device driver to
              perform communication. It also eliminates the need for direct memory access that requires
              system calls to the OS kernel. By removing these layers, the data transfer operation requires
              fewer clock cycles in the processor, improving the I/O performance.
70   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.1.10 Power management and system performance
          The POWER9 scale-out models introduced new features for EnergyScaleincluding new
          variable processor frequency modes that provide a significant performance boost beyond the
          static nominal frequency. The following modes can be modified or disabled.
          The default performance mode depends on the server model. For the Power S914 server
          (9009-41A), Dynamic Performance mode is enabled by default, and the Power S922
          (9009-22A) and Power S924 (9009-42A) servers have Maximum Performance mode enabled
          by default. The difference in the Power S914 setup is that some servers are used in the office
          workspace where the extra fan noise might be unacceptable. If acoustic concern is not an
          issue, you can change to Maximum Performance mode.
          The controls for all of these modes are available on the Advanced System Management
          Interface (ASMI) and can be dynamically modified.
          Table 2-7 Comparison of technology for the POWER9 processor and prior generations
           Characteristics         POWER9                  POWER8                   POWER7+
Technology 14 nm 22 nm 32 nm
Maximum cores 12 12 8
               Note: If you use the 4-core memory #EP10, the maximum of 64 GB RAM is supported.
               The following memory configurations are supported:
                Quantity of 2 or 4 or 6 or 8 of #EM60 (8 GB DDR4 DRAM)
                Quantity of 2 or 4 of #EM62 (16 GB DDR4 DRAM)
                Quantity of 2 of #EM63 (32 GB DDR4 DRAM)
              The Power S922 and Power S924 servers are two-socket servers that support up to two
              POWER9 processor modules. The servers support a maximum of 32 DDR4 DIMM slots, with
              16 DIMM slots per installed processor. Memory features that are supported are 8 GB, 16 GB,
              32 GB, 64 GB, and 128 GB, enabling a maximum system memory of 4 TB. Memory speeds
              vary depending on the DIMM size and modules placement, as shown in Table 2-8.
8 GB 2400 2133
16 GB 2666 2133
32 GB 2400 2133
64 GB 2400 2133
              The maximum theoretical memory bandwidth for POWER9 processor module is 170 GBps.
              The total maximum theoretical memory bandwidth for a two-socket system is 340 GBps.
72   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
          These servers support an optional feature that is called Active Memory Expansion that
          enables the effective maximum memory capacity to be much larger than the true physical
          memory on AIX. This feature runs innovative compression and decompression of memory
          content by using a dedicated coprocessor to provide memory expansion up to 125%,
          depending on the workload type and its memory usage.
          For example, a server with 256 GB RAM physically installed can effectively be expanded over
          512 GB RAM. This approach can enhance virtualization and server consolidation by allowing
          a partition to do more work with the same physical amount of memory or a server to run more
          partitions and do more work with the same physical amount of memory.
              Note: If you use the 4-core memory #EP10, the only memory sizes that are allowed is
              8 GB (#EM60), 16 GB (#EM62) and 32 GB (#EM63).
          All memory must be ordered in pairs, with a minimum of 32 GB for the Power S914,
          Power S924, and Power 922 servers that have a single processor module installed. 64 GB is
          the minimum for servers with two processors modules that are installed (the Power S924 and
          Power S922 servers) per server.
          The supported maximum memory is as follows for the Power S914 server:
           One processor module installed (4-core): 64 GB (eight 8 GB DIMMS, four 16 GB DIMMs,
            or two 32 GB DIMMs)
           One processor module installed (6-core or 8-core): 1 TB (Sixteen 64 GB DIMMs)
          The supported maximum memory is as follows for the Power S924 and the Power S922
          servers:
           One processor module installed: 2 TB (Sixteen 128 GB DIMMs)
           Two processors modules installed: 4 TB (Thirty-two 128 GB DIMMs)
          In general, the preferred approach is to install memory evenly across all processors in the
          system. Balancing memory across the installed processors enables memory access in a
          consistent manner and typically results in the best possible performance for your
          configuration. You should account for any plans for future memory upgrades when you decide
          which memory feature size to use at the time of the initial system order.
MCU Group 1
                                                                                                                                                                Front
               Rear
                                                                                       POWER9
                                                                                         SCM
MCU Group 0
Figure 2-8 Memory DIMM topology for the Power S914 server
              For systems with a single processor module that is installed, the plugging order for the
              memory DIMMs is as follows (see Figure 2-9):
               Pair installation:
                    – The first DIMM pair is installed at Red 1 (C33 DDR0-A and C17 DDR1-A).
                    – The second DIMM pair is installed at Gold 2 (C36 DDR4-A and C22 DDR5-A).
                    – The third DIMM pair is installed at Cyan 3 (C31 DDR2-A and C15 DDR3-A).
                    – The fourth DIMM pair is installed at Gray 4 (C38 DDR6-A and C20 DDR7-A).
               Quad installation:
                    – Two fifth DIMM pairs (or quad) are installed at Red 5 (C34 DDR0-B and C18 DDR1-B)
                      and at Cyan 5 (C32 DDR2-B and C16 DDR3-B).
                    – Two sixth DIMM pairs (or quad) are installed at Gold 6 (C35 DDR4-B and C21 DDR-B)
                      and at Gray 6 (C37 DDR6-B and C19 DDR7-B).
Figure 2-9 DIMM plug sequence for the Power S914 server
74   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
More considerations:
 You may not mix 1R and 2R DIMMs on a single channel within an MCU group because
  they run at different DIMM data rates.
   Table 2-9 lists the feature codes of the supported memory modules.
8 GB #EM60 1R
16 GB #EM62 1R
32 GB #EM63 2R
64 GB #EM64 2R
128 GB #EM65 2R
 DIMMs in the same color cells must be identical (same size and rank).
Note: Each color in Figure 2-9 represents a unique DIMM size and type.
Figure 2-10 shows the physical memory DIMM topology for the Power S922 and the
Power S924 servers.
MCU Group 1
                                  POWER9
                                   SCM1
MCU Group 0
MCU Group 1
                                  POWER9
                                   SCM0
MCU Group 0
Figure 2-10 Memory DIMM topology for the Power S922 and the Power S924
Figure 2-11 DIMM plug sequence for Power S922 and Power S924 servers
              More considerations:
               You may not mix 1R and 2R DIMMs on a single channel within an MCU group because
                they run at different DIMM data rates. For more information, see Table 2-9 on page 75.
               DIMMs at same color cells must be identical (same size and rank).
76   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.2.2 Memory bandwidth
          The POWER9 processor has exceptional cache, memory, and interconnect bandwidths. The
          next sections show the bandwidth capabilities of the Power S914, Power S922, and
          Power S924 servers.
          For an entire Power S922 server that is populated with two processor modules, the overall
          bandwidths are shown in Table 2-11.
              For an entire Power S914 system that is populated with one processor module, the overall
              bandwidths are what is shown in Table 2-13.
              Processor modules for the Power S922 and Power S924 servers run with higher frequency
              than the Power S914 server.
78   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Table 2-14 shows the maximum bandwidth estimates for a single core on the Power S924
server.
Table 2-14 The Power S924 single core bandwidth maximum estimates
 Single core                   Power S924 server                Power S924 server
For an entire Power S924 server that is populated with two processor modules, the overall
bandwidths are shown in Table 2-15.
 Note: There are several POWER9 design points to consider when comparing hardware
 designs that use SMP communication bandwidths as a unique measurement. POWER9
 provides:
    More cores per socket leading to lower inter-CPU communication.
    More RAM density (up to 2 TB per socket) that leads to less inter-CPU communication.
    Greater RAM bandwidth for less dependence on an L3 cache.
    Intelligent hypervisor scheduling that places RAM usage close to the CPU.
    New SMP routing so that multiple channels are available when congestion occurs.
              The Power S914, Power S924, and Power S922 servers have internal I/O connectivity
              through PCIe Gen4 and Gen3 (PCI Express Gen4/Gen3 or PCIe Gen4/Gen3) slots, and also
              external connectivity through SAS adapters.
              The internal I/O subsystem on the Power S914, Power S924, and Power S922 servers is
              connected to the PCIe controllers on a POWER9 processor in the system. An IBM Power
              System server in a two-socket configuration has a bus that has 80 PCIe G4 lanes running at a
              maximum of 16 Gbps full-duplex, which provides 320 GBps of I/O connectivity to the PCIe
              slots, SAS internal adapters, and USB ports. The Power S914 server with one processor
              module provides 160 GBps I/O bandwidth (maximum).
              Some PCIe slots are connected directly to the PCIe Gen4 buses on the processors, and PCIe
              Gen3 devices are connected to these buses through PCIe Gen3 Switches. For more
              information about which slots are connected directly to the processor and which ones are
              attached to a PCIe Gen3 Switch (referred as PEX), see Figure 2-3 on page 62.
Figure 2-12 compares the POWER8 and POWER9 I/O buses architecture.
POWER8 POWER9
                PCIe G3                                          PCIe G4
                96 GBps                       Internal SAS      160 GBps                        Internal SAS
                                PCIe Gen3                                         PCIe Gen3
                                 Switches                                          Switches
                                                 USB                                               USB
              Table 2-16 lists the I/O bandwidth of Power S914, Power S924, and Power S922 processor
              configurations.
80   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
           For PCIe Interconnect, each POWER9 processor module has 40 PCIe lanes running at 16
           Gbps full-duplex. The bandwidth formula is calculated as follows:
           All PCIe slots support enhanced error handling (EEH). PCI EEH-enabled adapters respond to
           a special data packet that is generated from the affected PCIe slot hardware by calling system
           firmware, which examines the affected bus, allows the device driver to reset it, and continues
           without a system restart.
           Table 2-17 PCIe slot locations and descriptions for the Power S922 server
            Slot availability                 Description                         Adapter size
Two slots (P1-C6 and P1-C12) PCIe Gen3 x8 Half-height and half-length
Two slots (P1-C7 and P1-C11) PCIe Gen3 x8 Half-height and half-length
            Three slots (P1-C3a, P1-C4a ,     PCIe Gen4 x16                       Half-height and half-length
            and P1-C9)
            Two slots (P1-C2a and P1-C8)      PCIe Gen4 x8 with x16               Half-height and half-length
                                              connector
              a. The slot is available when the second processor slot is populated.
Table 2-18 lists the PCIe adapter slot locations and details for the Power S922 server.
           Table 2-18 PCIe slot locations and details for the Power S922 server
            Location code       Description          Slot capabilities
              Figure 2-13 shows the rear view of the Power S922 server with the location codes for the
              PCIe adapter slots.
Figure 2-13 Rear view of a rack-mounted Power S922 server with PCIe slots location codes
82   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Table 2-19 provides information about the PCIe slots in the Power S914 and Power S924
servers.
Table 2-19 PCIe slot locations and descriptions for the Power S914 and Power S924 servers
 Slot availability                 Description                       Adapter size
 Two slots (P1-C6 and P1-C12)      PCIe Gen3 x8 with x16             Full-height and half-length
                                   connector
 Three slots (P1-C3a, P1-C4,       PCIe Gen4 x16                     Full-height and half-length
 and P1-C9)
 Two slots (P1-C2 and P1-C8)       PCIe Gen4 x8 with x16             Full-height and half-length
                                   connector
   a. The slot is available when the second processor slot of the Power S924 server is populated.
Table 2-20 lists the PCIe adapter slot locations and details for the Power S914 and
Power S924 servers.
Table 2-20 PCIe slot locations and details for the Power S914 and Power S924 servers
 Location code       Description          Slot capabilities
              Figure 2-14 shows the rear view of the Power S924 server with the location codes for the
              PCIe adapter slots.
Figure 2-14 Rear view of a rack-mounted Power S924 server with the PCIe slots location codes
84   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.5 Peripheral Component Interconnect adapters
          This section covers the various types and functions of the PCI adapters that are supported by
          the Power S914, Power S922, and Power S924 servers.
          The PCIe interfaces that are supported on this server are PCIe Gen4, which are capable of
          16 GBps simplex (32 GBps duplex) on a single x16 interface. PCIe Gen4 slots also support
          previous generations (Gen2 and Gen1) adapters, which operate at lower speeds, according
          to the following rules:
           Place x1, x4, x8, and x16 speed adapters in the same size connector slots first before
            mixing adapter speed with connector slot size.
           Adapters with smaller speeds are allowed in larger sized PCIe connectors, but larger
            speed adapters are not compatible in smaller connector sizes (that is, a x16 adapter
            cannot go in an x8 PCIe slot connector).
          All adapters support EEH. PCIe adapters use a different type of slot than PCI adapters. If you
          attempt to force an adapter into the wrong type of slot, you might damage the adapter or the
          slot.
          IBM POWER9 processor-based servers can support two different form factors of PCIe
          adapters:
           PCIe low-profile (LP) cards, which are used with the Power S922 PCIe slots. These cards
            are not compatible with Power S914 and Power S924 servers because of their low height,
            but there are similar cards in other form factors.
           PCIe full height and full high cards are not compatible with the Power S922 server and are
            designed for the following servers:
             – Power S914 server
             – Power S924 server
          Before adding or rearranging adapters, use the IBM System Planning Tool (SPT) to validate
          the new adapter configuration.
          If you are installing a new feature, ensure that you have the software that is required to
          support the new feature and determine whether there are any existing update prerequisites to
          install. To do this, go to the IBM Power Systems Prerequisite website.
          The following sections describe the supported adapters and provide tables of orderable
          feature numbers. The tables indicate operating system support (AIX, IBM i, and Linux) for
          each of the adapters.
Table 2-21 lists the LAN adapters that are available for the Power S922 server.
 EN0W        2CC4      PCIe2 2-port 10/1 GbE BaseT RJ45 Adapter   0              12          AIX, IBM i, and
                                                                                             Linux
 EN0U        2CC3      PCIe2 4-port (10 Gb+1 GbE) Copper          0              12          AIX, IBM i, and
                       SFP+RJ45 Adapter                                                      Linux
 EN0S        2CC3      PCIe2 4-Port (10 Gb+1 GbE) SR+RJ45         0              12          AIX, IBM i, and
                       Adapter                                                               Linux
 EN0X        2CC4      PCIe2 LP 2-port 10/1 GbE BaseT RJ45        0              9           AIX, IBM i, and
                       Adapter                                                               Linux
 EN0V        2CC3      PCIe2 LP 4-port (10 Gb+1 GbE) Copper       0              9           AIX, IBM i, and
                       SFP+RJ45 Adapter                                                      Linux
 EN0T        2CC3      PCIe2 LP 4-Port (10 Gb+1 GbE) SR+RJ45      0              9           AIX, IBM i, and
                       Adapter                                                               Linux
 EC2S        58FA      PCIe3 2-Port 10 Gb NIC & ROCE SR/Cu        0              4           AIX, IBM i, and
                       Adapter                                                               Linux
 EC37        57BC      PCIe3 LP 2-port 10 GbE NIC&RoCE SFP+       0              9           AIX, IBM i, and
                       Copper Adapter                                                        Linux
 EC38        57BC      PCIe3 2-port 10 GbE NIC&RoCE SFP+          0              12          AIX, IBM i, and
                       Copper Adapter                                                        Linux
 EC2U        58FB      PCIe3 2-Port 25/10 Gb NIC & ROCE SR/Cu     0              4           AIX, IBM i, and
                       Adapter                                                               Linux
 EC3B        57BD      PCIe3 2-Port 40 GbE NIC RoCE QSFP+         0              12          AIX, IBM i, and
                       Adapter                                                               Linux
 EC3B        57BD      PCIe3 2-Port 40 GbE NIC RoCE QSFP+         0              12          AIX, IBM i, and
                       Adapter                                                               Linux
 EN0K        2CC1      PCIe3 4-port (10 Gb FCoE & 1 GbE)          0              12          AIX, IBM i, and
                       SFP+Copper & RJ45                                                     Linux
 EN0H        2B93      PCIe3 4-port (10 Gb Fibre Channel over     0              12          AIX, IBM i, and
                       Ethernet (FCoE) & 1 GbE) SR & RJ45                                    Linux
86      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
 Feature    CCIN     Description                                    Minimum         Maximum       OS support
 code
 EC2R       58FA     PCIe3 LP 2-Port 10 Gb Network Interface Card   0               8             AIX, IBM i, and
                     (NIC) & ROCE SR/Cu Adapter                                                   Linux
 EC37       57BC     PCIe3 LP 2-port 10 GbE NIC&RoCE SFP+           0               9             AIX, IBM i, and
                     Copper Adapter                                                               Linux
 EC3L       2CEC     PCIe3 LP 2-port 100 GbE (NIC & RoCE)           0               3             AIX, IBM i, and
                     QSFP28 Adapter x16                                                           Linux
 EC2T       58FB     PCIe3 LP 2-Port 25/10 Gb NIC & ROCE SR/Cu      0               8             AIX, IBM i, and
                     Adapter                                                                      Linux
 EC3A       57BD     PCIe3 LP 2-Port 40 GbE NIC RoCE QSFP+          0               8             AIX, IBM i, and
                     Adapter                                                                      Linux
 EC3A       57BD     PCIe3 LP 2-Port 40 GbE NIC RoCE QSFP+          0               8             AIX, IBM i, and
                     Adapter                                                                      Linux
 EN0J       2B93     PCIe3 LP 4-port (10 Gb FCoE & 1 GbE) SR &      0               9             AIX, IBM i, and
                     RJ45                                                                         Linux
 EN0L       2CC1     PCIe3 LP 4-port(10 Gb FCoE & 1 GbE)            0               9             AIX, IBM i, and
                     SFP+Copper & RJ45                                                            Linux
 EC67       2CF3     PCIe4 LP 2-port 100 Gb ROCE EN LP adapter      0               3             AIX, IBM i, and
                                                                                                  Linux
Table 2-22 lists the available LAN adapters for a Power S914 server.
 EN0W       2CC4     PCIe2 2-port 10/1G bE BaseT RJ45 Adapter            0              12          AIX, IBM i,
                                                                                                    and Linux
 EN0U       2CC3     PCIe2 4-port (10 Gb+1 GbE) Copper SFP+RJ45          0              12          AIX, IBM i,
                     Adapter                                                                        and Linux
 EN0S       2CC3     PCIe2 4-Port (10 Gb+1 GbE) SR+RJ45 Adapter          0              12          AIX, IBM i,
                                                                                                    and Linux
 EC2S        58FA     PCIe3 2-Port 10 Gb NIC & ROCE SR/Cu Adapter       0         8         AIX, IBM i,
                                                                                            and Linux
 EC38        57BC     PCIe3 2-port 10 GbE NIC&RoCE SFP+ Copper          0         12        AIX, IBM i,
                      Adapter                                                               and Linux
 EC3M        2CEC     PCIe3 2-port 100 GbE (NIC & RoCE) QSFP28          0         1         AIX, IBM i,
                      Adapter x16                                                           and Linux
 EC2U        58FB     PCIe3 2-Port 25/10 Gb NIC & ROCE SR/Cu Adapter    0         8         AIX, IBM i,
                                                                                            and Linux
 EC3B        57BD     PCIe3 2-Port 40 GbE NIC RoCE QSFP+ Adapter        0         12        AIX, IBM i,
                                                                                            and Linux
 EN0K        2CC1     PCIe3 4-port (10 Gb FCoE & 1 GbE) SFP+Copper &    0         12        AIX, IBM i,
                      RJ45                                                                  and Linux
 EN0H        2B93     PCIe3 4-port (10 Gb FCoE & 1 GbE) SR & RJ45       0         12        AIX, IBM i,
                                                                                            and Linux
EC63 2CF1 PCIe4 1-port 100 Gb EDR InfiniBand CAPI adapter 0 1 Linux
EC65 2CF2 PCIe4 2-port 100 Gb EDR InfiniBand CAPI adapter 0 1 Linux
88      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                  Table 2-23 lists the available LAN adapters for a Power S924 server.
 EN0W      2CC4        PCIe2 2-port 10/1 GbE BaseT RJ45 Adapter        0             25           AIX, IBM i, and
                                                                                                  Linux
 EN0U      2CC3        PCIe2 4-port (10 Gb+1 GbE) Copper               0             25           AIX, IBM i, and
                       SFP+RJ45 Adapter                                                           Linux
 EN0S      2CC3        PCIe2 4-Port (10 Gb+1 GbE) SR+RJ45 Adapter      0            25            AIX, IBM i, and
                                                                                                  Linux
 EC2S      58FA        PCIe3 2-Port 10 Gb NIC & ROCE SR/Cu             0             13           AIX, IBM i, and
                       Adapter                                                                    Linux
 EC38      57BC        PCIe3 2-port 10 GbE NIC&RoCE SFP+ Copper        0             25           AIX, IBM i, and
                       Adapter                                                                    Linux
 EC3M      2CEC        PCIe3 2-port 100 GbE (NIC & RoCE) QSFP28        0             3            AIX, IBM i, and
                       Adapter x16                                                                Linux
 EC2U      58FB        PCIe3 2-Port 25/10 Gb NIC & ROCE SR/Cu          0             13           AIX, IBM i, and
                       Adapter                                                                    Linux
 EC3B      57BD        PCIe3 2-Port 40 GbE NIC RoCE QSFP+              0             25           AIX, IBM i, and
                       Adapter                                                                    Linux
 EN0K      2CC1        PCIe3 4-port (10 Gb FCoE & 1 GbE)               0             25           AIX, IBM i, and
                       SFP+Copper & RJ45                                                          Linux
 EN0H      2B93        PCIe3 4-port (10 Gb FCoE & 1 GbE) SR & RJ45     0             25           AIX, IBM i, and
                                                                                                  Linux
                  Table 2-24 on page 90 lists the available graphics accelerator adapter for the Power S922
                  server.
Table 2-25 lists the available graphics accelerator adapter for the Power S914 server.
Table 2-25 The graphics accelerator adapter that is supported in the Power S914 server
 Feature     CCIN       Description                                    Minimum        Maximum       OS support
 code
 5748        5269       POWER GXT145 PCI Express Graphics              0              4             AIX and Linux
                        Accelerator
Table 2-26 lists the available graphics accelerator adapter for the Power S924 server.
Table 2-26 The graphics accelerator card that is supported in the Power S924 server
 Feature     CCIN       Description                                        Minimum        Maximum   OS support
 code
 5748        5269       POWER GXT145 PCI Express Graphics                  0              7         AIX and Linux
                        Accelerator
Table 2-27 The PCIe SAS adapters that are available for the Power S922 server
 Feature     CCIN      Description                                         Minimum        Maximum     OS support
 code
 EJ0L        57CE      PCIe3 12 GB Cache RAID SAS Adapter Quad-port        0              19          AIX, IBM i,
                       6 Gb x8                                                                        and Linux
 EJ14        57B1      PCIe3 12 GB Cache RAID PLUS SAS Adapter             0              8           AIX, IBM i,
                       Quad-port 6 Gb x8                                                              and Linux
90      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                 Table 2-28 lists the SAS adapters that are available for the Power S914 server.
Table 2-28 The PCIe SAS adapters that are available for the Power S914 server
 Feature    CCIN    Description                                           Minimum           Maximum      OS support
 code
 EJ0L       57CE    PCIe3 12 GB Cache RAID SAS Adapter Quad-port 6        0                 19           AIX, IBM i,
                    Gb x8                                                                                and Linux
 EJ14       57B1    PCIe3 12 GB Cache RAID PLUS SAS Adapter               0                 8            AIX, IBM i,
                    Quad-port 6 Gb x8                                                                    and Linux
Table 2-29 lists the SAS adapters that are available for Power S924 servers.
Table 2-29 The PCIe SAS adapters that are available for Power S924 servers
 Feature    CCIN     Description                                          Minimum      Maximum         OS support
 code
 EJ0L       57CE     PCIe3 12 GB Cache RAID SAS Adapter Quad-port         0            19              AIX, IBM i,
                     6 Gb x8                                                                           and Linux
 EJ14       57B1     PCIe3 12 GB Cache RAID PLUS SAS Adapter              0            16              AIX, IBM i,
                     Quad-port 6 Gb x8                                                                 and Linux
                   Note: If you are attaching a device or switch with an SC type fiber connector, then an
                   LC-SC 50-Micron Fiber Converter Cable (#2456) or an LC-SC 62.5-Micron Fiber Converter
                   Cable (#2459) is required.
                 Table 2-30 summarizes the available Fibre Channel adapters for Power S922 servers. They
                 all have LC connectors.
Table 2-30 The PCIe Fibre Channel adapters that are available for Power S922 servers
 Feature    CCIN     Description                                     Minimum           Maximum        OS support
 code
 EN0A       577F     PCIe3 16 Gb 2-port Fibre Channel Adapter         0                12             AIX, IBM i, and
                                                                                                      Linux
 EN0B       577F     PCIe3 LP 16 Gb 2-port Fibre Channel Adapter      0                8              AIX, IBM i, and
                                                                                                      Linux
 EN0F         578D     PCIe2 LP 8 Gb 2-Port Fibre Channel Adapter     0                8             AIX, IBM i and
                                                                                                     Linuxa
 EN0G         578D     PCIe2 8 Gb 2-Port Fibre Channel Adapter        0                12            AIX, IBM i and
                                                                                                     Linuxa
 EN0Y         N/A      PCIe2 LP 8 Gb 4-port Fibre Channel Adapter     0                8             AIX, IBM i, and
                                                                                                     Linux
 EN1A         578F     PCIe3 32 Gb 2-port Fibre Channel Adapter       0                12            AIX, IBM i, and
                                                                                                     Linux
 EN1B         578F     PCIe3 LP 32 Gb 2-port Fibre Channel Adapter    0                8             AIX, IBM i, and
                                                                                                     Linux
 EN1C         578E     PCIe3 16 Gb 4-port Fibre Channel Adapter       0                12            AIX, IBM i, and
                                                                                                     Linux
 EN1D         578E     PCIe3 LP 16 Gb 4-port Fibre Channel Adapter    0                8             AIX, IBM i, and
                                                                                                     Linux
     a. IBM i requires VIOS
                    Table 2-31 summarizes the available Fibre Channel adapters for Power S914 servers. They
                    all have LC connectors.
Table 2-31 The PCIe Fibre Channel adapters that are available for Power S914 servers
 Feature      CCIN     Description                                        Minimum          Maximum   OS support
 code
                    Table 2-32 summarizes the available Fibre Channel adapters for Power S924 servers. They
                    all have LC connectors.
Table 2-32 The PCIe Fibre Channel adapters that are available for Power S924 servers
 Feature      CCIN     Description                                    Minimum          Maximum       OS support
 code
 EN0A         577F     PCIe3 16 Gb 2-port Fibre Channel Adapter       0                25            AIX, IBM i, and
                                                                                                     Linux
92      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
 Feature    CCIN      Description                                       Minimum         Maximum      OS support
 code
 EN0G       578D      PCIe2 8 Gb 2-Port Fibre Channel Adapter           0               25           AIX, IBM i, and
                                                                                                     Linux
 EN1A       578F      PCIe3 32 Gb 2-port Fibre Channel Adapter          0               25           AIX, IBM i, and
                                                                                                     Linux
 EN1C       578E      PCIe3 16 Gb 4-port Fibre Channel Adapter          0               25           AIX, IBM i, and
                                                                                                     Linux
                    Note: The usage of N_Port ID Virtualization (NPIV) through the Virtual I/O Server (VIOS)
                    requires an NPIV-capable Fibre Channel adapter, such as the #EN0A.
                 InfiniBand is an open set of interconnect standards and specifications. The main InfiniBand
                 specification is published by the IBTA and is available at the IBTA website.
                 InfiniBand is based on a switched fabric architecture of serial point-to-point links, where these
                 InfiniBand links can be connected to either host channel adapters (HCAs), which are used
                 primarily in servers, or target channel adapters (TCAs), which are used primarily in storage
                 subsystems.
                 The InfiniBand physical connection consists of multiple byte lanes. Each individual byte lane
                 is a four-wire, 2.5, 5.0, or 10.0 Gbps bidirectional connection. Combinations of link width and
                 byte lane speed allow for overall link speeds of 2.5 - 120 Gbps. The architecture defines a
                 layered hardware protocol and also a software layer to manage initialization and the
                 communication between devices. Each link can support multiple transport services for
                 reliability and multiple prioritized virtual communication channels.
                 For more information about InfiniBand, see HPC Clusters Using InfiniBand on IBM Power
                 Systems Servers, SG24-7767.
Table 2-33 lists the InfiniBand adapters that are available for Power S922 servers.
Table 2-34 lists the InfiniBand adapters that are available for Power S914 servers.
EC63 2CF1 PCIe4 1-port 100 Gb EDR InfiniBand CAPI adapter 0 1 Linux
EC65 2CF2 PCIe4 2-port 100 Gb EDR InfiniBand CAPI adapter 0 1 Linux
Table 2-35 lists the InfiniBand adapters available for Power S924 servers.
EC63 2CF1 PCIe4 1-port 100 Gb EDR InfiniBand CAPI adapter 0 3 Linux
EC65 2CF2 PCIe4 2-port 100 Gb EDR InfiniBand CAPI adapter 0 3 Linux
Table 2-36 The cryptographic coprocessor available for Power S9222 servers
 Feature    CCIN      Description                                       Minimum           Maximum       OS support
 code
 EJ33       4767      PCIe3 Crypto Coprocessor BSC-Gen3 4767            0                 12            AIX, IBM i, and
                                                                                                        Linux
                   The cryptographic coprocessor cards that are supported for the Power S914 server are
                   shown in Table 2-37.
94      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
 Feature    CCIN     Description                                          Minimum       Maximum        OS support
 code
                The cryptographic coprocessor cards that are supported for the Power S924 server are
                shown in Table 2-38.
 EJ32       4767     PCIe3 Crypto Coprocessor no BSC 4767             0                 10            AIX, IBM i, and
                                                                                                      Linux
 EJ33       4767     PCIe3 Crypto Coprocessor BSC-Gen3 4767           0                 18            AIX, IBM i, and
                                                                                                      Linux
 EC3L       2CEC     PCIe3 LP 2-port 100 GbE (NIC & RoCE)             0                 3             AIX, IBM i, and
                     QSFP28 Adapter x16                                                               Linux
 EC62       2CF1     PCIe4 LP 1-port 100 Gb EDR InfiniBand CAPI       0                 3             AIX, IBM i, and
                     adapter                                                                          Linux
 EC64       2CF2     PCIe4 LP 2-port 100 Gb EDR InfiniBand CAPI       0                 3             AIX, IBM i, and
                     adapter                                                                          Linux
                The CAPI-capable adapters that are available for Power S914 servers are shown in
                Table 2-40.
 EC3M       2CEC    PCIe3 2-port 100 GbE (NIC & RoCE) QSFP28 Adapter          0             1           AIX, IBM i,
                    x16                                                                                 and Linux
 EC63       2CF1    PCIe4 1-port 100 Gb EDR InfiniBand CAPI adapter           0             1           AIX, IBM i,
                                                                                                        and Linux
 EC65       2CF2    PCIe4 2-port 100 Gb EDR InfiniBand CAPI adapter           0             1           AIX, IBM i,
                                                                                                        and Linux
                The CAPI-capable adapters that are available for Power S924 servers are shown in
                Table 2-41 on page 96.
 EC3M        2CEC       PCIe3 2-port 100 GbE (NIC & RoCE) QSFP28          0         3              AIX, IBM i,
                        Adapter x16                                                                and Linux
 EC63        2CF1       PCIe4 1-port 100 Gb EDR InfiniBand CAPI adapter   0         3              AIX, IBM i,
                                                                                                   and Linux
 EC65        2CF2       PCIe4 2-port 100 Gb EDR InfiniBand CAPI adapter   0         3              AIX, IBM i,
                                                                                                   and Linux
 EC45        58F9       PCIe2 LP 4-Port USB 3.0 Adapter               0             8             AIX, IBM i, and
                                                                                                  Linux
 EC46        58F9       PCIe2 4-Port USB 3.0 Adapter                  0             12            AIX, IBM i, and
                                                                                                  Linux
                    The USB adapters that are available for Power S914 servers are shown in Table 2-40 on
                    page 95.
The USB adapters that are available for Power S924 servers are shown in Table 2-41.
96      IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.6 Internal storage
          The internal storage on the Power S922, S914, and Power S924 servers depends on the
          DASD/Media backplane that is used. The servers support various DASD/Media backplanes:
           Note: If #EC59 is ordered, a minimum of one #ES14 Mainstream 400 GB solid-state drive
           (SSD) NVMe M.2 module must be ordered.
           IBM i operating system performance: Clients with write-sensitive disk/hard disk drive
           (HDD) workloads should upgrade from the base storage backplane (#EJ1C/#EJ1E) to the
           expanded function storage backplanes (#EJ1M/#EJ1D) to gain the performance
           advantage of write cache
          The high-performance SAS controller provides RAID 0, RAID 5, RAID 6, or RAID 10 support
          for either HDDs or SSDs. JBOD support for HDD is also supported. The controller has no
          write cache.
          For servers that support split backplane capability, add #EJ1E. For write cache performance,
          use #EJ1D or #EJ1M instead of this backplane.
          Both 5xx and 4-KB sector HDDs/SSDs are supported. 5xx and 4-KB drives cannot be mixed
          in the same array.
          This FC provides a storage backplane with one integrated SAS adapter with no cache,
          running 12 SFF-3 SAS bays in the system unit and one RDX bay in the system unit.
The internal connections to the physical disks are shown in Figure 2-15 on page 98.
               The high-performance SAS controllers each provides RAID 0, RAID 5, RAID 6, or RAID 10
               support. JBOD support for HDDs is also supported. There is no write cache on either
               controller.
               Both 5xx and 4-KB sector HDDs/SSDs are supported. 5xx and 4-KB drives cannot be mixed
               in the same array.
               This FC provides a second integrated SAS adapter with no cache and internal cables to
               provide two sets of six SFF-3 bays in the system unit.
98    IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                 IBM i
The internal connections to the physical disks are shown in Figure 2-16 on page 99.
                The high-performance SAS controllers provide RAID 0, RAID 5, RAID 6, RAID 10, RAID 5T2,
                RAID 6T2, or RAID 10T2 support. Patented Active/Active configurations with at least two
                arrays are supported.
                The Easy Tier function is supported so that the dual controllers can automatically move hot
                data to attached SSDs and cold data to attached HDDs for AIX and Linux, and VIOS
                environments.
                SFF or 2.5-inch drives are mounted on a carrier/ tray that is specific to the system unit
                (SFF-3). The backplane has 18 SFF-3 bays.
                This backplane enables two SAS ports (#EJ0W) at the rear of the system unit to support the
                attachment of one EXP24S/EXP24SX I/O drawer in mode1 to hold HDDs or SSDs.
              This backplane does not support a split backplane. For a split backplane, use #EJ1C plus
              #EJ1E.
              Both 5xx and 4-KB sector HDDs/SSDs are supported. 5xx and 4-KB drives cannot be mixed
              in the same array.
              This FC provides a storage backplane with a pair of integrated SAS adapters with write
              cache, with an optional external SAS port running up to:
               A set of 18 SFF-3 SAS bays in the system unit
               Two SAS ports at the rear of the system unit to connect to a single EXP24S/EXP24SX I/O
                drawer
              The high-performance SAS controllers provide RAID 0, RAID 5, RAID 6, RAID 10, RAID 5T2,
              RAID 6T2, or RAID 10T2 support. Patented Active/Active configurations with at least two
              arrays are supported.
              The Easy Tier function is supported so that the dual controllers can automatically move hot
              data to attached SSDs and cold data to attached HDDs for AIX and Linux, and VIOS
              environments.
              SFF or 2.5-inch drives are mounted on a carrier/ tray that i specific to the system unit (SFF-3).
              The backplane has 12 SFF-3 bays.
              This backplane also enables two SAS ports (#EJ0W) at the rear of the system unit, which
              support the attachment of one EXP24S/EXP24SX I/O drawer in mode1, which holds HDDs or
              SSDs.
#EJ0W is an optional feature with #EJ1M, and one 8 PCIe slot is used by #EJ0W.
              This backplane does not support a split backplane. For a split backplane, use the #EJ1C with
              #EJ1E backplane features.
              Both 5xx and 4-KB sector HDDs/SSDs are supported. 5xx and 4-KB drives cannot be mixed
              in the same array.
100   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                This FC provides an expanded function storage backplane with a pair of integrated SAS
                adapters with a write cache, and an optional external SAS port running up to:
                 A set of 12 SFF-3 SAS bays in the system unit
                 One RDX bay in the system unit
                 Two SAS ports at the rear of the system unit to connect to a single EXP24S/EXP24SX I/O
                  drawer
                NVMe adapters
                Table 2-45 provides a list of available NVMe cards for the S922 server.
EC5G 58FC PCIe3 LP 1.6 TB SSD NVMe Adapter 0 5 AIX and Linux
EC5C 58FD PCIe3 LP 3.2 TB SSD NVMe adapter 0 5 AIX and Linux
EC5E 58FE PCIe3 LP 6.4 TB SSD NVMe adapter 0 5 AIX and Linux
Table 2-46 provides a list of available NVMe cards for the S914 server.
EC5B 58FC PCIe3 LP 1.6 TB SSD NVMe Adapter 0 3 AIX and Linux
EC5D 58FD PCIe3 LP 3.2 TB SSD NVMe adapter 0 3 AIX and Linux
EC5F 58FE PCIe3 LP 6.4 TB SSD NVMe adapter 0 3 AIX and Linux
EC6V 58FC PCIe3 x8 1.6 TB NVMe Flash Adapter for IBM i 0 3 IBM i
EC6X 58FD PCIe3 x8 3.2 TB NVMe Flash Adapter for IBM i 0 3 IBM i
EC6Z 58FE PCIe3 x8 6.4 TB NVMe Flash Adapter for IBM i 0 3 IBM i
Table 2-47 provides a list of available NVMe cards for the S924 server.
EC5B 58FC PCIe3 LP 1.6 TB SSD NVMe Adapter 0 5 AIX and Linux
EC5D 58FD PCIe3 LP 3.2 TB SSD NVMe adapter 0 5 AIX and Linux
EC5F 58FE PCIe3 LP 6.4 TB SSD NVMe adapter 0 5 AIX and Linux
EC6V 58FC PCIe3 x8 1.6 TB NVMe Flash Adapter for IBM i 0 3 IBM i
EC6X 58FD PCIe3 x8 3.2 TB NVMe Flash Adapter for IBM i 0 3 IBM i
EC6Z 58FE PCIe3 x8 6.4 TB NVMe Flash Adapter for IBM i 0 3 IBM i
                PCIe3 NVMe carrier card with two M.2 module slots (#EC59)
                The NVMe option offers fast start times, and is ideally suited to housing the rootvg of VIOS
                partitions.
                #EC59 is a carrier card for 400 GB Mainstream SSD NVMe module (#ES14). The maximum
                quantity is two of #ES14 per #EC59.
This FC provides an PCIe3 NVMe card with two M.2 module slots.
                If the NVMe carrier card #EC59 is selected, you do not have to order disk units. If you do not
                order SAN Boot (#EC59 or #0837), then you must order at least one disk unit. If you do not
                order HDD/SSD/SAN boot (#0837), then #EC59 (with at least one of #ES14) is the load
                source.
Figure 2-17 shows the location of the #EC59 feature code in a Power S922 server.
102     IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Figure 2-18 on page 103 shows an #EC59 feature code.
Figure 2-18 #EC59 feature code with the cover open showing two #ES14 modules fitted
Each NVMe device (#ES14) is a separate PCIe endpoint, which means that each NVMe
device can be assigned to a different logical partition (LPAR) or Virtual I/O Server (VIOS). At
the operating system level, each #ES14 device appears to the operating system as an
individual disk. For example, in AIX, the device might appear as hdisk0.
 Tip: If two #EC59 feature codes are configured, each with two #ES14 feature codes, it is
 possible to have the rootvg of the first VIOS mirrored to an #ES14 in each #EC59, and the
 second VIOS could be mirrored to the other two modules, which provides excellent
 performance and resilience.
Or, each #ES14 could be assigned to a separate partition or VIOS server as a boot device.
 Note: You must order, at a minimum, one #ES14 module with each #EC59 that you order.
 The maximum quantity is two #ES14 per #EC59.
Using this FC for other functions beyond boot support and non-intensive workloads might
result in throttled performance and high temperatures that lead to timeouts and critical
thermal warnings.
              This backplane option uses leading-edge, integrated SAS RAID controller technology that is
              designed and patented by IBM. A custom-designed PowerPC based ASIC chip is the basis of
              these SAS RAID controllers, and provides RAID 0, RAID 5, RAID 6, and RAID 10 functions
              with HDDs and SSDs. Internally, SAS ports are implemented and provide plentiful bandwidth.
              The integrated SAS controllers are placed in dedicated slots and do not reduce the number of
              available PCIe slots.
              The Storage Backplane option (#EJ1F) provides eight SFF-3 bays and one SAS controller
              with zero write cache.
              Optionally, by adding the Split Backplane (#EJ1H), a second integrated SAS controller with no
              write cache is provided, and the eight SSF-3 bays are logically divided into two sets of four
              bays. Each SAS controller independently runs one of the four-bay sets of drives.
              This backplane option supports HDDs or SSDs or a mixture of HDDs and SSDs in the SFF-3
              bays. Mixing HDDs and SSDs applies even within a single set of four bays of the split
              backplane option. If you are mixing HDDs and SSDs, they must be in separate arrays (unless
              you use the Easy Tier function).
              This backplane option can offer different drive protection options: RAID 0, RAID 5, RAID 6, or
              RAID 10. RAID 5 requires a minimum of three drives of the same capacity. RAID 6 requires a
              minimum of four drives of the same capacity. RAID 10 requires a minimum of two drives.
              Hot-spare capability is supported by RAID 5, RAID 6, or RAID 10.
104   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
          RAID 5 and RAID 6 result in more drive write activity than mirroring or unprotected drives.
          This backplane option is supported by AIX and Linux, and VIOS. As a preferred practice, the
          drives should be protected.
          Drive protection
          HDD/SSD drive protection can be provided by AIX, IBM i, and Linux, or by the HDD/SSD
          hardware controllers.
          Apart from the #EC59 option, all of the storage backplanes offer RAID. The default storage
          backplanes (#EJ1C for the Power S914 and Power S924 servers, and #EJ1F and #EJ1G for
          the Power S922 server) contain one SAS HDD/SSD controller and provide support for JBOD
          and RAID 0, 5, 6, and 10 for AIX or Linux. A secondary non-redundant controller is added
          when you use #EJ1E for the Power S914 and Power S924 servers or #EJ1H for the Power
          S922 server, so each of the six disk bays has a separated disk controller.
          When you choose the optional #EJ1D, #EJ1M, or #EJ1H storage backplane, the controller is
          replaced by a pair of high-performance RAID controllers with dual integrated SAS controllers
          with 1.8 GB of physical write cache. High-performance controllers run 18 SFF-3 SAS bays
          with 1.8-inch SSD bays. Dual controllers (also called dual I/O adapters or paired controllers)
          and their write cache are placed in integrated slots and do not use PCIe slots. Patented
          active/active configurations with at least two arrays are supported.
          The write cache, which is responsible for increasing write performance by caching data
          before it is written to the physical disks, can have its data compression capabilities activated,
          which provides up to 7.2 GB effective cache capacity. The write cache contents are protected
          against power loss by flash memory and super capacitors, which removes the need for
          battery maintenance.
          The high-performance SAS controllers provide RAID 0, RAID 5, RAID 6, and RAID 10
          support, and the Easy Tier variants (RAID 5T2, RAID 6T2, and RAID 10T2) if the server has
          both HDDs and SSDs installed.
          The Easy Tier function is supported, so the dual controllers can automatically move hot data
          to an attached SSD and cold data to an attached HDD for AIX and Linux, and VIOS
          environments. To learn more about Easy Tier, see 2.6.9, “Easy Tier” on page 106.
          AIX and Linux can use disk drives that are formatted with 512-byte blocks when they are
          mirrored by the operating system. These disk drives must be reformatted to 528-byte sectors
          when used in RAID arrays. Although a small percentage of the drive's capacity is lost, extra
              RAID 5T2, RAID 6T2, and RAID 10T2 are RAID levels with Easy Tier enabled. They require
              that both types of disks exist on the system under the same controller (HDDs and SSDs), and
              that both types are configured under the same RAID type.
              The high-function backplane (#EJ1D) can handle both types of storage in two different ways:
               Separate arrays: SSDs and HDDs coexist on separate arrays, just like the Standard SAS
                adapter can.
               Easy Tier: SSDs and HDDs coexist under the same array.
              When the SDDs and HDDS are under the same array, the adapter can automatically move
              the most accessed data to faster storage (SSDs) and less accessed data to slower storage
              (HDDs). This is called Easy Tier.
              There is no need for coding or software intervention after the RAID is configured correctly.
              Statistics on block accesses are gathered every minute, and after the adapter realizes that
              some portion of the data is being frequently requested, it moves this data to faster devices.
              The data is moved in chunks of 1 MB or 2 MB called bands.
106   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
From the operating system point-of-view, there is just a regular array disk. From the SAS
controller point-of-view, there are two arrays with parts of the data being serviced by one tier
of disks and parts by another tier of disks.
Figure 2-22 Tiered arrays (RAID 5T2, RAID 6T2, and RAID 10T2) example on AIX RAID Manager
              To support Easy Tier, make sure that the server is running at least the following minimum
              levels:
               VIOS 2.2.3.3 with interim fix IV56366 or later
               AIX 7.1 TL3 SP3 or later
108   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
            AIX 6.1 TL9 SP3 or later
            RHEL 6.5 or later
            SUSE Linux Enterprise Server 11 SP3 or later
           Power S914 and Power S924 servers that use the high-performance RAID feature support
           two external SAS ports. The external SAS ports are used for expansion to an external SAS
           drawer.
           More drawers and the IBM System Storage 7226 Tape and DVD Enclosure Express (Model
           1U3) can be attached by installing more SAS adapters.
            Note: Only one SAS drawer is supported from the external SAS port. More SAS drawers
            can be supported through SAS adapters.
              If you feel that you do need a DVD drive, IBM offers a stand-alone external USB unit (#EUA5),
              which is shown in Figure 2-23.
               Note: If you use an external/stand-alone USB drive, which does not have its own power
               supply, you should use a USB socket at the front of the system to ensure enough current is
               available.
If #EJ1C or #EJ1M is configured, an internal bay is available and can be populated by #EU00.
110   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.7 External IO subsystems
          This section describes the PCIe Gen3 I/O expansion drawer that can be attached to the
          Power S922, Power S914, and Power S924 servers.
          The PCIe fan-out module has two CXP ports, which are connected to two CXP ports on a
          PCIe Optical Cable Adapter (#EJ05, #EJ07, or #EJ08, depending on the server that is
          selected). A pair of active optical CXP cables (AOCs) or a pair of CXP copper cables are used
          for this connection.
          A BSC is used to house the full-high adapters that go into these slots. The BSC is the same
          BSC that is used with the previous generation server's #5802/5803/5877/5873 12X attached
          I/O drawers.
Figure 2-24 shows the back view of the PCIe Gen3 I/O expansion drawer.
Figure 2-24 Rear view of the PCIe Gen3 I/O expansion drawer
               Cable lengths: Use the 3.0 m cables for intra-rack installations. Use the 10.0 m cables for
               inter-rack installations.
               Limitation: You cannot mix copper and optical cables on the same PCIe Gen3 I/O drawer.
               Both fan-out modules use copper cables or both use optical cables.
              A minimum of one PCIe3 Optical Cable Adapter for PCIe3 Expansion Drawer is required to
              connect to the PCIe3 6-slot fan-out module in the I/O expansion drawer. The fan-out module
              has two CXP ports. The top CXP port of the fan-out module is cabled to the top CXP port of
              the PCIe3 Optical Cable Adapter. The bottom CXP port of the fan-out module is cabled to the
              bottom CXP port of the same PCIe3 Optical Cable Adapter.
112   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Figure 2-25 shows the connector locations for the PCIe Gen3 I/O expansion drawer.
Figure 2-25 Connector locations for the PCIe Gen3 I/O expansion drawer
              The PCIe3 cable adapter for the PCIe3 EMX0 expansion drawer (#EJ08) is supported in
              P1-C9 for the Power S914 and Power S924 single processor systems. It is supported in
              P1-C9, P1-C3, and P1-C4 in the Power S924 double processor systems.
              Table 2-49 shows PCIe adapter slot priorities and maximum adapters that are supported in
              the Power S922, Power S914, and Power S924 systems.
              Table 2-49 PCIe adapter slot priorities and maximum adapters that are supported
               System                   Feature code            Slot priorities          Maximum number of
                                                                                         adapters supported
2.7.3 PCIe Gen3 I/O expansion drawer system power control network cabling
              There is no system power control network (SPCN) that is used to control and monitor the
              status of power and cooling within the I/O drawer. SPCN capabilities are integrated into the
              optical cables.
2.8.1 EXP24SX SAS Storage Enclosure and EXP12SX SAS Storage Enclosure
              The EXP24SX is a storage expansion enclosure with 24 2.5-inch SFF SAS bays. It supports
              up to 24 hot-swap HDDs or SSDs in only 2 EIA of space in a 19-inch rack. The EXP24SX SFF
              bays use SFF gen2 (SFF-2) carriers/ trays that are identical to the carrier/trays in the previous
              EXP24S drawer. With AIX and Linux, or VIOS, the EXP24SX can be ordered with four sets of
              six bays (mode 4), two sets of 12 bays (mode 2), or one set of 24 bays (mode 1). With IBM i,
              one set of 24 bays (mode 1) is supported.
114   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
There can be no mixing of HDDs and SSDs in the same mode 1 drawer. HDDs and SSDs can
be mixed in a mode 2 or mode 4 drawer, but they cannot be mixed within a logical split of the
drawer. For example, in a mode 2 drawer with two sets of 12 bays, one set can hold SSDs
and one set can hold HDDs, but you cannot mix SSDs and HDDs in the same set of 12 bays.
The EXP12SX is a storage expansion enclosure with twelve 3.5-inch large form factor (LFF)
SAS bays. It supports up to 12 hot-swap HDDs in only 2 EIA of space in a 19-inch rack. The
EXP12SX SFF bays use LFF gen1 (LFF-1) carriers/trays. 4-KB sector drives (4096 or 4224)
are supported. With AIX and Linux, and VIOS, the EXP12SX can be ordered with four sets of
three bays (mode 4), two sets of six bays (mode 2), or one set of 12 bays (mode 1). Only
4-KB sector drives are supported in the EXP12SX drawer.
Four mini-SAS HD ports on the EXP24SX or EXP12SX are attached to PCIe Gen3 SAS
adapters or attached to an integrated SAS controller in the Power S922, S914, or S924
systems. The following PCIe3 SAS adapters support the EXP24SX and EXP 12SX:
 PCIe3 RAID SAS Adapter Quad-port 6 Gb x8 (#EJ0J, #EJ0M, #EL3B, or #EL59)
 PCIe3 12 GB Cache RAID Plus SAS Adapter Quad-port 6 Gb x8 (#EJ14)
Earlier generation PCIe2 or PCIe1 SAS adapters are not supported by the EXP24SX.
The attachment between the EXP24SX or EXP12SX and the PCIe3 SAS adapters or
integrated SAS controllers is through SAS YO12 or X12 cables. All ends of the YO12 and X12
cables have mini-SAS HD narrow connectors. The cable options are:
   X12 cable: 3-meter copper (#ECDJ)
   YO12 cables: 1.5-meter copper (#ECDT) or 3-meter copper (#ECDU)
   3M 100 GbE Optical Cable QSFP28 (AOC) (#EB5R)
   5M 100 GbE Optical Cable QSFP28 (AOC) (#EB5S)
   10M 100 GbE Optical Cable QSFP28 (AOC) (#EB5T)
   15M 100 GbE Optical Cable QSFP28 (AOC) (#EB5U)
   20M 100 GbE Optical Cable QSFP28 (AOC) (#EB5V)
   30M 100 GbE Optical Cable QSFP28 (AOC) (#EB5W)
   50M 100 GbE Optical Cable QSFP28 (AOC) (#EB5X)
   100M 100 GbE Optical Cable QSFP28 (AOC) (#EB5Y)
There are six SAS connectors at the rear of the EXP24SX and EXP12SX to which SAS
adapters or controllers are attached. They are labeled T1, T2, and T3; there are two T1, two
T2, and two T3 connectors.
 In mode 1, two or four of the six ports are used. Two T2 ports are used for a single SAS
  adapter, and two T2 and two T3 ports are used with a paired set of two adapters or a dual
  adapters configuration.
 In mode 2 or mode 4, four ports are used, two T2s and two T3s, to access all SAS bays.
Figure 2-27 Connector locations for the EXP24SX and EXP12SX storage enclosures
              For more information about SAS cabling, see the “Connecting an ESLL or ESLS storage
              enclosure to your system” topic in IBM Knowledge Center.
              The EXP24SX and EXP12SX drawers have many high-reliability design points:
               SAS bays that support hot-swap.
               Redundant and hot-plug power and fan assemblies.
               Dual power cords.
               Redundant and hot-plug ESMs.
               Redundant data paths to all drives.
               LED indicators on drives, bays, ESMs, and power supplies that support problem
                identification.
               Through the SAS adapters/controllers, drives that can be protected with RAID and
                mirroring and hot-spare capability.
116   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
        IBM Flash Storage
        The next generation of IBM Flash Storage delivers the extreme performance and efficiency
        you need to succeed, with a new pay-as-you-go option to reduce your costs and
        scale-on-demand. For more information, see Flash Storage and All Flash Arrays.
        In addition, the VIOS can be installed in special partitions that provide support to other
        partitions running AIX, IBM i, or Linux operating systems for using features such as virtualized
        I/O devices, PowerVM Live Partition Mobility (LPM), or PowerVM Active Memory Sharing.
              IBM periodically releases maintenance packages (service packs or technology levels) for the
              AIX operating system. For more information about these packages, downloading, and
              obtaining the CD-ROM, see Fix Central.
              The Fix Central website also provides information about how to obtain the fixes that are
              included on CD-ROM.
              The Service Update Management Assistant (SUMA), which can help you automate the task
              of checking and downloading operating system downloads, is part of the base operating
              system. For more information about the suma command, see IBM Knowledge Center.
              The following minimum levels of AIX support the Power S922, Power S914, and Power S924
              servers:
               If you are installing the AIX operating system LPAR with any I/O configuration:
                 – AIX Version 7.2 with the 7200-02 Technology Level and Service Pack
                   7200-02-02-1810, or later
                 – AIX Version 7.1 with the 7100-05 Technology Level and Service Pack
                   7100-05-02-1810, or later
                 – AIX Version 6.1 with the 6100-09 Technology Level and Service Pack
                   6100-09-11-1810, or later
                 – AIX Version 7.2 with the 7200-01 Technology Level and Service Pack
                   7200-01-04-1806, or later
                 – AIX Version 7.2 with the 7200-00 Technology Level and Service Pack
                   7200-00-06-1806, or later
                 – AIX Version 7.1 with the 7100-04 Technology Level and Service pack
                   7100-04-06-1806, or later
               If you are installing the AIX operating system Virtual I/O only LPAR:
                 – AIX Version 7.2 with the 7200-02 Technology Level and Service Pack
                   7200-02-01-1732, or later
                 – AIX Version 7.2 with the 7200-01 Technology Level and Service Pack
                   7200-01-01-1642, or later
                 – AIX Version 7.2 with the 7200-00 Technology Level and Service Pack
                   7200-00-01-1543, or later
                 – AIX Version 7.1 with the 7100-05 Technology Level and Service Pack
                   7100-05-01-1731, or later
                 – AIX Version 7.1 with the 7100-04 Technology Level and Service Pack
                   7100-04-01-1543, or later
                 – AIX Version 6.1 with the 6100-09 Technology Level and Service Pack
                   6100-09-06-1543, or later (AIX 6.1 service extension required.)
118   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.9.2 IBM i
              IBM i is supported on the Power S922, Power S914, and Power S924 servers with the
              following minimum required levels:
               IBM i 7.3 TR4
               IBM i 7.2 TR8
              IBM periodically releases maintenance packages (service packs or technology levels) for the
              IBM i operating system. For more information about these packages, downloading, and
              obtaining the CD-ROM, see IBM Fix Central.
              For compatibility information for hardware features and the corresponding AIX and IBM i
              Technology Levels, see IBM Prerequisites.
              The supported versions of Linux on the Power S922, Power S914, and Power S924 servers
              are as follows:
               If you are installing the Linux operating system LPAR:
                 – Red Hat Enterprise Linux 7 for Power LE Version 7.4, or later (POWER8 mode).
                 – SUSE Linux Enterprise Server 12 Service Pack 3, or later.
                 – Ubuntu Server 16.04.4, or later (POWER8 mode).
               If you are installing the Linux operating systems LPAR in non-production SAP
                implementations:
                 – SUSE Linux Enterprise Server 12 Service Pack 3, or later.
                 – SUSE Linux Enterprise Server for SAP with SUSE Linux Enterprise Server 12 Service
                   Pack 3, or later.
                 – Red Hat Enterprise Linux 7 for Power LE, version 7.4, or later (POWER8 mode).
                 – Red Hat Enterprise Linux for SAP with Red Hat Enterprise Linux 7 for Power LE
                   version 7.4, or later (POWER8 mode). Linux supports almost all of the Power Systems
                   I/O, and the configurator verifies support on order.
              To learn about developing on the IBM Power Architecture®, find packages, get access to
              cloud resources, and discover tools and technologies, see Linux on IBM Power Systems
              Developer Portal.
              The IBM Advance Toolchain for Linux on Power is a set of open source compilers, runtime
              libraries, and development tools that users can use to take leading-edge advantage of IBM
              POWER hardware features on Linux. For more information, see Advanced Toolchain for Linux
              on Power.
For more information about Red Hat Enterprise Linux, see Red Hat Enterprise Linux.
                  IBM regularly updates the VIOS code. To find information about the latest updates, see
                  Fix Central.
                  Table 2-50 provides a list of the Power Systems RAS capabilities by operating system. The
                  HMC is an optional feature on scale-out Power Systems servers.
Table 2-50 Selected reliability, availability, and serviceability features by operating system
 RAS feature                                                             AIX              IBM i       Linux
Processor
I/O subsystem
Memory availability
120     IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
RAS feature                                                          AIX              IBM i            Linux
Serviceability
Inventory collection X X X
EED collection X X X
Redundant HMCs X X X
EPOW
              Various service interfaces are used, depending on the state of the system and its operating
              environment. Here are the primary service interfaces:
               Light Path, which provides indicator lights to help a service technical find a component in
                need of service.
               Service processor.
               ASMI.
               Operator panel.
               An operating system service menu, which obtains error codes directly from the hardware.
               Service Focal Point (SFP) on the HMC.
              Service processor
              The service processor is a controller that is running its own operating system. It is a
              component of the service interface card.
              The service processor operating system has specific programs and device drivers for the
              service processor hardware. The host interface is a processor support interface that is
              connected to the POWER processor. The service processor is always working, regardless of
              the main system unit’s state. The system unit can be in the following states:
               Standby (power off)
               Operating, ready to start partitions
               Operating with running logical partitions (LPARs)
              The service processor is used to monitor and manage the system hardware resources and
              devices. The service processor checks the system for errors, ensuring that the connection to
              the management console for manageability purposes is functioning, and accepting ASMI
              Secure Sockets Layer (SSL) network connections. The service processor can view and
              manage the machine-wide settings by using the ASMI, which enables complete system and
              partition management from the HMC.
               Analyzing a system that does not start: The flexible service processor (FSP) can
               analyze a system that does not start. Reference codes and detailed data are available in
               the ASMI and are transferred to the HMC.
122   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
The service processor uses two Ethernet ports that run at 1-Gbps speed. Consider the
following information:
 Both Ethernet ports are visible only to the service processor and can be used to attach the
  server to an HMC or to access the ASMI. The ASMI options can be accessed through an
  HTTP server that is integrated into the service processor operating environment.
 Both Ethernet ports support only auto-negotiation. Customer-selectable media speed and
  duplex settings are not available.
 Both Ethernet ports have a default IP address, as follows:
    – Service processor eth0 (HMC1 port) is configured as 169.254.2.147.
    – Service processor eth1 (HMC2 port) is configured as 169.254.3.147.
    – DHCP using the HMC for the HMC management networks is also possible.
The ASMI is accessible through the management console. It is also accessible by using a
web browser on a system that is connected directly to the service processor (in this case,
either a standard Ethernet cable or a crossed cable) or through an Ethernet network. ASMI
can also be accessed from an ASCII terminal, but this is available only while the system is in
the platform powered-off mode.
Figure 2-28 Starting the Advanced System Management Interface through the HMC GUI
124   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
You are prompted for confirmation about which FSP to use, and then a login window opens,
as shown in Figure 2-29.
               Tip: If you click Expand all menus, as shown in the red ring, you can then use the search
               function (Ctrl+f) in your browser to find quickly menu items.
              Use the ASMI to change the service processor IP addresses or to apply certain security
              policies and prevent access from unwanted IP addresses or ranges.
              You might be able to use the service processor’s default settings. In that case, accessing the
              ASMI is not necessary. To access ASMI, use one of the following methods:
               Use a management console.
                 If configured to do so, the management console connects directly to the ASMI for a
                 selected system from this task.
                 To connect to the ASMI from a management console, complete the following steps:
                 a. Open Systems Management from the navigation pane.
                 b. From the work window, select one of the managed systems.
                 c. From the System Management tasks list, click Operations → Launch Advanced
                    System Management (ASMI).
126   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
 Use a web browser.
   At the time of writing, the supported web browsers are Microsoft Internet Explorer (Version
   10.0.9200.16439), Mozilla Firefox ESR (Version 24), and Chrome (Version 30). Later
   versions of these browsers might work, but are not officially supported. The JavaScript
   language and cookies must be enabled, and TLS 1.2 might need to be enabled.
   The web interface is available during all phases of system operation, including the initial
   program load (IPL) and run time. However, several of the menu options in the web
   interface are unavailable during IPL or run time to prevent usage or ownership conflicts if
   the system resources are in use during that phase. The ASMI provides an SSL web
   connection to the service processor. To establish an SSL connection, open your browser
   go to the following address:
   https://<ip_address_of_service_processor>
    Note: To make the connection through Internet Explorer, click Tools Internet Options.
    Clear the Use TLS 1.0 check box, and click OK.
The operator panel is formed of two parts: One is always installed, and the second might be
optional.
              The LCD operator panel is optional in some systems, but there must be at least one in a
              server in a rack containing any IBM Power S914, Power S922, or Power S924 server. The
              panel can be accessed by using the switches on the front panel.
              The System Management Services (SMS) error log is accessible through the SMS menus.
              This error log contains errors that are found by partition firmware when the system or partition
              is starting.
The service processor’s error log can be accessed on the ASMI menus.
              You can also access the system diagnostics from a Network Installation Management (NIM)
              server.
              IBM i and its associated machine code provide dedicated service tools (DSTs) as part of the
              IBM i licensed machine code (Licensed Internal Code) and System Service Tools (SSTs) as
              part of IBM i. DSTs can be run in dedicated mode (no operating system is loaded). DSTs and
              diagnostic tests are a superset of those available under SSTs.
              The IBM i End Subsystem (ENDSBS *ALL) command can shut down all IBM and customer
              applications subsystems except for the controlling subsystem QTCL. The Power Down
              System (PWRDWNSYS) command can be set to power down the IBM i partition and restart the
              partition in DST mode.
              You can start SSTs during normal operations, which keep all applications running, by using
              the IBM i Start Service Tools (STRSST) command (when signed on to IBM i with the
              appropriately secured user ID).
              With DSTs and SSTs, you can look at various logs, run various diagnostic tests, or take
              several kinds of system memory dumps or other options.
128   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
          Depending on the operating system, the following service-level functions are what you
          typically see when you use the operating system service menus:
           Product activity log
           Trace Licensed Internal Code
           Work with communications trace
           Display/Alter/Dump
           Licensed Internal Code log
           Main storage memory dump manager
           Hardware service manager
           Call Home/Customer Notification
           Error information menu
           LED management menu
           Concurrent/Non-concurrent maintenance (within scope of the OS)
           Managing firmware levels:
            – Server
            – Adapter
           Remote support (access varies by OS)
          Each LPAR reports errors that it detects and forwards the event to the SFP application that is
          running on the management console, without determining whether other LPARs also detect
          and report the errors. For example, if one LPAR reports an error for a shared resource, such
          as a managed system power supply, other active LPARs might report the same error.
          By using the Manage Serviceable Events task in the management console, you can avoid
          long lists of repetitive Call Home information by recognizing that they are repeated errors and
          consolidating them into one error.
          In addition, you can use the Manage Serviceable Events task to initiate service functions on
          systems and LPARs, including the exchanging of parts, configuring connectivity, and
          managing memory dumps.
          Firmware entitlement
          With the new HMC Version V8R8.1.0.0 and Power Systems servers, the firmware installations
          are restricted to entitled servers. The customer must be registered with IBM and entitled by a
          service contract. During the initial machine warranty period, the access key is installed in the
          machine by manufacturing. The key is valid for the regular warranty period plus some extra
          time. The Power Systems Firmware is relocated from the public repository to the access
          control repository. The I/O firmware remains on the public repository, but the server must be
          entitled for installation. When the lslic command is run to display the firmware levels, a new
          value, update_access_key_exp_date, is added. The HMC GUI and the ASMI menu show the
          Update access key expiration date.
              Any firmware release that was made available during the entitled time frame can still be
              installed. For example, if the entitlement period ends on 31 December 2014, and a new
              firmware release is release before the end of that entitlement period, then it can still be
              installed. If that firmware is downloaded after 31 December 2014, but it was made available
              before the end of the entitlement period, it still can be installed. Any newer release requires a
              new update access key.
               Note: The update access key expiration date requires a valid entitlement of the system to
               perform firmware updates.
              Firmware updates
              System firmware is delivered as a release level or a service pack. Release levels support the
              general availability (GA) of new functions or features, and new machine types or models.
              Upgrading to a higher release level is disruptive to customer operations. IBM intends to
              introduce no more than two new release levels per year. These release levels will be
              supported by service packs. Service packs are intended to contain only firmware fixes and
              not introduce new functions. A service pack is an update to an existing release level.
              If the system is managed by a management console, you use the management console for
              firmware updates. By using the management console, you can take advantage of the CFM
              option when concurrent service packs are available. CFM is the IBM Power Systems
              Firmware updates that can be partially or wholly concurrent or nondisruptive. With the
              introduction of CFM, IBM is increasing its clients’ opportunity to stay on a given release level
              for longer periods. Clients that want maximum stability can defer until there is a compelling
              reason to upgrade, such as the following reasons:
               A release level is approaching its end of service date (that is, it has been available for
                about a year, and soon service will not be supported).
               Move a system to a more standardized release level when there are multiple systems in
                an environment with similar hardware.
               A new release has a new function that is needed in the environment.
               A scheduled maintenance action causes a platform restart, which provides an opportunity
                to also upgrade to a new firmware release.
              The updating and upgrading of system firmware depends on several factors, such as whether
              the system is stand-alone or managed by a management console, the firmware that is
              installed, and what operating systems are running on the system. These scenarios and the
              associated installation instructions are comprehensively outlined in the firmware section of
              Fix Central.
              You might also want to review the preferred practice white papers that are found at
              Service and support best practices for Power Systems.
130   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                Firmware update steps
                The system firmware consists of service processor microcode, Open Firmware microcode,
                and SPCN microcode.
                The firmware and microcode can be downloaded and installed either from an HMC, from a
                running partition, or from USB port number 1 on the rear, if that system is not managed by an
                HMC.
                Power Systems Firmware has a permanent firmware boot side (A side) and a temporary
                firmware boot side (B side). New levels of firmware must be installed first on the temporary
                side to test the update’s compatibility with existing applications. When the new level of
                firmware is approved, it can be copied to the permanent side.
                For access to the initial websites that address this capability, see Support for IBM Systems.
                For Power Systems, select the Power link.
                Although the content under the Popular links section can change, click the Firmware and
                HMC updates link to go to the resources for keeping your system’s firmware current.
                If there is an HMC to manage the server, the HMC interface can be used to view the levels of
                server firmware and power subsystem firmware that are installed and that are available to
                download and install.
                Each IBM Power Systems server has the following levels of server firmware and power
                subsystem firmware:
                 Installed level
                    This level of server firmware or power subsystem firmware is installed and will be installed
                    into memory after the managed system is powered off and then powered on. It is installed
                    on the temporary side of system firmware.
                 Activated level
                    This level of server firmware or power subsystem firmware is active and running in
                    memory.
                 Accepted level
                    This level is the backup level of the server or power subsystem firmware. You can return to
                    this level of server or power subsystem firmware if you decide to remove the installed
                    level. It is installed on the permanent side of system firmware.
                IBM provides the CFM function on selected Power Systems servers. This function supports
                applying nondisruptive system firmware service packs to the system concurrently (without
                requiring a restart operation to activate changes). For systems that are not managed by an
                HMC, the installation of system firmware is always disruptive.
PP Package identifier 01 -
01VL910,73,73 = POWER9 Entry Systems Firmware for 9009-41A, 9009-22A, and 9009-42A
              Otherwise, an installation is concurrent if the service pack level (FFF) of the new firmware is
              higher than the service pack level that is installed on the system and the conditions for
              disruptive installation are not met.
              With the Power-On Reset Engine (PORE), the firmware can now dynamically power off
              processor components, change the registers, and reinitialize while the system is running,
              without discernible impact to any applications running on a processor, which potentially
              allows concurrent firmware changes in POWER9, which in earlier designs required a restart
              to take effect.
              Activating new firmware functions requires the installation of a firmware release level
              (upgrades). This process is disruptive to server operations, and requires a scheduled outage
              and full server restart.
132   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
2.11.4 Electronic Services and Electronic Service Agent
           IBM transformed its delivery of hardware and software support services to help you achieve
           higher system availability. Electronic Services is a web-enabled solution that offers an
           exclusive, no additional charge enhancement to the service and support that is available for
           IBM servers. These services provide the opportunity for greater system availability with faster
           problem resolution and preemptive monitoring. The Electronic Services solution consists of
           two separate, but complementary, elements:
            Electronic Services news page
            Electronic Service Agent
           Early knowledge about potential problems enables IBM to deliver proactive service that can
           result in higher system availability and performance. In addition, information that is collected
           through the ESA is made available to IBM Support Services Representatives (IBM SSRs)
           when they help answer your questions or diagnose problems. Installation and use of ESA for
           problem reporting enables IBM to provide better support and service for your IBM server.
           To learn how Electronic Services can work for you, see IBM Electronic Services (an IBM ID is
           required).
134   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                The easiest way to set up the ESA is by using the wizard in the HMC GUI, which is shown in
                Figure 2-32.
                The wizard guides the user through the necessary steps, including entering details about the
                location of the system, the contact details, and other details. The user can select which HMC
                or HMCs should be used, as shown in Figure 2-33.
Figure 2-33 Managing which HMCs are used for Call Home
                 The user can then navigate the SEM menus to see the events for this server, as shown in
                 Figure 2-35.
136     IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
                                                                                                      3
    Chapter 3.   Virtualization
                 Virtualization is a key factor for productive and efficient use of IBM Power Systems servers. In
                 this chapter, you find a brief description of virtualization technologies that are available for
                 POWER9. The following IBM Redbooks publications provide more information about the
                 virtualization features:
                    IBM PowerVM Best Practices, SG24-8062
                    IBM PowerVM Virtualization Introduction and Configuration, SG24-7940
                    IBM PowerVM Virtualization Active Memory Sharing, REDP-4470
                    IBM PowerVM Virtualization Managing and Monitoring, SG24-7590
                    IBM Power Systems SR-IOV: Technical Overview and Introduction, REDP-5065
              Combined with features in the POWER9 processors, the IBM POWER Hypervisor delivers
              functions that enable other system technologies, including logical partitioning (LPAR)
              technology, virtualized processors, IEEE VLAN-compatible virtual switch, virtual SCSI
              adapters, virtual Fibre Channel adapters, and virtual consoles. The POWER Hypervisor is a
              basic component of the system’s firmware and offers the following functions:
               Provides an abstraction between the physical hardware resources and the LPARs that use
                them.
               Enforces partition integrity by providing a security layer between LPARs.
               Controls the dispatch of virtual processors to physical processors.
               Saves and restores all processor state information during a logical processor context
                switch.
               Controls hardware I/O interrupt management facilities for LPARs.
               Provides virtual LAN channels between LPARs that help reduce the need for physical
                Ethernet adapters for inter-partition communication.
               Monitors the service processor and performs a reset or reload if it detects the loss of the
                service processor, notifying the operating system if the problem is not corrected.
              The POWER Hypervisor is always active, regardless of the system configuration and also
              when not connected to the managed console. It requires memory to support the resource
              assignment to the LPARs on the server. The amount of memory that is required by the
              POWER Hypervisor firmware varies according to several factors:
               Memory that is required for hardware page tables (HPT).
               Memory that is required to support I/O devices.
               Memory that is required for virtualization.
138   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
The amount of memory for the HPT is based on the maximum memory size of the partition
and the HPT ratio. The default HPT ratio is either 1/64th of the maximum (for IBM i partitions)
or 1/128th (for AIX, Virtual I/O Server (VIOS), and Linux partitions) of the maximum memory
size of the partition. AIX, VIOS, and Linux use larger page sizes (16 and 64 KB) instead of
using 4 KB pages. Using larger page sizes reduces the overall number of pages that must be
tracked, so the overall size of the HPT can be reduced. As an example, for an AIX partition
with a maximum memory size of 256 GB, the HPT would be 2 GB.
When defining a partition, the maximum memory size that is specified should be based on the
amount of memory that can be dynamically added to the dynamic partition (DLPAR) without
having to change the configuration and restart the partition.
In addition to setting the maximum memory size, the HPT ratio can also be configured. The
hpt_ratio parameter for the chsyscfg Hardware Management Console (HMC) command can
be issued to define the HPT ratio that is used for a partition profile. The valid values are 1:32,
1:64, 1:128, 1:256, or 1:512. Specifying a smaller absolute ratio (1/512 is the smallest value)
decreases the overall memory that is assigned to the HPT. Testing is required when changing
the HPT ratio because a smaller HPT might incur more CPU consumption because the
operating system might need to reload the entries in the HPT more frequently. Most
customers have chosen to use the IBM provided default values for the HPT ratios.
              The POWER Hypervisor provides the following types of virtual I/O adapters:
                 Virtual SCSI
                 Virtual Ethernet
                 Virtual Fibre Channel
                 Virtual (TTY) console
140   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
3.2 POWER processor modes
       Although they are not virtualization features, the POWER processor modes are described
       here because they affect various virtualization features.
       On Power System servers, partitions can be configured to run in several modes, including the
       following modes:
        POWER7 compatibility mode
          This is the mode for POWER7+ and POWER7 processors, implementing Version 2.06 of
          the IBM Power Instruction Set Architecture (ISA). For more information, see
          IBM Knowledge Center.
        POWER8 compatibility mode
          This is the native mode for POWER8 processors implementing Version 2.07 of the IBM
          Power ISA. For more information, see IBM Knowledge Center.
        POWER9 compatibility mode
          This is the native mode for POWER9 processors implementing Version 3.0 of the IBM
          Power ISA. For more information, see IBM Knowledge Center.
                 Table 3-1 shows an example where the processor mode must be selected when a migration
                 from POWER9 to POWER8 is planned.
 POWER9           POWER9           Fails because the wanted processor      Fails because the wanted processor
                                   mode is not supported on destination.   mode is not supported on destination.
 POWER9           POWER8           Fails because the wanted processor      Fails because the wanted processor
                                   mode is not supported on destination.   mode is not supported on destination.
                 This feature enables memory expansion on the system. By using compression and
                 decompression of memory, content can effectively expand the maximum memory capacity,
                 providing more server workload capacity and performance.
                 Active Memory Expansion is a technology that allows the effective maximum memory
                 capacity to be much larger than the true physical memory maximum. Compression and
                 decompression of memory content can allow memory expansion up to 125% for AIX
                 partitions, which in turn enables a partition to perform more work or support more users with
                 the same physical amount of memory. Similarly, it can allow a server to run more partitions
                 and do more work for the same physical amount of memory.
                  Note: The Active Memory Expansion feature is not supported by IBM i and the Linux
                  operating system.
142    IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
3.4 Single Root I/O Virtualization
         Single Root I/O Virtualization (SR-IOV) is an extension to the Peripheral Component
         Interconnect Express (PCIe) specification that allows multiple operating systems to
         simultaneously share a PCIe adapter with little or no runtime involvement from a hypervisor or
         other virtualization intermediary.
         SR-IOV is PCI standard architecture that enables PCIe adapters to become self-virtualizing. It
         enables adapter consolidation, through sharing, much like logical partitioning enables server
         consolidation. With an adapter capable of SR-IOV, you can assign virtual slices of a single
         physical adapter to multiple partitions through logical ports; all of this is done without a VIOS.
         For more information, see IBM Power Systems SR-IOV: Technical Overview and Introduction,
         REDP-5065.
3.5 PowerVM
         The PowerVM platform is the family of technologies, capabilities, and offerings that delivers
         industry-leading virtualization on IBM Power Systems servers. It is the umbrella branding
         term for Power Systems virtualization (logical partitioning, IBM Micro-Partitioning®, POWER
         Hypervisor, VIOS, LPM, and more). As with Advanced Power Virtualization in the past,
         PowerVM is a combination of hardware enablement and software.
          The Power S914, Power S922, and Power S924 servers come with PowerVM Enterprise
          Edition (#5228) by default. Furthermore, a temporary PowerVM Enterprise license
          (#ELPM) is delivered for an old server to support a seamless move to POWER9 at no
          additional cost.
         Logical partitions
         LPARs and virtualization increase the usage of system resources and add a level of
         configuration possibilities.
         Logical partitioning is the ability to make a server that is run as though it were two or more
         independent servers. When you logically partition a server, you divide the resources on the
         server into subsets called LPARs. You can install software on an LPAR, and the LPAR runs as
         an independent logical server with the resources that you have allocated to the LPAR. LPAR
         is the equivalent of a virtual machine (VM).
         You can assign processors, memory, and input/output devices to LPARs. You can run AIX,
         IBM i, Linux, and the VIOS in LPARs. The VIOS provides virtual I/O resources to other LPARs
         with general-purpose operating systems.
         LPARs share a few system attributes, such as the system serial number, system model, and
         processor feature code (FC). All other system attributes can vary from one LPAR to another.
              On the POWER9 processors, a partition can be defined with a processor capacity as small as
              0.05processing units. The shared processor partitions are dispatched and time-sliced on the
              physical processors under the control of the POWER Hypervisor. The shared processor
              partitions are created and managed by the HMC.
              The Power S914 server supports up to eight cores in a single system. Here are the maximum
              numbers:
               Eight dedicated partitions
               160 micropartitions (maximum of 20 micropartitions per physical active core)
              The Power S922 server supports up to 20 cores in a single system. Here are the maximum
              numbers:
               20 dedicated partitions
               400 micropartitions (maximum of 20 micropartitions per physical active core)
              The Power S924 server supports up to 24 cores in a single system. Here are the maximum
              numbers:
               24 dedicated partitions
               480 micropartitions (maximum of 20 micropartitions per physical active core)
              The maximum amounts are supported by the hardware, but the practical limits depend on
              application workload demands.
              Processing mode
              When you create an LPAR, you can assign entire processors for dedicated use, or you can
              assign partial processing units from a shared processor pool. This setting defines the
              processing mode of the LPAR.
              Dedicated mode
              In dedicated mode, physical processors are assigned as a whole to partitions. The SMT
              feature in the POWER9 processor core allows the core to run instructions from two, four, or
              eight independent software threads simultaneously.
144   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
           Shared dedicated mode
           On POWER9 processor technology-based servers, you can configure dedicated partitions to
           become processor donors for idle processors that they own, allowing for the donation of spare
           CPU cycles from dedicated processor partitions to a shared processor pool. The dedicated
           partition maintains absolute priority for dedicated CPU cycles. Enabling this feature can help
           increase system usage without compromising the computing power for critical workloads in a
           dedicated processor.
           Shared mode
           In shared mode, LPARs use virtual processors to access fractions of physical processors.
           Shared partitions can define any number of virtual processors (the maximum number is 20
           times the number of processing units that are assigned to the partition). The POWER
           Hypervisor dispatches virtual processors to physical processors according to the partition’s
           processing units entitlement. One processing unit represents one physical processor’s
           processing capacity. All partitions receive a total CPU time equal to their processing unit’s
           entitlement. The logical processors are defined on top of virtual processors. So, even with a
           virtual processor, the concept of a logical processor exists, and the number of logical
           processors depends on whether SMT is turned on or off.
           Micropartitions are created and then identified as members of either the default processor
           pool or a user-defined shared processor pool. The virtual processors that exist within the set
           of micropartitions are monitored by the POWER Hypervisor, and processor capacity is
           managed according to user-defined attributes.
           If the Power Systems server is under heavy load, each micropartition within a shared
           processor pool is assured of its processor entitlement, plus any capacity that it might be
           allocated from the reserved pool capacity if the micropartition is uncapped.
           If certain micropartitions in a shared processor pool do not use their capacity entitlement, the
           unused capacity is ceded and other uncapped micropartitions within the same shared
           processor pool are allocated the additional capacity according to their uncapped weighting. In
           this way, the entitled pool capacity of a shared processor pool is distributed to the set of
           micropartitions within that shared processor pool.
           All Power Systems servers that support the MSPPs capability have a minimum of one (the
           default) shared processor pool and up to a maximum of 64 shared processor pools.
                                                                                                 Virtual SCSI
                                Physical Ethernet         Virtual Ethernet                         Adapter
                                    Adapter                   Adapter
      Physical
       Disk
                                                                                               Virtual I/O Client 2
                                  Physical Disk               Virtual SCSI
                                    Adapter                     Adapter                         Virtual Ethernet
                                                                                                    Adapter
      Physical
       Disk                                                                                      Virtual SCSI
                                                                                                   Adapter
                     By using the SEA, several client partitions can share one physical adapter, and you can
                     connect internal and external VLANs by using a physical adapter. The SEA service can be
                     hosted only in the VIOS, not in a general-purpose AIX or Linux partition, and acts as a Layer
                     2 network bridge to securely transport network traffic between virtual Ethernet networks
                     (internal) and one or more (Etherchannel) physical network adapters (external). These virtual
                     Ethernet network adapters are defined by the POWER Hypervisor on the VIOS.
                     Virtual SCSI
                     Virtual SCSI is used to see a virtualized implementation of the SCSI protocol. Virtual SCSI is
                     based on a client/server relationship. The VIOS LPAR owns the physical resources and acts
                     as a server or, in SCSI terms, a target device. The client LPARs access the virtual SCSI
                     backing storage devices that are provided by the VIOS as clients.
146     IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
            The virtual I/O adapters (virtual SCSI server adapter and a virtual SCSI client adapter) are
            configured by using a managed console or through the Integrated Virtualization Manager on
            smaller systems. The virtual SCSI server (target) adapter is responsible for running any SCSI
            commands that it receives. It is owned by the VIOS partition. The virtual SCSI client adapter
            allows a client partition to access physical SCSI and SAN-attached devices and LUNs that
            are assigned to the client partition. The provisioning of virtual disk resources is provided by
            the VIOS.
            N_Port ID Virtualization
            N_Port ID Virtualization (NPIV) is a technology that allows multiple LPARs to access
            independent physical storage through the same physical Fibre Channel adapter. This adapter
            is attached to a VIOS partition that acts only as a pass-through, managing the data transfer
            through the POWER Hypervisor.
            Each partition has one or more virtual Fibre Channel adapters, each with their own pair of
            unique worldwide port names, enabling you to connect each partition to independent physical
            storage on a SAN. Unlike virtual SCSI, only the client partitions see the disk.
            For more information and requirements for NPIV, see IBM PowerVM Virtualization Managing
            and Monitoring, SG24-7590.
            LPM provides systems management flexibility and improves system availability by:
             Avoiding planned outages for hardware upgrade or firmware maintenance.
             Avoiding unplanned downtime. With preventive failure management, if a server indicates a
              potential failure, you can move its LPARs to another server before the failure occurs.
            For more information and requirements for NPIV, see IBM PowerVM Live Partition Mobility,
            SG24-7460.
            The physical memory of an IBM Power System can be assigned to multiple partitions in either
            dedicated or shared mode. A system administrator can assign some physical memory to a
            partition and some physical memory to a pool that is shared by other partitions. A single
            partition can have either dedicated or shared memory:
             With a pure dedicated memory model, the system administrator’s task is to optimize
              available memory distribution among partitions. When a partition suffers degradation
              because of memory constraints and other partitions have unused memory, the
              administrator can manually issue a dynamic memory reconfiguration.
             With a shared memory model, the system automatically decides the optimal distribution of
              the physical memory to partitions and adjusts the memory assignment based on partition
              load. The administrator reserves physical memory for the shared memory pool, assigns
              partitions to the pool, and provides access limits to the pool.
              Active Memory Deduplication allows the POWER Hypervisor to map dynamically identical
              partition memory pages to a single physical memory page within a shared memory pool. This
              way enables a better usage of the Active Memory Sharing shared memory pool, increasing
              the system’s overall performance by avoiding paging. Deduplication can cause the hardware
              to incur fewer cache misses, which also leads to improved performance.
              Active Memory Deduplication depends on the Active Memory Sharing feature being available,
              and it uses CPU cycles that are donated by the Active Memory Sharing pool’s VIOS partitions
              to identify deduplicated pages. The operating systems that are running on the Active Memory
              Sharing partitions can suggest to the POWER Hypervisor that some pages (such as
              frequently referenced read-only code pages) are good for deduplication.
              The HMC version V9R1 brings following enhancements to the Remote Restart feature.
               Remote restart a partition with reduced or minimum CPU/memory on the target system.
               Remote restart by choosing a different virtual switch on the target system.
               Remote restart the partition without turning on the partition on the target system.
               Remote restart the partition for test purposes when the source-managed system is in the
                Operating or Standby state.
               Remote restart through the REST API.
148   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Related publications
                 The publications that are listed in this section are considered suitable for a more detailed
                 description of the topics that are covered in this paper.
IBM Redbooks
                 The following IBM Redbooks publications provide more information about the topics in this
                 document. Some publications that are referenced in this list might be available in softcopy
                 only.
                  IBM PowerAI: Deep Learning Unleashed on IBM Power Systems Servers, SG24-8409
                  IBM Power System AC922 Introduction and Technical Overview, REDP-5472
                  IBM Power System AC922 Introduction and Technical Overview, REDP-5494
                  IBM Power System L922 Introduction and Technical Overview, REDP-5496
                  IBM Power Systems LC921 and LC922 Introduction and Technical Overview, REDP-5495
                  IBM Power Systems H922 and H924 Introduction and Technical Overview, REDP-5498
                  IBM Power System S822LC for High Performance Computing Introduction and Technical
                   Overview, REDP-5405
                  IBM PowerVM Best Practices, SG24-8062
                  IBM PowerVM Virtualization Introduction and Configuration, SG24-7940
                  IBM PowerVM Virtualization Managing and Monitoring, SG24-7590
                 You can search for, view, download, or order these documents and other Redbooks
                 publications, Redpapers, web docs, drafts, and additional materials, at the following website:
                 ibm.com/redbooks
Online resources
                 These websites are also relevant as further information sources:
                  IBM Fix Central website
                     http://www.ibm.com/support/fixcentral/
                  IBM Knowledge Center
                     http://www.ibm.com/support/knowledgecenter/
                  IBM Knowledge Center: IBM Power Systems Hardware
                     http://www-01.ibm.com/support/knowledgecenter/api/redirect/powersys/v3r1m5/inde
                     x.jsp
                  IBM Knowledge Center: Migration combinations of processor compatibility modes for
                   active Partition Mobility
                     http://www-01.ibm.com/support/knowledgecenter/api/redirect/powersys/v3r1m5/topi
                     c/p7hc3/iphc3pcmcombosact.htm
150   IBM Power Systems S922, S914, and S924: Technical Overview and Introduction
Back cover
REDP-5497-00
ISBN 0738456934
Printed in U.S.A.
                                     ®
             ibm.com/redbooks