IBM z15 (8561) Technical Guide: Books
IBM z15 (8561) Technical Guide: Books
IBM z15 (8561) Technical Guide: Books
Octavian Lascu
Bill White
John Troy
Jannie Houlbjerg
Kazuhiro Nakajima
Paul Schouten
Anna Shugol
Frank Packheiser
Hervey Kamga
Bo Xu
Redbooks
IBM Redbooks
August 2020
SG24-8851-00
Note: Before using this information and the product it supports, read the information in “Notices” on
page xiii.
This edition applies to IBM z15 Model T01, Machine Type 8561.
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiv
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Chapter 1. Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Design considerations for the IBM z15 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.1 Complementing and augmenting cloud solutions . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.2 Compliance, resiliency, and performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.3 Pervasive encryption. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.4 IBM Z Data Privacy Passports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.5 Blending open source with IBM Z state-of-the-art technologies . . . . . . . . . . . . . . . 3
1.2 z15 server highlights . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.1 Processor and memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.2 Capacity and performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.3 Virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.4 I/O subsystem and I/O features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.2.5 Reliability, availability, and serviceability design . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.3 z15 server technical overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.3.1 Model and features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.3.2 Model upgrade paths . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3.3 Frames . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.3.4 CPC drawer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.3.5 I/O connectivity: PCIe+ Generation 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
1.3.6 I/O subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
1.3.7 I/O and special purpose features in the PCIe I/O drawer . . . . . . . . . . . . . . . . . . . 21
1.3.8 Storage connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
1.3.9 Network connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
1.3.10 Coupling and Server Time Protocol connectivity . . . . . . . . . . . . . . . . . . . . . . . . 25
1.3.11 Cryptography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
1.4 Reliability, availability, and serviceability. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
1.5 Hardware Management Consoles and Support Elements . . . . . . . . . . . . . . . . . . . . . . 31
1.6 Operating systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
1.6.1 Supported operating systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
1.6.2 IBM compilers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
Contents v
4.5.1 I/O feature card ordering information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165
4.5.2 Physical channel ID report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166
4.6 Connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
4.6.1 I/O feature support and configuration rules. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
4.6.2 Storage connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
4.6.3 Network connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 181
4.6.4 Parallel Sysplex connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
4.7 Cryptographic functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
4.7.1 CPACF functions (FC 3863) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
4.7.2 Crypto Express7S feature (FC 0898 and FC 0899) . . . . . . . . . . . . . . . . . . . . . . 200
4.7.3 Crypto Express6S feature (FC 0893) as carry forward only . . . . . . . . . . . . . . . . 201
4.7.4 Crypto Express5S feature (FC 0890) as carry forward only . . . . . . . . . . . . . . . . 202
4.8 Integrated Firmware Processor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
Contents vii
8.5 On/Off Capacity on Demand . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357
8.5.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357
8.5.2 Capacity Provisioning Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 358
8.5.3 Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 359
8.5.4 On/Off CoD testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 362
8.5.5 Activation and deactivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
8.5.6 Termination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 364
8.6 z/OS Capacity Provisioning. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365
8.7 System Recovery Boost Upgrade . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369
8.8 Capacity for Planned Event. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 370
8.9 Capacity Backup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 372
8.9.1 Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 372
8.9.2 CBU activation and deactivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374
8.9.3 Automatic CBU enablement for GDPS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375
8.10 Planning for nondisruptive upgrades. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376
8.10.1 Components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376
8.10.2 Concurrent upgrade considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 377
8.11 Summary of Capacity on-Demand offerings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 380
Contents ix
12.3 Fundamental components of workload performance . . . . . . . . . . . . . . . . . . . . . . . . 479
12.3.1 Instruction path length. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 480
12.3.2 Instruction complexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 480
12.3.3 Memory hierarchy and memory nest. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 480
12.4 Relative Nest Intensity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 481
12.5 LSPR workload categories based on RNI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 483
12.6 Relating production workloads to LSPR workloads . . . . . . . . . . . . . . . . . . . . . . . . . 483
12.7 CPU MF counter data and LSPR workload type . . . . . . . . . . . . . . . . . . . . . . . . . . . . 484
12.8 Workload performance variation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 485
12.9 Capacity planning consideration for z15 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 485
12.9.1 Collect CPU MF counter data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 485
12.9.2 Creating EDF file with CP3KEXTR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 486
12.9.3 Loading EDF file to the capacity planning tool . . . . . . . . . . . . . . . . . . . . . . . . . 486
12.9.4 Tips to maximize z15 server capacity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 486
Contents xi
xii IBM z15 (8561) Technical Guide
Notices
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
AIX® IBM z13® System z®
Bluemix® IBM z13s® System z10®
CICS® IBM z14® System z9®
Db2® IBM z15™ VIA®
DB2® Interconnect® VTAM®
Distributed Relational Database Language Environment® Watson™
Architecture™ MVS™ WebSphere®
DS8000® OMEGAMON® z Systems®
FICON® Parallel Sysplex® z/Architecture®
FlashCopy® Passport Advantage® z/OS®
GDPS® PowerPC® z/VM®
Global Technology Services® RACF® z/VSE®
HyperSwap® Redbooks® z13®
IBM® Redbooks (logo) ® z13s®
IBM Watson® Resource Link® z15™
IBM Z® S/390® z9®
IBM z Systems® System Storage™ zEnterprise®
The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.
Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States,
other countries, or both.
Java, and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its
affiliates.
Red Hat, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the United States and
other countries.
UNIX is a registered trademark of The Open Group in the United States and other countries.
VMware, and the VMware logo are registered trademarks or trademarks of VMware, Inc. or its subsidiaries in
the United States and/or other jurisdictions.
Other company, product, or service names may be trademarks or service marks of others.
This IBM® Redbooks® publication describes the features and functions the latest member of
the IBM Z® platform, the IBM z15™ (machine type 8561). It includes information about the
IBM z15 processor design, I/O innovations, security features, and supported operating
systems.
The z15 is a state-of-the-art data and transaction system that delivers advanced capabilities,
which are vital to any digital transformation. The z15 is designed for enhanced modularity,
which is in an industry standard footprint. This system excels at the following tasks:
Making use of multicloud integration services
Securing data with pervasive encryption
Accelerating digital transformation with agile service delivery
Transforming a transactional platform into a data powerhouse
Getting more out of the platform with IT Operational Analytics
Accelerating digital transformation with agile service delivery
Revolutionizing business processes
Blending open source and Z technologies
This book explains how this system uses new innovations and traditional Z strengths to
satisfy growing demand for cloud, analytics, and open source technologies. With the z15 as
the base, applications can run in a trusted, reliable, and secure environment that improves
operations and lessens business risk.
Authors
This book was produced by a team of specialists from around the world working at IBM
Redbooks, Poughkeepsie Center.
Octavian Lascu is an IBM Redbooks Project Leader and a Senior IT Consultant for IBM
Romania with over 25 years of experience. He specializes in designing, implementing, and
supporting complex IT infrastructure environments (systems, storage, and networking),
including high availability and disaster recovery solutions and high-performance computing
deployments. He has developed materials for and taught over 50 workshops for technical
audiences around the world. He is the author of several IBM publications.
Bill White is an IBM Redbooks Project Leader and Senior Networking and Connectivity
Specialist at IBM Redbooks, Poughkeepsie Center.
John Troy is an IBM Z and storage hardware National Top Gun in the northeast area of the
United States. He has 40 years of experience in the service field. His areas of expertise
include IBM Z servers and high-end storage systems technical and customer support and
services. John has also been an IBM Z hardware technical support course designer,
developer, and instructor for the last eight generations of IBM high-end servers.
Jannie Houlbjerg is a Systems Programmer working at JN Data in Denmark. She has more
than 20 years of experience in the IBM Z field. Her areas of expertise include IBM Z hardware
and infrastructure, IBM Parallel Sysplex®, connectivity, performance, IBM GDPS®, and
technical project management and documentation.
Paul Schouten is an IBM Z Client Technical Specialist based in Sydney, Australia. During his
40 years supporting mainframe systems, he has performed many roles, including Certified IT
Architect, Systems Software developer, and Systems Programming. He has extensive
experience developing and documenting high availability solutions for IBM’s enterprise
customers.
Anna Shugol is a mainframe technical specialist with IBM UK. She has over 8 years of
Mainframe experience and has worked with clients in various geographies. Large system
performance is one of her key areas of expertise. Anna holds a Computer Science degree
from Bauman Moscow State Technical University.
Frank Packheiser is a Senior zIT Specialist at the Field Technical Sales Support office in
Germany. He has over 25 years of experience in IBM Z. Frank has worked for 10 years in the
IBM Education Center in Germany, developing and providing professional training. He also
provides professional services to IBM Z and mainframe clients. In 2008 and 2009, Frank
supported clients in Middle East/North Africa (MENA) as a zIT Architect. In addition to
co-authoring several IBM Redbooks publications since 1999, he has been an official ITSO
presenter at ITSO workshops since 2010.
Hervey Kamga is an IBM Z Product Engineer with the EMEA I/O Connectivity Team in
Montpellier, France. Before serving in his current role, he was a Support Engineer and
Engineer On Site for 13 years with Sun MicroSystems and Oracle in EMEA. Hervey’s areas of
expertise include Oracle Solaris (operating system and hardware products), virtualization
(VMware and virtualBox), Linux, and IBM Z I/O features and protocols (IBM FICON® and
OSA) while co-authoring several IBM Redbooks publications.
Robert Haimowitz
IBM Redbooks, Poughkeepsie Center
Dave Surman, Darelle Gent, David Hutton, Tom Dewkett, Frank Bosco, Patty Driever, Anthony
Sofia, Brian Valentine, Purvi Patel, Les Geer III, Bill Bitner, Christine Smith, Barbara Weiler,
Dean St Piere, Tom Morris, Ellen Carbarnes, Gregory Hutchison, Riaz Ahmad, Franco Pinto,
Walter Niklaus, Martin Recktenwald, Roan Dawkins
IBM
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an email to:
redbooks@us.ibm.com
Mail your comments to:
IBM Corporation, IBM Redbooks
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
Preface xvii
xviii IBM z15 (8561) Technical Guide
1
Chapter 1. Introduction
This chapter describes the basic concepts and design considerations around IBM z15TM
servers and includes the following topics:
1.1, “Design considerations for the IBM z15” on page 2
1.2, “z15 server highlights” on page 4
1.3, “z15 server technical overview” on page 12
1.4, “Reliability, availability, and serviceability” on page 30
1.5, “Hardware Management Consoles and Support Elements” on page 31
1.6, “Operating systems” on page 31
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, Features Max34, Max71, Max108,
Max145, and Max190, which is further identified as IBM z15 Model T01, or z15 T01,
unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, Features Max4, Max13, Max21, Max32,
and Max65, which is further identified as IBM z15 Model T02, or z15 T02, unless
otherwise specified.
In the remainder of this chapter, IBM z15 (z15) refers to both machine types.
There is no doubt that the technology you choose determines business success and
differentiates you from the competition. Technology choices must help meet the expectations
of rapid dynamic development cycles and give the confidence that new and existing services
are resilient and can be delivered quickly and securely. Therefore, the correct balance of open
source technologies and the IT platform on which they run is also key.
The IBM Integrated Accelerator for zEnterprise® Data Compression (zEDC) provides on-chip
compression (processor nest compression accelerator), which supports DEFLATE-compliant
compression and decompression and GZIP CRC/ZLIB Adler. Compared to the zEDC
Express (PCIe feature), the z15 on-chip compression provides low latency and high
bandwidth while eliminating the need for the virtualization layer. The nest accelerator is now
accessible as a designed instruction, which is available for problem state programs.
New with z15, the System Recovery Boost feature offers the customer more CP capacity
during system recovery operations, such as shutdown, IPL, and stand-alone dumps, to
“speed up” the recovery. The feature is embedded in the z15 firmware and is available to
operating systems (opt-in for supported OS-es) to accelerate recovery from maintenance
tasks by providing more CP capacity to opt in LPARs.
For more information about System Recovery Boost, see Chapter 7, “Operating system
support” on page 253 and Appendix B, “System Recovery Boost” on page 493.
Pervasive encryption can dramatically simplify data protection and reduce the costs that are
associated with regulatory compliance. By using simple policy controls, z15 pervasive
computing streamlines data protection for mission critical IBM Db2® for z/OS, IBM IMS, and
Virtual Storage Access Method (VSAM) datasets.
The Central Processor Assist for Cryptographic Function (CPACF), which is standard on
every core, supports pervasive encryption and provides hardware acceleration for encryption
operations. The new Crypto-Express7S gets a performance boost on z15. Combined, these
enhancements perform encryption more efficiently on the z15 than on earlier IBM Z servers.
With up to 40 TB of memory, z15 T01 can open opportunities, such as in-memory data marts
and in-memory analytics, while giving you the necessary room to tune applications for optimal
performance. By using the Vector Packed Decimal Facility that allows packed decimal
operations to be performed in registers rather than memory, and new fast mathematical
computations, compilers (such as Enterprise COBOL for z/OS, V6.2, Enterprise PL/I for z/OS,
V5.2, z/OS V2.4 XL C/C++), the COBOL optimizer, Automatic Binary Optimizer for z/OS,
V1.3, and Java, are optimized on z15. These compilers and optimizers are designed to
improve application performance and reduce CPU usage and operating costs.
Chapter 1. Introduction 3
Smoothly handling the data tsunami requires robust infrastructure that is designed specifically
for high-volume data transactions. To take advantage of new unstructured data,1 businesses
on IBM Z can use application programming interfaces (APIs) that can help with creating and
delivering innovative services.
Linux on IBM Z, which is optimized for open source software, brings more value to the
platform. Linux on IBM Z supports a wealth of new products that are familiar to application
developers, such as Python, Scala, Spark, MongoDB, PostgreSQL, and MariaDB. By
accessing core business data directly on platform (without the need for ETL2, and hence no
data offload off Z platform), you can develop new intelligent applications and business
processes.
As your business technology needs evolve to compete in today’s digital economy, IBM stands
ready to help with intelligent, robust, and comprehensive technology solutions. The IBM
approach integrates server, software, and storage solutions to ensure that each member of
the stack is designed and optimized to work together. The new IBM z15™ leads that
approach by delivering the power and speed users demand, the security users and regulators
require, and the operational efficiency that maximizes your bottom line.
Terminology note: The remainder of this book uses the designation CPC to refer to the
central processor complex.
IBM continues its technology leadership with the z15 server. The z15 T01 server is built by
using the IBM modular multi-processor drawer design that supports 1 - 5 processor drawers
per CPC. Each processor drawer contains four Processor Unit (PU) single-chip modules
(SCMs) and one Storage Controller (SC) SCM.
Both SCMs are redesigned by using 14 nm FINFET SOI technology.3 Each PU SCM has 12
processor units (PUs, or cores). In addition to SCMs, CPC drawers host memory DIMMs,
connectors for I/O, redundant power supplies, combined Flexible Service Processors and
Oscillators (FSP/OSC), and cooling manifolds.
1 This data accounts for 80% of all data that is generated today and is expected to grow to over 93% by 2020.
2
Extract, Transform, Load (ETL) database operations that extract data from one or more databases, and transform
and load it into another database for analyzing data in a different context.
3 FINFET is the industry solution; SOI is the IBM solution for SER.
Also featured is an expanded instruction set with Vector Packed Decimal Facility, Guarded
Storage Facility, Vector Facility enhancements, Semaphore Assist Facility, Order Preserving
Compression, Entropy Encoding for Co-processor Compression for better performance in
several different areas, and the IBM Integrated Accelerator for zEnterprise Data Compression
(on-chip compression accelerator).
Depending on the model, the z15 T01 server can support 512 GB - 40 TB of usable memory,
with up to 8 TB of usable memory per CPC drawer. In addition, a fixed amount of 256 GB is
reserved for the hardware system area (HSA) and is not part of customer-purchased memory.
Memory is implemented as a redundant array of independent memory (RAIM) and uses extra
physical memory as spare memory. The RAIM function accounts for 20% of the physical
installed memory in each CPC drawer.
New with z15 T01, Virtual Flash Memory (VFM) feature is offered from the main memory
capacity in 0.5 TB units (versus 1.5 TB per feature on z14 M0x) increasing granularity for the
feature. VFM provides much simpler management and better performance by eliminating the
I/O to the adapters in the PCIe+ I/O drawers. VFM does not require any application changes
when moving from IBM zFlash Express (previously available on z13® and zEC12).
The increased performance and the total system capacity available (with potential energy
savings) allows consolidation of diverse applications on a single platform with significant
financial savings. The introduction of new technologies and an expanded and enhanced
instruction set ensure that the z15 server is a high-performance, reliable, and rich-security
platform. The z15 server is designed to maximize the use of resources and allows you to
integrate and consolidate applications and data across the enterprise IT infrastructure.
z15 Model T01 server is offered with five maximum processor features, with 1 - 190
configurable PUs. The processor features Max34, Max71, Max108, and Max145 have one,
two, three, respective four CPC drawers with 41 active PUs per CPC drawer. The
high-capacity feature (Max190) has five processor (CPC) drawers with 43 PUs per drawer.
The z15 T01 feature Max190 is estimated to provide up to 25% more total system capacity
than the z14 Model M05, with the same amount of memory and power requirements. With up
to 40 TB of main storage and enhanced SMT, the performance of the z15 T01 processors
deliver considerable improvement. Uniprocessor performance also increased significantly. A
z15 Model 701 offers average performance improvements of up to 14%5 over the z14 Model
701.
The Integrated Facility for Linux (IFL) and IBM Z Integrated Information Processor (zIIP)
processor units on the z15 server can be configured to run two simultaneous threads per
clock cycle in a single processor (SMT). This feature increases the capacity of these
4
Simultaneous multithreading is two threads per core.
5 Observed performance increases vary depending on the workload types.
Chapter 1. Introduction 5
processors with 25% in average5 over processors that are running single thread. SMT is also
enabled by default on System Assist Processors (SAPs).
The z15 T01 server expands the subcapacity settings, offering three subcapacity levels (in
models 4xx, 5xx and 6xx) for up to 34 processors that are characterized as CPs (compared to
up to 33 processors for z14). This configuration gives a total of 292 distinct capacity settings.
The z15 servers deliver scalability and granularity to meet the needs of medium-sized
enterprises, while also satisfying the requirements of large enterprises that have demanding,
mission-critical transaction and data processing requirements.
This comparison is based on the Large System Performance Reference (LSPR) mixed
workload analysis. For more information about performance and workload variation on z15
servers, see Chapter 12, “Performance” on page 475.
z15 servers continue to offer all specialty engine types that are available on z14.
Workload variability
Consult the LSPR when considering performance on z15 servers. The range of performance
ratings across the individual LSPR workloads is likely to have a large spread. More
performance variation of individual logical partitions (LPARs) is available when an increased
number of partitions and more PUs are available. For more information, see Chapter 12,
“Performance” on page 475.
For more information about millions of service units (MSUs) ratings, see the IBM Z Software
Contracts website.
Capacity on demand
Capacity on demand (CoD) enhancements enable clients to have more flexibility in managing
and administering their temporary capacity requirements. The z15 server supports the same
architectural approach for CoD offerings as the z14 (temporary or permanent). Within the z15
server, one or more flexible configuration definitions can be available to solve multiple
temporary situations, and multiple capacity configurations can be active simultaneously.
Prepaid OOCoD tokensa: Beginning with IBM z15, new prepaid OOCoD tokens that are
purchased do not carry forward to future systems.
a. IBM’s statements regarding its plans, directions, and intent are subject to change or withdrawal
without notice at IBM’s sole discretion.
Up to 200 staged records can be created to handle many scenarios. Up to eight of these
records can be installed on the server at any time. After the records are installed, the
activation of the records can be done manually, or the z/OS Capacity Provisioning Manager
can automatically start the activation when Workload Manager (WLM) policy thresholds are
reached. Tokens are available that can be purchased for On/Off CoD before or after workload
execution (pre- or post-paid).
LPAR capping
IBM Processor Resource/Systems Manager (IBM PR/SM) offers different options to limit the
amount of capacity that is assigned to and used by an LPAR or a group of LPARs. By using
the Hardware Management Console (HMC), a user can define an absolute or a relative
capping value for LPARs that are running on the system.
z15 servers support IBM z/Architecture® mode only, which can be initialized in LPAR mode
(also known as PR/SM) or Dynamic Partition Manager (DPM) mode.
PR/SM mode
PR/SM is Licensed Internal Code (LIC) that manages and virtualizes all the installed and
enabled system resources as a single large symmetric multiprocessor (SMP) system. This
virtualization enables full sharing of the installed resources with high security and efficiency.
On z15 T01, the PR/SM supports configuring up to 85 LPARs, each of which includes logical
processors, memory, and I/O resources. Resources of these LPARs are assigned from the
installed CPC drawers and features. For more information about PR/SM functions, see 3.7,
“Logical partitioning” on page 132.
LPAR configurations can be dynamically adjusted to optimize the virtual servers’ workloads.
z15 servers provide improvements to the PR/SM HiperDispatch function. HiperDispatch
provides alignment of logical processors to physical processors that ultimately improves
cache utilization, minimizes inter-CPC drawer communication, and optimizes operating
system work dispatching, which combined results in increased throughput. For more
information, see “HiperDispatch” on page 96.
z15 PR/SM implements an Improved memory affinity algorithm and improved logical partition
placement algorithms based on z14 experience. PR/SM reoptimization is the process of
identifying “homes” for the partitions. PR/SM on z15 tries to assign all memory in one drawer
(single SC SCM, shared L4 cache) and attempts to consolidate storage onto drawers with the
most processor entitlement.
PR/SM also tries to assign all logical processors to one CPC drawer and packed into chips of
that drawer, in cooperation with operating system use of HiperDispatch. In z15, all processor
types can be dynamically reassigned, except IFPs.
HiperSockets
z15 servers support defining up to 32 IBM HiperSockets. HiperSockets provide for
memory-to-memory communication across LPARs without the need for any I/O adapters and
have virtual LAN (VLAN) capability.
Chapter 1. Introduction 7
– IBM z/VM®
– IBM z/VSE®
– z/TPF
– Linux on IBM Z
Coupling Facility: Coupling Facility Control Code (CFCC)
Linux only:
– Linux on IBM Z
– z/VM
z/VM
Secure Service Container:
– VNA (z/VSE Network Appliance)
– IBM High Security Business Network (HSBN)6
IBM Z servers also offer other virtual appliance-based solutions and support other the
following hypervisors and containerization:
IBM GDPS Virtual Appliance
KVM for IBM Z
Docker Enterprise Edition for Linux on IBM Systems7
For example, in a z/VM-mode LPAR, z/VM can manage Linux on IBM Z guests that are
running on IFL processors while also managing z/VSE and z/OS guests on CPs. It also
allows z/OS to fully use zIIPs.
SSC is a part of the Pervasive Encryption concept that was introduced with IBM z14, which is
aimed at delivering best IBM Security hardware and software enhancements, services, and
practices for 360-degree infrastructure protection.
SSC is a powerful IBM technology for providing the extra protection of the most sensitive
workloads. The integration with other applications in transparent; all services can be called
externally by standard REST APIs.
Chapter 1. Introduction 9
Coordinated IBM HyperSwap®
Single point of control
For a five CPC drawer system, up to 60 PCIe+ fanout slots can be configured for data
communications between the CPC drawers and the I/O infrastructure, and for coupling. The
multiple channel subsystem (CSS) architecture allows up to six CSSs, each with 256
channels.
For I/O constraint relief, four subchannel sets are available per CSS, which allows access to
many logical volumes. The fourth subchannel set allows extending the amount of addressable
external storage for Parallel Access Volumes (PAVs), Peer-to-Peer Remote Copy (PPRC)
secondary devices, and IBM FlashCopy® devices. z15 T01 supports Initial Program Load
(IPL) from subchannel set 1 (SS1), subchannel set 2 (SS2), or subchannel set 3 (SS3), and
subchannel set 0 (SS0). For more information, see “Initial program load from an alternative
subchannel set” on page 208.
The system I/O buses use the Peripheral Component Interconnect® Express (PCIe)
technology, which also is used in coupling links.
z15 T01 connectivity supports the following I/O or special purpose features:
Storage connectivity:
– Fibre Channel connection (IBM FICON):
• FICON Express16SA 10 KM long wavelength (LX) and short wavelength (SX)
• FICON Express16S+ 10 KM LX and SX (carry forward only)
• FICON Express16S 10 KM LX and SX (carry forward only)
• FICON Express8S 10 KM LX and SX (carry forward only)
– IBM zHyperLink Express1.1 (new build)
– IBM zHyperLink Express (carry forward)
Network connectivity:
– Open Systems Adapter (OSA):
• OSA-Express7S 25 GbE Short Reach1.1 (new build)
• OSA-Express7S 10 GbE long reach (LR) and short reach (SR) (new build)
• OSA-Express7S GbE LX and SX (new build)
• OSA-Express7S 1000BASE-T Ethernet (new build)
• OSA-Express7S 25GbE SR (carry forward)
• OSA-Express6S 10 GbE LR and SR (carry forward)
• OSA-Express6S GbE LX and SX (carry forward)
• OSA-Express6S 1000BASE-T Ethernet (carry forward)
• OSA-Express5S 10 GbE LR and SR (carry forward)
• OSA-Express5S GbE LX and SX (carry forward)
• OSA-Express5S 1000BASE-T Ethernet (carry forward)
– IBM HiperSockets
– Shared Memory Communication - Remote Direct Memory Access (SMC-R):
IBM Z servers are designed to enable highest availability and lowest downtime. These facts
are recognized by various IT analysts, such as ITIC8 and IDC9. Comprehensive, multi-layered
strategy includes the following features:
Error Prevention
Error Detection and Correction
Error Recovery
System Recovery Boost
With a properly configured z15 server, further reduction of outages can be attained through
First Failure Data Capture (FFDC), which is designed to reduce service times and avoid
subsequent errors. It also improves nondisruptive replace, repair, and upgrade functions for
memory, drawers, and I/O adapters. z15 servers support the nondisruptive download and
installation of LIC updates.
IBM z15™ RAS features provide unique high-availability and nondisruptive operational
capabilities that differentiate the Z servers in the marketplace. z15 RAS enhancements are
made on many components of the CPC (processor chip, memory subsystem, I/O, and
service) in areas, such as error checking, error protection, failure handling, error checking,
faster repair capabilities, sparing, and cooling.
The ability to cluster multiple systems in a Parallel Sysplex takes the commercial strengths of
the z/OS platform to higher levels of system management, scalable growth, and continuous
availability.
8
For more information, see ITIC Global Server Hardware, Server OS Reliability Report.
9 For more information, see Quantifying the Business Value of IBM Z.
Chapter 1. Introduction 11
The z15 processor builds upon the RAS of the z14 family with the following RAS
improvements:
System Recovery Boost
System Recovery Boost was introduced with IBM z15. It offers customers more Central
Processor (CP) capacity during system recovery operations to accelerate the system
startup (IPL), shutdown or stand-alone dump operations. System Recovery Boost requires
operating system support. No other IBM software charges are made during the boost
period.
System Recovery Boost might be used during LPAR IPL or LPAR shutdown to make the
running operating system and services available in a shorter period.
The System Recovery Boost provides the following options for the capacity increase:
– Subcapacity CP speed boost: During the boost period, subcapacity engines are
transparently activated at their full capacity (CP engines).
– zIIP Capacity Boost: During the boost period, all active zIIPs that are assigned to an
LPAR are used to extend the CP capacity (CP workload is dispatched to zIIP
processors during the boost period).
At the time of this writing, the main System Recovery Boost users are z/OS (running in an
LPAR), z/VM, and z/TPF.
z/VM uses the System Recovery Boost if it runs on subcapacity CP processors only (IFLs
are always at their full clock speed). Second-level z/VM guest operating systems10 can
inherit the boost if they are running on CPs.
Level 3 and Level 4 cache enhancements use symbol ECC to extend the reach of prior
z14 cache and memory improvements for improved availability. The level 3 and level 4
cache powerful symbol ECC is designed to make it resistant to more failure mechanisms.
Preemptive DRAM marking is added to the main memory to isolate and recover failures
more quickly.
On-chip compression accelerating compression operations on core level, for all LPARs,
which eliminates the virtualization layer that was needed for zEDC Express. The
technology replaces zEDC Express PCIe cards, which improves reliability and availability.
10 z/OS configured as a guest system under z/VM does not use the boost.
Systems with up to four drawers have 41 active PUs per CPC drawer, while feature Max190
(five CPC drawers) has 43 active PUs per drawer. A PU is the generic term for the IBM
z/Architecture processor unit (processor core) on the CP SCM.
On z15 servers, some PUs are part of the system base; that is, they are not part of the PUs
that can be purchased by clients. They include the following characteristics:
System assist processor (SAP) that is used by the channel subsystem. The number of
predefined SAPs depends on the z15 model.
One integrated firmware processor (IFP). The IFP is used in support of select features,
such as and RoCE Express.
Two spare PUs that can transparently assume any characterization during a permanent
failure of another PU.
The PUs that clients can purchase can assume any of the following characteristics:
CP for general-purpose use.
Integrated Facility for Linux (IFL) for the use of Linux on Z.
IBM Z Integrated Information Processor (zIIP) is designed to help free-up general
computing capacity and lower overall total cost of computing for select data and
transaction processing workloads.
zIIPs: At least one CP must be purchased with, or before, a zIIP can be purchased.
Clients can purchase up to two zIIPs for each purchased CP (assigned or unassigned)
on the system (2:1). However, during System Recovery Boost periods, the zIIP to CP
ratio can be greater than 2:1.
A PU that is not characterized cannot be used, but is available as a spare. The following rules
apply:
In the five-feature structure, at least one CP, ICF, or IFL must be purchased and activated
for any model.
PUs can be purchased in single PU increments and are orderable by feature code.
The total number of PUs purchased cannot exceed the total number that are available for
that model.
The number of installed zIIPs cannot exceed twice the number of installed CPs.
The multi-CPC drawer system design provides the capability to concurrently increase the
capacity of the system in the following ways:
Add capacity by concurrently activating more CPs, IFLs, ICFs, or zIIPs on a CPC drawer.
Add a CPC drawer concurrently and activate more CPs, IFLs, ICFs, or zIIPs.
Chapter 1. Introduction 13
Add a CPC drawer to provide more memory, or one or more adapters to support a larger
number of I/O features.
1.3.3 Frames
The z15 Model T01 system is designed in a new format that provides configuration flexibility
to fit customer requirements. The z15 T01 is build in a 19-inch form factor and can be
configured with one, two, three, or four frames, depending on processor and I/O
requirements.
Attention: The z15 system cannot be installed in a customer supplied frame. The system
comes configured and build in an 19-inch IBM frame.
The z15 T01 also offers two power choices: Bulk Power Assembly (BPA) and Intelligent Power
Distribution Units (iPDU, or PDU). The frames are A, B, C, and Z, bolted together, and feature
the following components:
Up to three CPC drawers in Frame A
Up to two CPC drawers in Frame B (CPC drawers in frame B are factory-installed only)
Up to 12 PCIe+ I/O drawers that hold I/O features and special purpose features
All CPC drawers and PCIe+ I/O drawers have redundant power supplies
For BPA systems only: Bulk Power Assemblies in Frames A and B with Optional Internal
Battery Feature (IBF)
For PDU systems only: Power Distribution Units in Frames A, B, and C (configuration
dependant)
CPC Drawer cooling units for either air or water cooling in Frames A and B
Two Ethernet switches in Frame A and two in Frame B (configuration dependent) to
interconnect the CPC components through Ethernet
Two 1U rack-mounted Support Elements (mounted in Frame A). The Support Elements
have a new service console (stored in Frame A), which can be connected in the front or
rear of a system.
Two CPC drawer sizes are available in z15 T01, depending on the number of active PU
(cores). The z15 model T01 Max190 has 43 active cores (PUs) per CPC drawer. All other z15
models have 41 active cores. The PU SCM has 12 cores by design, with 9, 10, or 11 active
cores, which can be characterized as CPs, IFLs, ICFs, zIIPs, SAPs, or IFPs.
The SCM provides a significant increase in system scalability and an extra opportunity for
server consolidation. All CPC drawers are fully interconnected by using high-speed
communication links through the L4 cache (in the SC SCM). This configuration allows the z15
Chapter 1. Introduction 15
server to be controlled by the PR/SM facility as a memory-coherent and cache-coherent SMP
system.
The PU configuration includes two designated spare PUs per system and a variable number
of SAPs. The SAPs scale with the number of CPC drawers that are installed in the server. For
example, four standard SAPs are available with one CPC drawer that is installed, and up to 22
standard SAPs for five CPC drawers installed. In addition, one PU is used as an IFP and is
not available for client use. The remaining PUs can be characterized as CPs, IFL processors,
zIIPs, ICF processors, or extra SAPs. For z15, the SAPs operate in Simultaneous
Multi-Threading (enabled by default, cannot be disabled).
The PU SCMs of the z15 T01 are cooled by a cold plate that is connected to an internal water
cooling loop. In an air-cooled system, the radiator units (RUs) exchange the heat from the
internal water loop with air. The RU has redundant pumps and blowers. The SC SCM is
air-cooled.
The z15 T01 server offers also a water-cooling option (using data center chilled water supply)
for increased system and data center energy efficiency. Water-cooling option is only available
with Bulk Power Assembly (BPA)-based systems. The water cooling units (WCUs) are fully
redundant in an N+1 arrangement.
Processor features
The processor core of the z15 T01 operates at 5.2 GHz. Depending on the z15 T01 feature,
41 - 215 active PUs are available on 1 - 5 CPC drawers.
Each core on the PU SCM includes an enhanced dedicated coprocessor for data
compression and cryptographic functions, which are known as the Central Processor Assist
for Cryptographic Functions (CPACF)11. Having standard clear key cryptographic
coprocessors that are integrated with the processor provides high-speed cryptography for
protecting data.
The z15 supports 64-bit addressing mode and uses Complex Instruction Set Computer
(CISC), including highly capable and thus complex instructions. Most of the instructions are
implemented at the hardware or firmware level for most optimal and effective execution.
Each PU is a superscalar processor, which can decode up to six complex instructions per
clock cycle, running instructions out-of-order. The PU uses a high-frequency, low-latency
pipeline that provides robust performance across a wide range of workloads.
Compared to its predecessor, the z15 processor design includes the following improvements
and architectural extensions:
Better performance and throughput:
– Fast processor units with enhanced microarchitecture
– Larger L2, L3 caches
– Up to 12 cores per processor chip
– More capacity (up to 190 characterizable processor units versus 170 on the z14)
– Larger cache (and shorter path to cache) means faster uniprocessor performance
11 Feature code (FC) 3863 must be ordered to enable CPACF. This feature code is available for no extra fee.
The instructions are issued, but previous state values are saved in transactional memory. If
the transaction succeeds, the saved values are discarded. If it fails, they are used to restore
Chapter 1. Introduction 17
the original values. Software can test the success of execution and redrive the code (if
needed) by using the same or a different path.
Simultaneous multithreading
Simultaneous multithreading (SMT) is built into the z15 IFLs, zIIPs, and SAPs, which allows
more than one thread to simultaneously run in the same core and share all of its resources.
This function improves the use of the cores and increases processing capacity.
When a program accesses a memory location that is not in the cache, it is called a cache
miss. Because the processor must then wait for the data to be fetched before it can continue
to run, cache misses affect the performance and capacity of the core to run instructions. By
using SMT, when one thread in the core is waiting (such as for data to be fetched from the
next cache levels or from main memory), the second thread in the core can use the shared
resources rather than remain idle.
Adjusted with the growth in the core cache and TLB2, third-generation SMT on z15 improves
thread balancing, supports multiple outstanding translations, optimizes hang avoidance
mechanisms, and delivers improved virtualization performance to benefit Linux. z15 provides
economies of scale with next generation multithreading (SMT) for Linux and zIIP-eligible
workloads while adding support for the I/O System Assist Processor (SAP).
SIMD is designed for parallel computing and can accelerate code that contains integer, string,
character, and floating point data types. This system enables better consolidation of analytics
workloads and business transactions on the Z platform.
Processor RAS
In the unlikely case that a permanent core failure occurs, cores can be individually replaced
by one of the available spares. Core sparing is transparent to the operating system and
applications.
During an outage, a redundant I/O interconnect allows a PCIe+ Gen3 fanout feature to be
concurrently repaired without loss of access to its associated I/O domains. Up to 12 PCIe+
Chapter 1. Introduction 19
fanout slots are available per CPC drawer. The PCIe fanout slots can also be used for the ICA
SR. If redundancy in coupling link connectivity is ensured, the PCIe+ fanout features can be
concurrently repaired.
z15 servers offer PCIe+ I/O drawers that host PCIe features. PCIe I/O drawers and I/O
drawers that were used in previous IBM Z servers are not supported on z15.
With z15 (as with the z14), the number of PCIe resource groups is four, which helps mitigate
the effect of the disruptive Resource Group Microcode Change Level (MCL) installations. This
firmware management approach contributes to the RAS of the server.
During the ordering process of the native PCIe features, features of the same type are evenly
spread across the four resource groups (RG1, RG2, RG3, and RG4) for availability and
serviceability reasons. Resource groups are automatically activated when these features are
present in the CPC.
In addition to the 10GbE RoCE Express features, the following native PCIe I/O features are
also managed by the resource groups:
Coupling Express Long Reach (CE LR)
zHyperLink Express1.1 and zHyperLink Express
RoCE Express2.1, RoCE Express2, and RoCE Express features.
When carried forward on an upgrade, the z15 T01 server also supports the following features
in the PCIe+ I/O drawers:
Storage connectivity:
– FICON Express16S+ Short Wave (SX)
– FICON Express16S+ Long Wave (LX) 10 km (6.2 miles)
– zHyperLink Express
– FICON Express 16S SX
– FICON Express 16S LX
– FICON Express 8S SX
– FICON Express 8S LX
Network connectivity:
– OSA-Express7S 25GbE Short Reach (SR)
– OSA-Express6S 10GbE Long Reach (LR)
– OSA-Express6S 10GbE Short Reach (SR)
– OSA-Express6S GbE LX
– OSA-Express6S GbE SX
– OSA-Express6S 1000BASE-T
– 25GbE RoCE Express2
– 10GbE RoCE Express2
– OSA-Express5S 10 GbE Long Reach (LR)
– OSA-Express5S 10 GbE Short Reach (SR)
– OSA-Express5S GbE LX
– OSA-Express5S GbE SX
– OSA-Express5S 1000BASE-T
– 10GbE RoCE Express
Coupling and Server Time Protocol connectivity: Coupling Express LR
Cryptography:
Chapter 1. Introduction 21
– Crypto-Express6S
– Crypto-Express5S
Although they are used for coupling connectivity, the IBM Integrated Coupling Adapter (ICA
SR) features are not listed here because they are attached directly to the CPC drawer.
Note: The IBM zHyperLink Express1.1 (FC 0451, new build for z15) and IBM zHyperLink
Express (FC 0431, carry forward from z14) are functionally equivalent. Unless specified,
these features are referred to as zHyperLink Express or zHyperLink.
zHyperLink Express feature directly connects the z15 central processor complex (CPC) to the
I/O Bay of the DS8000 series (R8.5 and newer). This short distance of up to 150 m (492 feet)
direct connection is intended to reduce I/O latency and improve storage I/O throughput.
The improved performance of zHyperLink Express allows the z15 PU to make a synchronous
request for the data that is in the DS8880 cache. This feature eliminates the undispatch of the
running request, the queuing delays to resume the request, and the PU cache disruption.
The IBM zHyperLink Express is a two-port feature in the PCIe+ I/O drawer. Up to 16 features
with up to 32 zHyperLink Express ports are supported in a z15 CPC. The zHyperLink Express
feature uses PCIe Gen3 technology, with x16 lanes that are bifurcated into x8 lanes for
storage connectivity. It is designed to support a link data rate of 8 GigaBytes per second
(GBps)12.
The point-to-point link is established by 24x fiber optic cable with Multi-fiber Termination
Push-on (MTP) connectors. For more information, see “zHyperLink Express1.1 (FC 0451)” on
page 179.
FICON channels
Up to 160 features with up to 320 FICON Express16SA channels are supported on a new
build z15. FICON Express 16SA supports 8 or 16 Gbps data link rate (NO 4 Gbps support).
FICON Exptess16S+ and FICON Express 16S (both carry forward only) support link data
rates of 4, 8, or 16 Gbps. FICON Express8S features (carry forward only) support link data
rates of 2, 4, or 8 Gbps.
FICON Express16SA offers the same performance as FICON Express16S+ with its IBM I/O
ASIC that supports up to 3x the I/O start rate of previous FICON/FCP solutions. As with the
FICON Express16S+, both ports of a feature must be defined as the same CHPID type (no
mix of FC and FCP CHPID for the same feature).
12
The link data rates do not represent the performance of the links. The actual performance is dependent upon
many factors, including latency through the adapters, cable lengths, and the type of workload.
OSA-Express features provide important benefits for TCP/IP traffic by reducing latency and
improving throughput for standard and jumbo frames. Data router function that is present in all
OSA-Express features enables performance enhancements.
On z15, an OSA feature that is configured as an integrated console controller CHPID type
(OSC) supports the configuration and enablement of secure connections by using the
Transport Layer Security (TLS) protocol versions 1.0, 1.1, and 1.2.
Chapter 1. Introduction 23
For more information about the OSA features, see 4.6, “Connectivity” on page 167.
HiperSockets
The HiperSockets function (also known as internal queued direct input/output or internal
QDIO or iQDIO) is an integrated function of the z15 server that provides users with
attachments to up to 32 high-speed virtual LANs with minimal system and network processor
usage.
For communications between LPARs in the same z15 server, HiperSockets eliminate the
need to use I/O subsystem features to traverse an external network. Connection to
HiperSockets offers significant value in server consolidation by connecting many virtual
servers.
HiperSockets can also be used for Dynamic cross-system coupling, which is a z/OS
Communications Server feature that creates trusted, internal links to other stacks within a
Parallel Sysplex.
RoCE Express features reduce CPU consumption for applications that use the TCP/IP stack
(sockets communication), such as IBM WebSphere Application Server that accesses a Db2
database. It is transparent to applications and also might help to reduce network latency with
memory-to-memory transfers that use SMC-R in supported z/OS releases and Linux on Z.
IBM Z server generations continue to enhance the RoCE architecture. The 10GbE RoCE
Express feature (carry forward only) supports sharing among 31 LPARs running z/OS or
Linux on Z, while the RoCE Express2 and RoCE Express2.1 (10 GbE and 25 GbE) support
4x the number of LPARs and performance improvements. RoCE Express2 and RoCE
Express2.1 support 63 Virtual Functions (VFs) per port for up to 126 VFs per PCHID
(physical channel ID).
The 10GbE RoCE Express2, 10GbE RoCE Express2.1, and 10GbE RoCE Express features
use SR optics and support the use of a multimode fiber optic cable that ends with an LC
Duplex connector. Both support point-to-point and switched connections with an
enterprise-class 10 GbE switch. A maximum of eight RoCE Express features can be installed
in PCIe+ I/O drawers of z15.
The 25GbE RoCE Express2 and 25GbE RoCE Express2.1 also feature SR optics and
supports the use of 50-micron multimode fiber optic that ends with an LC duplex connector.
These features support point-to-point and switched connections with 25GbE capable switch
(support only for 25 Gbps, no down negotiation to 10 Gbps).
Introduced with z13 GA2 and z13s, SMC-D enables high-bandwidth LPAR-to-LPAR TCP/IP
traffic (sockets communication) by using the direct memory access software protocols over
virtual Internal Shared Memory PCIe devices (vPCIe). SMC-D maintains the socket-API
transparency aspect of SMC-R so that applications that use TCP/IP communications can
benefit immediately without requiring any application software or IP topology changes.
z15 continues to support SMC-D with its lightweight design that improves throughput, latency,
and CPU consumption and complements HiperSockets, OSA, or RoCE without sacrificing
quality of service.
All physical coupling link types can be used to carry STP messages.
The ICA SR1.1 is designed to drive distances up to 150 m and support a link data rate of
8 GBps. The ICA SR1.1 fanout takes one PCIe fanout slot in the z15 CPC drawer. It is used
for coupling connectivity between z15, z14, z13, and z13s CPCs, and cannot be connected to
HCA3-O or HCA3-O LR coupling fanouts. The ICA SR1.1 is compatible with another ICA
SR1.1 or ICA SR only.
Chapter 1. Introduction 25
The ICA SR is designed to drive distances up to 150 m (492 feet) and support a link data rate
of 8 GBps. The ICA SR fanout takes one PCIe I/O fanout slot in the z15 CPC drawer. It is
used for coupling connectivity between z15, z14, z13, and z13s CPCs, and cannot be
connected to HCA3-O or HCA3-O LR coupling fanouts. The ICA SR is compatible with
another ICA SR or ICA SR 1.1 only.
The CE LR link allows for more granularity when scaling up or completing maintenance and
uses Single Mode fiber (similar to InfiniBand 1x coupling links). The CE LR link provides
point-to-point coupling connectivity at distances of 10 km (6.2 miles) unrepeated and 100 km
(62.1 miles) with a qualified dense wavelength division multiplexing (DWDM) device.
CFCC Level 24
CFCC level 24 is delivered on the z15 with driver level 41. CFCC Level 24 introduces the
following enhancements:
CFCC Fair Latch Manager2
Message Path SYID Resiliency Enhancement
Shared-Engine CF Default is changed to “DYNDISP=THIN”
Coupling Facility monopolization avoidance
z15 servers with CFCC Level 24 require z/OS V2R1 or later, and z/VM V6R4 or later for
virtual guest coupling. For more information, see “Coupling Facility Enhancements with
CFCC level 24” on page 118.
Although the CF LPARs are running on different server generations, different levels of CFCC
can coexist in the same sysplex, which enables upgrade from one CFCC level to the next. CF
LPARs that are running on the same server share a single CFCC level.
A CF running on a z15 server (CFCC level 24) can coexist in a sysplex with CFCC levels 23,
22, 21 and 20. For more information about determining the CF LPAR size by using the
CFSizer tool, see the System z Coupling Facility Structure Sizer Tool web page.
STP is a server-wide facility that is implemented in the LIC, which presents a single view of
time to PR/SM. Any IBM Z CPC (including CPCs that are running as stand-alone CFs) can be
enabled for STP by installing the STP feature.
HMC can be configured as an NTP client or an NTP server. To ensure secure connectivity,
HMC NTP broadband authentication can be enabled on z15, z14, and z13 servers.
The time accuracy of an STP-only CTN can be improved by using an NTP server with the
pulse per second (PPS) output signal as ETS. This type of ETS is available from various
vendors that offer network timing solutions.
Attention: A z15 server can coexist in the same (STP-only) CTN with z15, z14, and z13
servers. No older servers are supported in the same CTN with z15.
The z15 supports coupling and timing connectivity by using Integrated Coupling Adapter
Short Reach (ICA SR) and Coupling Express Long Reach (CE LR). No InfiniBand
connectivity is supported on the z15. For STP role playing servers, planning must consider
only ICA SR and CE LR coupling or timing links.
1.3.11 Cryptography
A strong synergy exists between cryptography and security. Cryptography provides the
primitives to support security functions. Similarly, security functions help to ensure authorized
use of key material and cryptographic functions.
Cryptography on IBM Z is built on the platform with integrity. IBM Z platform offers
hardware-based cryptography features that are used by the following environments and
functions:
Java
Db2/IMS encryption tool
Db2 built in encryption z/OS Communication Server
IPsec/IKE/AT-TLS
z/OS System SSL
z/OS
z/OS Encryption Facility
Linux on Z
CP Assist for Cryptographic Functions
Crypto Express7S
Crypto-Express6S
Trusted Key Entry workstation
Chapter 1. Introduction 27
Support for CPACF is available through a group of instructions that are known as the
Message-Security Assist (MSA).
z/OS Integrated Cryptographic Service Facility (ICSF) callable services and the z90crypt
device driver that is running on Linux on Z also start CPACF functions. ICSF is a base
element of z/OS. It uses the available cryptographic functions, CPACF, or PCIe cryptographic
features to balance the workload and help address the bandwidth requirements of your
applications.
With z15, a new Elliptic Curve Cryptography-supporting Modulo Arithmetic unit was
implemented on each PU (core) along with a new message security assist extension 9 and
Elliptic Curve Signature Authentication (ECSA) instruction. This provides hardware support
for verification and signing using NIST P256, P384, P521 curves, Ed25519, Ed448-Goldilocks
curves. The expected use cases include SSL libraries (authentication on the web) and
Blockchain cryptography.
With z14 and carried on to z15, CPACF is enhanced to support pervasive encryption to
provide faster encryption and decryption than previous servers. For every Processor Unit that
is defined as a CP or an IFL, the following benefits are realized:
Reduced overhead on short data (hashing and encryption)
Up to 4x throughput for AES compared to z13
Special instructions for elliptic curve cryptography (ECC)/RSA
New hashing algorithms (for example, SHA-3)
Support for authenticated encryption (combined encryption and hashing; for example,
AES-GCM)
True random number generator (for example, for session keys)
The z13 CPACF provides (supported by z15 and z14 also) the following features:
For data privacy and confidentially: DES, Triple Data Encryption Standard (TDES), and
AES for 128-bit, 192-bit, and 256-bit keys.
For data integrity: Secure Hash Algorithm-1 (SHA-1) 160-bit, and SHA-2 for 224-, 256-,
384-, and 512-bit support. SHA-1 and SHA-2 are included as enabled on all z14s and do
not require the no-charge enablement feature.
For key generation: Pseudo Random Number Generation (PRNG), Random Number
Generation Long (RNGL) (1 - 8192 bytes), and Random Number Generation Long (RNG)
with up to 4096-bit key RSA support for message authentication.
CPACF must be explicitly enabled by using a no-charge enablement feature (FC 3863). This
requirement excludes the SHAs, which are enabled by default with each server.
The enhancements to CPACF are exclusive to the IBM Z servers and are supported by z/OS,
z/VM, z/VSE, z/TPF, and Linux on Z.
Crypto Express7S
Crypto Express7S represents the newest generation of cryptographic features. Cryptographic
performance improvements with new Crypto Express7S, which is available with two features,
with one (FC 0899) or two (FC 0898) cryptographic co-processors14 that allow more data to
be securely transferred across the internet. Crypto Express7S is designed to complement the
cryptographic capabilities of the CPACF. It is an optional feature of the z15 server generation.
14 The IBM PCIe Cryptographic Co-processor (PCIeCC) is implemented as a Hardware Security Module (HSM).
z15 T01 servers allow sharing of a cryptographic coprocessor across 85 domains (the
maximum number of LPARs on the system for z15 T01 is 85).
Crypto Express7S provides higher density of the PCIe cryptographic features with enhanced
PCIe Cryptographic Coprocessor (IBM 4769) and carries forward the functionality of the
Crypto Express6S, described in the following paragraph.
Crypto-Express6S
Crypto-Express6S (FC #0893) introduced with z14 server generation can be carried forward
to z15. For availability reasons, a minimum of two features is required.
z15 servers allow sharing of a cryptographic coprocessor across 85 domains (the maximum
number of LPARs on the system for z15 is 85).
Federal Information Processing Standard (FIPS) 140-2 certification is supported only when
Crypto-Express6S is configured as a CCA or an EP11 coprocessor.
Crypto-Express6S supports several ciphers and standards that are described next. For more
information about cryptographic algorithms and standards, see Chapter 6, “Cryptographic
features” on page 215.
15 Federal Information Processing Standard (FIPS) 140-2 Security Requirements for Cryptographic Modules
Chapter 1. Introduction 29
Trusted Key Entry workstation
The Trusted Key Entry (TKE) feature is an integrated solution that is composed of workstation
firmware, hardware, and software to manage cryptographic keys in a secure environment.
The TKE is network-connected or isolated, in which case smart cards are used.
The TKE workstation offers a security-rich solution for basic local and remote key
management. It provides authorized personnel with a method for key identification, exchange,
separation, update, and backup, and a secure hardware-based key loading mechanism for
operational and master keys. TKE also provides secure management of host cryptographic
module and host capabilities.
Support for an optional smart card reader that is attached to the TKE workstation allows the
use of smart cards that contain an embedded microprocessor and associated memory for
data storage. Access to and the use of confidential data on the smart cards are protected by
a user-defined personal identification number (PIN).
TKE workstation and the most recent TKE 9.2 LIC are optional features on the z15. TKE
workstation is offered in two types: TKE Tower (FC 0088) and TKE Rack Mount (FC 0087).
TKE 9.x16 requires the crypto-adapter FC 4769. You can use an older TKE version to collect
data from previous generations of cryptographic modules and apply the data to
Crypto-Express7S and Crypto-Express6S coprocessors.
TKE 9.x is required if you choose to use the TKE to manage a Crypto-Express7S. TKE 9.1
and later is also required to manage the new CCA mode PCI-HSM settings that are available
on the Crypto-Express6S and Crypto-Express7S. A TKE is required to manage any
Crypto-Express feature that is running in IBM Enterprise PKCS #11 (EP11) mode. If EP11 is
to be defined, smart cards that are used require FIPS certification.
For more information about the cryptographic features, see Chapter 6, “Cryptographic
features” on page 215.
For more information about the most current ICSF updates that are available, see the Web
Deliverables download web page.
The initial focus is on preventing failures from occurring. This goal is accomplished by using
Hi-Rel (highest reliability) components that use screening, sorting, burn-in, and run-in, and by
taking advantage of technology integration.
For LIC and hardware design, failures are reduced through rigorous design rules; design
walk-through; peer reviews; element, subsystem, and system simulation; and extensive
engineering and manufacturing testing.
The RAS strategy is focused on a recovery design to mask errors and make them transparent
to client operations. An extensive hardware recovery design is implemented to detect and
16
TKE 9.0 LIC, TKE 9.1 LIC, and TKE 9.2 LIC have the same hardware requirements. TKE 9.0 LIC and 9.1 LIC can
be upgraded to TKE 9.2 LIC.
System Recovery Boost (see also Appendix B, “System Recovery Boost” on page 493) is a
new function that is implemented in the z15 firmware that brings a new dimension to the
overall RAS approach. System Recovery Boost is designed to provide more CP capacity to
LPARs running on a z15 CPC, capacity that is used for boosting (speeding up) operations
during maintenance periods, such as system IPL, system shutdown, and stand-alone dumps.
For more information, see Chapter 9, “Reliability, availability, and serviceability” on page 383.
HMC is offered as a Tower (FC 0062) and a Rack Mount (FC 0063) feature. Rack Mount HMC
can be placed in a customer-supplied 19-inch rack and occupies 1U rack space. z15 includes
Driver level 41 and HMC application Version 2.15.0.
IBM z15 also introduces a new management feature, the IBM Hardware Management
Appliance (FC 0100). With this feature, the need for a stand-alone HMC is eliminated, both
Hardware Management Console appliance and Support Element Appliance running
virtualized on the Support Element hardware servers integrated with the z15 CPC.
For more information, see Chapter 10, “Hardware Management Console and Support
Element” on page 411.
Chapter 1. Introduction 31
z/VSE: Version 6 Release 2
z/TPF Version 1 Release 1
Linux on IBM Z distributions18:
– SUSE SLES 15 SP1 with service, SUSE SLES 12 SP4 with service, and SUSE SLES
11 SP4 with service.
– Red Hat RHEL 8.0 with service, Red Hat RHEL 7.7 with service, and Red Hat RHEL
6.10 with service.
– Ubuntu 18.04.1 LTS with service and Ubuntu 16.04.5 LTS with service.
KVM: Supported by Linux on IBM Z distributions
For more information about supported Linux on Z distribution levels, see the Tested platforms
for Linux page of the IBM Z website.
For more information about features and functions that are supported on z14 by operating
system, see Chapter 7, “Operating system support” on page 253.
z/VM support
z/VM 7.2 (Announced April 14, 2020, planned availability Sept. 2020) brings, in addition to
features and functions included with z/VM 7.1 and subsequent 7.1 PTFs, the following new
functionality:
Centralized Service Management for non-SSI environments to deploy service to multiple
systems, regardless of geographic location, from a centralized primary location.
Multiple Subchannel Set (MSS) Multi-Target Peer-To-Peer Remote Copy (MT-PPRC) z/VM
support for the GDPS environment, allowing a device to be the primary to up to three
secondary devices, each defined in a separate alternate subchannel set (supporting up to
3 alternate subchannel sets). Also provides the CP updates necessary for VM/HCD
support of alternate subchannel sets.
New Architecture Level Set of z13, z13s, or newer processor families
z/VM 7.1 (available as of September 2018) increases the level of engagement with the z/VM
user community. z/VM 7.1 includes the following new features:
Single System Image and Live Guest Relocation included in the base (no extra charge).
Enhances the dump process to reduce the time that is required to create and process
dumps.
Upgrades to a new Architecture Level Set (requires an IBM zEnterprise EC12 or BC12, or
later).
Provides the base for more functionality to be delivered as service after general
availability.
Enhances the dynamic configuration capabilities of a running z/VM system with Dynamic
Memory Downgrade* support. For more information, see this web page.
Includes SPE19s shipped for z/VM 6.4, including Virtual Switch Enhanced Load Balancing,
DS8000 z-Thin Provisioning, and Encrypted Paging.
To support new functionality that was announced October 2018, z/VM requires fixes for the
following APARs:
PI99085
VM66130
18
Customers should monitor for new distribution releases supported.
19 Small Program Enhancements, part of the continuous delivery model, see http://www.vm.ibm.com/newfunction/
Support for z15 is provided in z/VM 7.1 and 6.4 with fixes for the following APARs:
VM66248: z15 processor compatibility support (including Crypto-Express7S)
PI99085: TCP/IP Stack support for OSA-Express7S
VM66239: VMHCD support
VM65598: VMHCM support
PH00902: HLASM support
VM66240: IOCP support.
For more information about the features and functions that are supported on z14 by operating
system, see Chapter 7, “Operating system support” on page 253.
z/OS support
z/OS uses many of the following new functions and features of z14 (depending on version
and release; PTFs might be required to support new functions):
Up to 190 processors per LPAR or up to 128 physical processors per LPAR in SMT mode
(SMT for zIIP)
Up to 16 TB of real memory per LPAR (4 TB maximum for z/OS)
Two-way simultaneous multithreading (SMT) optimization and support of SAPs (SAP SMT
enabled by default) in addition to zIIP engines
XL C/C++ ARCH(13) and TUNE(13) complier options
Use of faster CPACF
Pervasive Encryption:
– Coupling Facility Encryption
– Dataset and network encryption
HiperDispatch Enhancements
On-chip compression accelerator (transparent zEDC replacement)
z15 Hardware Instrumentation Services (HIS)
Entropy-Encoding Compression Enhancements
Guarded Storage Facility (GSF)
Instruction Execution Protection (IEP)
IBM Virtual Flash Memory (VFM)
Improved memory management in Real Storage Manager (RSM)
CF use of VFM: CFCC Level 24
Coupling Express Long Reach (CE LR) CHPID type CL5
zHyperLink Express1.1
FICON Express16SA; OSA Express7S
RoCE-Express2.1 (25GbE and 10GbE)
Cryptography:
– Crypto-Express7S:
• Next Generation Coprocessor support
• Support for Coprocessor in PCI-HSM Compliance Mode
Chapter 1. Introduction 33
• Designed for up to 85 domains
– TKE 9.2 workstation
For more information about the features and functions that are supported on z15 by operating
system, see Chapter 7, “Operating system support” on page 253.
The compilers increase the return on your investment in IBM Z hardware by maximizing
application performance by using the compilers’ advanced optimization technology for
z/Architecture. Through their support of web services, XML, and Java, they allow for the
modernization of assets in web-based applications. They also support the latest IBM
middleware products (CICS, Db2, and IMS), which allows applications to use their latest
capabilities.
To fully use the capabilities of z15 servers, you must compile it by using the minimum level of
each compiler. To obtain the best performance, you must specify an architecture level of 13 by
using the ARCH(13) option.
For more information, see 7.5.4, “z/OS XL C/C++ considerations” on page 324.
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, Features Max34, Max71, Max108,
Max145, and Max190, which is further identified as IBM z15 Model T01, or z15 T01,
unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, Features Max4, Max13, Max21, Max32,
and Max65, which is further identified as IBM z15 Model T02, or z15 T02, unless
otherwise specified.
In the remainder of this chapter, IBM z15 (z15) refers to both machine types.
The redesigned CPC drawer and I/O infrastructure also lowers power consumption, reduces
the footprint, and allows installation in virtually any data center. The z15 server is rated for
ASHRAE class A31 data center operating environment.
The z15 server differentiates itself from previous Z server generations through the following
significant changes to the modular hardware:
All external cabling (power, I/O, and management) is performed at the rear of the system
Flexible configurations: Frame quantity is determined by the system configuration (1 - 4
frames)
Choice of power Intelligent Power Distribution Unit (iPDU or PDU) or Bulk Power Assembly
(BPA)
Feature codes that reserve slots for plan-ahead CPC drawers
Added internal water cooling plumbing for systems with more than three CPC drawers
New PCIe+ Gen3 I/O drawers (19-inch format) supporting 16 PCIe adapters
The power options include PDU-based power or BPA-based power. The z15 T01 server can
be configured as a radiator (air) cooled or water cooled (that uses data center chilled water
supply) system. Only BPA-based power system can be (optionally) configured for water
cooling and with Internal Battery Feature (IBF), while a radiator (air) cooled system has
PDU-based power (no Internal Battery Feature [IBF] available for PDU-based systems).
The z15 T01 includes the following basic hardware building blocks:
19-inch 42u frame (1 - 4)
CPC (Processor) drawers (1 - 5)
PCIe+ Gen3 I/O drawers (up to 12)
CPC drawer Cooling Units: Radiator cooling assembly (RCA) or Water Cooling Unit
(WCU)
Power, with choice of:
– Intelligent Power Distribution Units (iPDU) pairs (2 - 4 per frame, depending on the
configuration).
– Bulk Power Regulators (1 - 6 pairs, depending on the configuration)
Support Elements (two):
– Single KMM2 device (USB-C connection)
– Optional extra hardware for IBM Hardware Management Appliance feature
24-port 1GbE Switches (two or four, depending on the system configuration)
Hardware for cable management at the rear of the system
1
For more information, see Chapter 2, Environmental specifications in IBM Z 8561Installation Manual for Physical
Planning, GC28-7002.
2 KMM - Keyboard, Mouse, Monitor
An example of a BPA-based powered system with IBF, and a maximum of five CPC drawers
and 11 PCIe+ I/O drawers is shown in Figure 2-2.
The key features that are used to build the system are listed in Table 2-1 on page 38. For
more information about the various configurations, see Appendix D, “Frame configurations”
on page 515.
2271 CPC1 reserve Reserve A15 location for future add CPC1
(Max34 to Max71 upgrade)
2272 CPC2 reserve Reserve A20 location for future add CPC2
(Max71 to Max108 upgrade)
PDU power
0630 380-415V 60A 3 Phase (Wye- Worldwide (except North America and Japan)
”Y”)
BPA power
3016 Bulk Power Regulator (BPR) Quantity per BPA feature: Min. 2, max 6 (in
pairs)
3217 Internal Battery Feature (IBF) Quantity per BPA feature: Min. 2, max 6 (in
pairs)
I/O
7928 Top Exit Cabling without Tophat Uses rear slide plates at top of frame
The front doors of the z15 T01 for systems with 1 - 4 frames also is shown in Figure 2-3.
The Top Exit feature code (FC 7917) provides an optional Top Exit cover enclosure. The
optional Top Exit cover enclosure provides cable retention hardware and mounting locations
to secure Fiber Quick Connector MPO3 brackets on the top of the frames.
All external cabling enters the system at the rear of the frames for all I/O adapters,
management LAN, and power connections.
Feature code 7917 provides a top hat assembly to be installed at the top of each frame in the
configuration. This assembly is designed to assist with fiber trunking management.
3 MPO - Multi-fiber Push On connector
A view of the top rear of the frame and the openings for top exit cables and power is shown in
Figure 2-4. When FC 7917 is installed, the plated adjustable shields are removed and the top
exit enclosure is installed.
A newly designed vertical cable management guide (“spine”) can assist with proper cable
management for fiber, copper, and coupling cables. Depending on the configuration, a spine
is present from manufacturing with cable organizer clips installed.
The cable retention clips can be relocated for best usage. All external cabling to the system
(from top or bottom) can use the spines to minimize interference with the PDUs that are
mounted on the sides of the rack.
The z15 T01 can be configured with 1 - 5 CPC drawers (three in the A frame and two in the B
frame). A CPC drawer and its components are shown in Figure 2-6 on page 42.
The z15 Model T01 5u CPC drawer always contains four Processor Unit (PU) SCMs, one
System Controller (SC) SCM, and up to 20 memory DIMMs.
Depending on the feature, the z15 T01 contains the following CPC components:
The number of CPC drawers installed is driven by the following feature codes:
– FC 0655: One CPC drawer, Max34, up to 34 characterizable PUs
– FC 0656: Two CPC drawers, Max71, up to 71 characterizable PUs
– FC 0657: Three CPC drawers, Max108, up to 108 characterizable PUs
– FC 0658: Four CPC drawers, Max145, up to 145 characterizable PUs
– FC 0659: Five CPC drawers, Max190, up to 190 characterizable PUs
The following SCMs are used:
– PU SCM uses 14nm SOI technology, 17 layers of metal, 9.2 billion transistors, core
running at 5.2GHz: (with 12 cores design per PU SCM).
– SC SCM, 17 layers of metal, 12.2 billion transistors, 960 MB shared eDRAM L4 cache.
Memory plugging:
– Four memory controllers per drawer (one per PU SCM)
– Each memory controller supports five DIMM slots
– Four or three memory controllers per drawer are populated (up to 20 DIMMs)
– Different memory controllers can have different size DIMMs
Up to 12 PCIe+ Gen3 fanout slots that can host:
– 2-Port PCIe+ Gen3 I/O fanout for PCIe+ I/O drawers (ordered and used in pairs for
availability)
– ICA SR and ICA SR1.1 PCIe fanout for coupling (two ports per feature)
Management elements: Two dual function flexible service processor (FSP) and oscillator
cards (OSC) for system control and to provide system clock (N+1 redundancy).
The front view of the CPC drawer, which includes the cooling fans, FSP/OSC and bulk
(power) distribution cards (BDC), is shown in Figure 2-7.
PPS ports
Figure 2-7 Front view of the CPC drawer
The rear view of a fully populated CPC Drawer is shown in Figure 2-8 on page 44. Dual port
I/O fanouts and ICA SR adapters are plugged in specific slots for best performance and
availability. Redundant power supplies and four SMP ports also are shown.
The CPC drawer logical structure, component connections (including the PU SCMs), and the
storage control SCMs are shown in Figure 2-9.
Memory is connected to the SCMs through memory control units (MCUs). Up to four MCUs
are available in a CPC drawer (one per PU SCM) and provide the interface to the DIMM
controller. A memory controller uses five DIMM slots.
CPC2
CPC4
CPC1
CPC0
CPC3
The CPC drawers that are installed in Frame A and Frame B are populated from bottom to
top.
For simplification, the z15 (2.15.0) removes the Support Element “Sysplex/System Timer”
task panels. HMC level 2.15.0 (Driver 41) is required to manage system time for z15.
The accuracy of an STP-only CTN is improved by using an NTP or PTP server with the PPS
output signal as the External Time Source (ETS). Devices with PPS output are available from
several vendors that offer network timing solutions.
Tip: STP is available as FC 1021. It is implemented in the Licensed Internal Code (LIC),
and allows multiple servers to maintain time synchronization with each other and
synchronization to an ETS.
For more information, see the Redbook: IBM Server Time Protocol Guide, SG24-8480.
With z15, the CPC drawer FSP card is combined with the Oscillator card in a single Field
Replaceable Unit (FRU). Two combined FSP/OSC cards are used per CPC drawer.
Also, the PCIe+ I/O drawer has a new FSP. Each FSP card has one Ethernet port that
connects to the internal Ethernet LANs through the internal network switches (SW1, SW2,
and SW3, SW4, if configured). The FSPs communicate with the SEs and provide a
subsystem interface (SSI) for controlling components.
Note: The maximum z15 T01 system configuration features four GbE switches, five CPC
drawers, and up to 12 PCIe I/O drawers.
A typical FSP operation is to control a power supply. An SE sends a command to the FSP to
start the power supply. The FSP cycles the various components of the power supply, monitors
the success of each step and the resulting voltages, and reports this status to the SE.
Most SEs are duplexed (N+1), and each element has at least one FSP. Two internal Ethernet
LANs and two SEs, for redundancy, and crossover capability between the LANs, are available
so that both SEs can operate on both LANs.
The Hardware Management Consoles (HMCs) and SEs are connected directly to one or two
Ethernet Customer LANs. One or more HMCs can be used.
The two types of SCMs (PU and SC) are shown in Figure 2-13. For both SCMs, a thermal cap
is placed over the chip. Each PU SCM is water cooled by way of a cold plate manifold
assembly. The SC SCM is air cooled by using CPC drawer fans and heat sink.
The SCMs are plugged into a socket that is part of the CPC drawer packaging. The
interconnectivity between the CPC drawers is accomplished through SMP connectors and
cables. Four inter-drawer connections are available on each CPC drawer. This configuration
allows a multidrawer system to act as an SMP system.
The z15 PU chip (installed as a PU SCM) is an evolution of the z14 chip design. It includes
the following features and improvements:
CMOS 14nm SOI technology
12 core design (versus 10 for z14) with increased on-chip cache sizes
Three PCIe Gen4 interfaces (GX bus was dropped)
DDR4 memory controller
Two X-buses support cluster connectivity (PU SCM-to-PU SCM and PU SCM-to-SC SCM
connectivity by way of X bus).
New EDRAM macro design with 2x macro density. Compared to z14 PU:
– L3 was increased from 128 MB to 256 MB per chip
– L2-I was increased from 2 MB to 4 MB per core
– L2-L3 protocol was changed to reduce latency
On-chip compression accelerator (Nest Acceleration Unit - NXU)
Further optimization of the nest-core staging
2.3.3 PU characterization
The PUs are characterized for client use. The characterized PUs can be used in general to
run supported operating systems, such as z/OS, z/VM, and Linux on Z. They also can run
specific workloads, such as Java, XML services, IPSec, and some Db2 workloads, or
clustering functions, such as the Coupling Facility Control Code (CFCC).
The maximum number of characterizable PUs depends on the z15 CPC drawer feature code.
Some PUs are characterized for system use; some are characterized for client workload use.
By default, one spare PU is available to assume the function of a failed PU. The maximum
number of PUs that can be characterized for client use are listed in Table 2-3.
The rule for the CP to zIIP purchase ratio is that for every CP purchased, up to two zIIPs can
be purchased. Java and XML workloads can run on zIIPs.
However, an LPAR definition can go beyond the 1:2 ratio. For example, a maximum of four
physical zIIPs can be installed on a system with two physical CPs.
Converting a PU from one type to any other type is possible by using the Dynamic Processor
Unit Reassignment process. These conversions occur concurrently with the system
operation.
Note: The addition of ICFs, IFLs, zIIPs, and SAP to the z15 does not change the system
capacity setting or its million service units (MSU) rating.
A schematic representation of the SC chip is shown in Figure 2-16. Consider the following
points:
A Bus (SC-SC off drawer): Minor changes to reflect protocol improvements and new
system topology
960 MB shared eDRAM L4 Cache
L4 Directory is built with eDRAM
New L4 Cache Management: Ratio of L3 to L4 cache capacity is increasing
The following configuration examples and how the configurations are different with the power
selection, number of CPC drawers and I/O features ordered, the layout of the PCIe+ I/O
drawers, and PCHID numbering are shown in Figure 2-19:
The first, a single frame system, ordered with PDU power, radiator cooling, one CPC
drawer, and greater than 32 I/O features to drive three I/O drawers.
PCHID numbering is consecutive from top to bottom.
The second, a three frame system, ordered with PDU power, radiator cooling, four CPC
drawers, and greater than 128 I/O features to drive nine I/O drawers.
PCHID numbering starts in the A-frame, resumes in the B-frame from top down, and
continues to the Z-frame working from the bottom up.
The third, a two frame system, ordered with BPA, IBF power options, radiator cooling, one
CPC drawer, two reserved CPC drawer slots for future CPC drawer add MES, and greater
than 64 I/O features to drive five I/O drawers
PCHID numbering starts in the Z-frame, from the bottom and working up.
The vertical card collapsed horizontal and the awareness of the port and PCHID layout where
the top of the adapter (port D1) is closest to the location panel on both sides of the drawer are
shown in Figure 2-20.
Figure 2-20 I/O feature orientation in PCIe I/O drawer (rear view)
Note: The CHPID Mapping Tool (available on ResourceLink) can be used to print a CHPID
Report that displays the drawer and PCHID/CHPID layout.
2.5 Memory
The maximum physical memory size is directly related to the number of CPC drawers in the
system. Each CPC drawer can contain up to 8 TB of customer memory, for a total of 40 TB of
memory per system.
The minimum and maximum memory sizes that you can order for each z15 feature are listed
in Table 2-4.
Note: The Plan Ahead Memory feature is not offered with a new order z15 system. The
Plan Ahead Memory feature that is available on z13 or z14 can be carried forward to z15.
The memory granularity, which is based on the installed customer memory, is listed in
Table 2-5.
64 512 - 768
The CPC drawer memory topology of a z15 server is shown in Figure 2-21.
The RAIM design requires the addition of one memory channel that is dedicated for reliability,
availability, and serviceability (RAS).
The fifth channel in each MCU enables memory to be implemented as a RAIM. This
technology features significant error detection and correction capabilities. Bit, lane, DRAM,
DIMM, socket, and complete memory channel failures can be detected and corrected,
including many types of multiple failures. Therefore, RAIM takes 20% of DIMM capacity (a
non-RAIM option is not available).
DIMM plugging for the configurations in each CPC drawer do not have to be similar. Each
memory 5 slot DIMM bank must have the same DIMM size; however, a drawer can have a mix
of DIMM banks. Table 2-7 lists the memory population by DIMM bank for the 15
configurations that are listed in Table 2-6 on page 59.
As an example, for configuration #14, memory positions MD06-MD10 are populated with five
128 GB DIMMS.
The support element View Hardware Configuration task can be used to determine the size
and quantity of the memory plugged in each drawer. Figure 2-23 shows an example of
configuration number 16 from the previous tables, and displays the location and description of
the installed memory modules.
Figure 2-24 shows the CPC drawer and DIMM locations for a z15.
If more storage is ordered by using other feature codes, such as Virtual Flash Memory, or
Flexible Memory, the extra storage is installed and plugged as necessary.
For example, a customer orders FC 1528 that features 1920 GB memory and Max145 (4
CPC drawers). The drawer configurations include the following components:
CPC0 (768GB), CPC3 (768GB) - Configuration #11 (from Table 2-6 on page 59)
CPC1 (384GB), CPC2 (384GB) - Configuration #19
Total 768 + 768 + 384 + 384 - 256 HSA= 2048GB (Dial Max)
Increments
Increments
Memory
Customer
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Drawer CPC4
Max Max Max Max Max
Increments
Increments
Memory
Customer
Dial Dial Dial Dial Dial
Drawer CPC0
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Drawer CPC4
Max Max Max Max Max
Increments
Increments
Memory
Customer
Dial Dial Dial Dial Dial
Drawer CPC0
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Drawer CPC4
Max Max Max Max Max
For a model upgrade that results in the addition of a CPC drawer, the minimum memory
increment is added to the system. Each CPC drawer has a minimum physical memory size of
480 GB.
During a model upgrade, adding a CPC drawer is a concurrent operation. Adding physical
memory to the added drawer is also concurrent. If all or part of the added memory is enabled
for use, it might become available to an active LPAR if the partition includes defined reserved
storage. (For more information, see 3.7.3, “Reserved storage” on page 141.) Alternatively, the
added memory can be used by a defined LPAR that is activated after the memory is added.
Note: Memory downgrades within a z15 are not supported. Feature downgrades (removal
of a CPC quantity feature) are not supported.
Removing a CPC drawer often results in removing active memory. With the flexible memory
option, removing the affected memory and reallocating its use elsewhere in the system is
possible. For more information, see 2.5.7, “Flexible Memory Option”. This process requires
more available memory to compensate for the memory that is lost with the removal of the
drawer.
No application changes are required to change from IBM Flash Express to VFM. Consider the
following points:
Dialed memory + VFM = total hardware plugged
Dialed memory + VFM + Flex memory option = total hardware plugged
VFM is offered in 0.5 TB increment size; VFM for z15 is FC 0643 - Min=0, Max=12
VFM is designed to help improve availability and handling of paging workload spikes when
z/OS V2.1, V2.2, V2.3, or V2.4 is run. With this support, z/OS is designed to help improve
system availability and responsiveness by using VFM across transitional workload events,
such as market openings and diagnostic data collection. z/OS is also designed to help
improve processor performance by supporting middleware use of pageable large (1 MB)
pages.
VFM can also be used by coupling facility images to provide extended capacity and
availability for workloads that are use IBM WebSphere MQ Shared Queues structures. The
use of VFM can improve availability by reducing latency from paging delays that can occur at
the start of the workday or during other transitional periods. It is also designed to help
eliminate delays that can occur when collecting diagnostic data.
VFM can help organizations meet their most demanding service level agreements and
compete more effectively. VFM is easy to configure in the LPAR Image Profile and provides
rapid time to value.
When you order memory, you can request extra flexible memory. The extra physical memory,
if required, is calculated by the configuration and priced accordingly.
The flexible memory sizes that are available for the z15 T01 are listed in Table 2-9.
The IBM Z hardware has decades of intense engineering behind it, which results in a robust
and reliable platform. The hardware has many RAS features that are built into it. For more
information, see Chapter 9, “Reliability, availability, and serviceability” on page 383.
DIMM level failures, including components, such as the memory controller application-specific
integrated circuit (ASIC), power regulators, clocks, and system board, can be corrected.
Memory channel failures, such as signal lines, control lines, and drivers and receivers on the
MCM, can be corrected.
Upstream and downstream data signals can be spared by using two spare wires on the
upstream and downstream paths. One of these signals can be used to spare a clock signal
line (one upstream and one downstream). The following improvements were also
implemented in the z15 server:
No cascading of memory DIMMs
Independent channel recovery
Double tabs for clock lanes
Separate replay buffer per channel
The internal intelligent Power Distribution Unit (iPDU, or PDU) provide the following
capabilities:
Individual outlet control by way of Ethernet:
– Provide a System Reset capability
– Power cycle an SE if a hang occurs
– Verify a power cable at installation
System Reset Function:
– No EPO switch is available on the z15. This function provides a means to put a server
into a known state similar to past total power reset.
7 Voltage Regulator Module stick converts the DC bulk power that is delivered by the PSUs (12V) into localized low
voltage that is used by the installed components (for example, PU SCMs, SC SCM, memory DIMMs, and other
circuitry).
8 The air density sensor measures air pressure and is used to control blower speed.
The power service and control network (PSCN) is used to control and monitor the elements in
the system and include the following components:
Ethernet Top of Rack (TOR) switches provide the internal PSCN connectivity:
– Switches are redundant (N+1)
– Concurrently maintainable
– Each switch has one integrated power supply
– FSPs are cross wired to the Ethernet switches
Redundant SEs
Each SE has two power supplies (N+1) and input power is cross-coupled from the PDUs.
Concurrent CPC upgrades
CPC1 to (CPC1 + CPC2) and (CPC1 + CPC2) to (CPC1+CPC2+CPC3) - provided CPC1
Reserve and/or CPC2 Reserve features are part of the initial system order (FC 2271
and/or FC 2272)
All PCIe+ I/O drawer MESs are concurrent
All LICC model changes are concurrent
IBM z15 servers continue to deliver robust server designs through new technologies,
hardening both new and classic redundancy. For more information, see Chapter 9,
“Reliability, availability, and serviceability” on page 383.
9 For z/OS, boost period is 60 minutes for IPL and 30 minutes for shutdown
Figure 2-25 shows the location of the fanout slots. Each slot is identified with a location code
(label) of LGxx.
A fanout can be repaired concurrently with the use of redundant I/O interconnect. For more
information, see 2.7.1, “Redundant I/O interconnect” on page 71.
When configured for availability, the channels and coupling links are balanced across CPC
drawers. In a system that is configured for maximum availability, alternative paths maintain
access to critical I/O devices, such as disks and networks. The CHPID Mapping Tool can be
used to assist with configuring a system for high availability.
Enhanced (CPC) drawer availability (EDA) allows a single CPC drawer in a multidrawer CPC
to be removed and reinstalled (serviced) concurrently for an upgrade or a repair. Removing a
CPC drawer means that the connectivity to the I/O devices that are connected to that CPC
drawer is lost. To prevent connectivity loss, the redundant I/O interconnect feature allows you
to maintain connection to critical I/O devices (except for ICA SR1.1) when a CPC drawer is
removed.
The PCIe+ I/O drawer supports up to 16 PCIe features, which are organized in two hardware
domains (for each drawer). The infrastructure for the fanout to I/O drawers and external
coupling is shown in Figure 2-26.
Figure 2-26 Infrastructure for PCIe+ I/O drawer (system with two PCIe+ I/O drawers)
The new PCIe+ Gen3 fanout cards are used to provide the connection from the PU SCM
PCIe Bridge Unit (PBU), which uses split the PCIe Gen4 (@32GBps) processor busses into
two PCIe Gen3 x16 (@16 GBps) interfaces to the PCIe switch card in the PCIe+ I/O drawer.
The PCIe switch card spreads the x16 PCIe bus to the PCIe I/O slots in the domain.
In the PCIe+ I/O drawer, the two PCIe switch cards (LG06 and LG16, see Figure 2-27)
provide a backup path (Redundant I/O Interconnect - RII) for each other through the passive
connection in the PCIe+ I/O drawer backplane.
During a PCIe fanout or cable failure, all 16 PCIe cards in the two domains can be driven
through a single PCIe switch card (see Figure 2-28).
Note: The PCIe Interconnect (switch) adapter must be installed in the PCIe+ I/O drawer to
maintain the interconnect across I/O domains. If the adapter is removed (for a service
operation), the I/O cards in that domain (up to eight) become unavailable.
Before removing the CPC drawer, the contents of the PUs and memory of the drawer must be
relocated. PUs must be available on the remaining CPC drawers to replace the deactivated
drawer. Also, sufficient redundant memory must be available if no degradation of applications
is allowed. To ensure that the CPC configuration supports removal of a CPC drawer with
minimal effect on the workload, consider the flexible memory option. Any CPC drawer can be
replaced, including the first CPC drawer that initially contains the HSA.
Removal of a CPC drawer also removes the CPC drawer connectivity to the I/O drawers,
PCIe I/O drawers, and coupling links. The effect of the removal of the CPC drawer on the
system is limited by the use of redundant I/O interconnect. (For more information, see 2.7.1,
“Redundant I/O interconnect” on page 71.) However, all ICA SR1.1 links that are installed in
the removed CPC drawer must be configured offline.
If the enhanced drawer availability and flexible memory options are not used when a CPC
drawer must be replaced, the memory in the failing drawer is also removed. This process
might be necessary during an upgrade or a repair action. Until the removed CPC drawer is
replaced, a power-on reset of the system with the remaining CPC drawers is supported. The
CPC drawer can then be replaced and added back into the configuration concurrently.
A minimum of one PU that is characterized as a CP, IFL, or ICF is required per system. The
maximum number of characterizable PUs is 190. The maximum number of zIIPs is up to twice
the number of PUs that are characterized as CPs.
The following components are present in the z15 server, but they are not part of the PUs that
clients purchase and require no characterization:
SAP to be used by the channel subsystem. The number of predefined SAPs depends on
the z15 model.
One IFP, which is used in the support of designated features and functions, such as RoCE
(all features), Coupling Express LR, Internal Shared Memory (ISM) SMC-D, and other
management functions.
Two spare PUs, which can transparently assume any characterization during a permanent
failure of another PU.
The z15 uses features to define the number of PUs that are available for client use in each
configuration. The models are listed in Table 2-10.
drawer
PUs per
SAPs
Opt
SAPs
Base
Spares
CPs IFLs ICFs uIFLs
Not all PUs available on a model are required to be characterized with a feature code.
Only the PUs purchased by a customer are identified with a feature code.
zIIP maximum quantity for new build systems follows the 2:1 ratio. It might be greater if
present during MES upgrades.
All PU conversions can be performed concurrently.
A capacity marker identifies the number of CPs that were purchased. This number of
purchased CPs is higher than or equal to the number of CPs that is actively used. The
capacity marker marks the availability of purchased but unused capacity that is intended to be
used as CPs in the future. This status often is present for software-charging reasons.
2.8.1 Upgrades
Concurrent upgrades of CPs, IFLs, ICFs, zIIPs, or SAPs are available for the z15 server.
However, concurrent PU upgrades require that more PUs are installed but not activated.
Spare PUs are used to replace defective PUs. Two spare PUs always are on a z15 T01
server. In the rare event of a PU failure, a spare PU is activated concurrently and
transparently and is assigned the characteristics of the failing PU.
The following upgrade paths for the z15 are shown in Figure 2-29:
Note: The z15 CPC cannot be member in an Ensemble managed by Unified Resource
Manager.
The following distinct model capacity identifier ranges are recognized (one for full capacity
and three for granular capacity):
For full-capacity engines, model capacity identifiers 701 - 7J0 are used. They express
capacity settings for 1 - 190 characterized CPs.
Three model capacity identifier ranges offer a unique level of granular sub-capacity
engines at the low end. They are available when no more than 34 CPs are characterized.
These three subcapacity settings are applied to up to 34 CPs, which combined offer 102
more capacity settings. For more information, see “Granular capacity”.
Granular capacity
The z15 T01 server offers 102 capacity settings (granular capacity) for up to 34 CPs.. When
subcapacity settings are used, other PUs beyond 34 can be characterized only as specialty
engines. For models with more that 34 CPs, all CPs are running at full capacity (7xx).
The three defined ranges of subcapacity settings have model capacity identifiers numbered
401- 434, 501 - 534, and 601 - 634.
Consideration: All CPs have the same capacity identifier. Specialty engines (IFLs, zIIPs,
and ICFs) operate at full speed.
Max34 701 - 734, 601 - 634, 501 - 534, and 401 - 434
Max71 701 - 771, 601 - 634, 501 - 534, and 401 - 434
Max108 701 - 7A8, 601 - 634, 501 - 534, and 401 - 434
Max145 701 - 7E5, 601 - 634, 501 - 534, and 401 - 434
Max190 701 - 7J0, 601 - 634, 501 - 534, and 401 - 434
For more information about temporary capacity increases, see Chapter 8, “System upgrades”
on page 333.
The PDUs are controlled by using an Ethernet port and support the following input:
3-phase 200 - 240 V AC (wired as “Delta”)
3-phase 380 - 415 V AC (wired as “Wye”)
The power supply units convert the AC power to DC power that is used as input for the Points
of Load (POLs) in the CPC drawer and the PCIe+ I/O drawers.
The power requirements depend on the number of CPC drawers (1 - 5), number of PCIe I/O
drawers (0 - 12) and I/O features that are installed in the PCIe I/O drawers.
PDUs are installed and serviced from the rear of the frame. Unused power ports should never
be used by any external device.
Each PDU installed requires a customer supplied power feed. The number of power cords
that are required depends on the system configuration.
Note: For initial installation, all power sources are required to run the system checkout
diagnostics successfully.
PDUs are installed in pairs. A system can have 2, 4, 6, or 8 PDUs, depending on the
configuration. Consider the following points:
Paired PDUs are A1/A2, A3/A4, B1/B2, and C1/C2.
From the rear of the system, the odd-numbered PDUs are on the left side of the rack, and
the even-numbered PDUs are on the right side of the rack.
The total loss of one PDU in a pair has no effect on the system operation.
Components that plug into the PDUs for redundancy (using two power cords) include the
following features:
CPC Drawers, PCIe+ I/O drawers, Radiators, and Support Elements
The redundancy for each component is achieved by plugging the power cables into the
paired PDUs.
For example, the top Support Element (1), has one power supply plugged into PDU A1
and the second power supply plugged into the paired PDU A2 for redundancy.
As a best practice, connect the odd-numbered PDUs (A1, B1, C1, and D1) to one power
source or distribution panel, and the even-numbered PDUs (A2, B2, C2, and D2) to a
separate power source or distribution panel.
The frame count rules (number of frames) for z15 T01 are listed in Table 2-12.
CPC drawers 0 1 2 3 4 5 6 7 8 9 10 11 12
1 1 1 1 1 2 2 2 2 2 3 3 3 3
2 1 1 1 2 2 2 2 2 3 3 3 3 3
3 1 1 2 2 2 2 2 3 3 3 3 3 NA
4 2 2 2 2 2 3 3 3 3 3 4 4 4
5 2 2 2 2 3 3 3 3 3 4 4 4 4
The number of CPC drawers and I/O drawers determines the number of racks in the system
and the number of PDUs in the system.
The PDU/line cord rules (number of PDU/Cord pairs) for z15 T01 are listed in Table 2-13
Table 2-13 PDU/line cord rules (# PDU/Cord pairs) for z15 T01
PDU/Linecord I/O drawers
CPC drawers 0 1 2 3 4 5 6 7 8 9 10 11 12
1 1 1 1 1 2 2 2 2 2 3 3 3 3
2 2 2 2 2 2 2 2 2 3 3 3 3 3
3 2 2 2 2 2 2 2 3 3 3 3 3 NA
4 3 3 3 3 3 3 3 3 3 4 4 4 4
5 3 3 3 3 3 3 3 3 3 4 4 4 4
No PDUs are installed in this configuration; instead, 1 or 2 pairs of BPAs are installed that
house the bulk power distribution to all the components, the Bulk Power Regulators (BPR),
and the optional Internal Battery Feature.
The power supply units convert the regulated high-voltage DC power (supplied by BPA) to DC
power that is used as input for the Points of Load (POLs) in the CPC drawer and the PCIe+
I/O drawers.
The power requirements depend on the number of CPC drawers (1 - 5), number of PCIe I/O
drawers (0 - 11), and I/O features that are installed in the PCIe+ I/O drawers.
A schematic view of a maximum configured system with BPA power is shown in Figure 2-31.
Each BPA installed requires a customer supplied power feed. The number of power cords that
are required depends on the system configuration. The total loss of one BPA has no effect on
system operation.
Note: For initial installation, all power sources are required to run the system checkout
diagnostics successfully.
BPAs are installed in pairs (2 or 4), depending on the configuration. Consider the following
points:
BPA1 and BPA2 are always installed initially in Frame A
BPA3 and BPA4 are installed to support more components for larger configurations (CPC
and I/O drawers)
BPA1/BPA2 and BPA3/BPA4 for paired redundancy
From the rear of the system, the odd-numbered BPAs are above the even-numbered BPAs
in both frames
The total loss of one BPA in a pair has no effect on system operation
Note: Customer power sources should always maintain redundancy across BPA pairs; that
is, one power source or distribution panel supplies power for BPA1 and the separate power
source or distribution panel supplies power for BPA2.
As a best practice, connect the odd-numbered BPAs (BPA1 and BPA3) to one power
source or distribution panel, and the even-numbered BPAs (BPA2 and BPA4) to a separate
power source or distribution panel.
Removal of IBF supporta: IBM z15 is planned to be the last IBM Z server to offer an
Internal Battery Feature (IBF). As client data centers continue to improve power stability
and uninterruptible power supply (UPS) coordination, IBM Z continues to innovate to help
clients take advantage of common power efficiency and monitoring across their
ecosystems. Additional support for data center power planning can be requested through
your IBM Sales contact.
a. Statements by IBM regarding its plans, directions, and intent are subject to change or
withdrawal without notice at the sole discretion of IBM.
The IBF further enhances the robustness of the power design, which increases power line
disturbance immunity. It provides battery power to preserve processor data during a loss of
power on all power feeds from the computer room. The IBF can hold power briefly during a
brownout, or for orderly shutdown for a longer outage. For information about the hold times,
which depend on the I/O configuration and amount of CPC drawers, see Chapter 11,
“Environmentals” on page 453.
This tool estimates the power consumption for the specified configuration. The tool does not
verify that the specified configuration can be physically built.
Tip: The exact power consumption for your system varies. The object of the tool is to
estimate the power requirements to aid you in planning for your system installation. Actual
power consumption after installation can be confirmed by using the HMC Monitors
Dashboard task.
For all z15 T01 servers, the CPC drawer components (except for PU SCMs) and the PCIe+
I/O drawers are air cooled by redundant fans. Airflow of the system is directed from front (cool
air) to the back of the system (hot air).
Although the PU SCMs are cooled by water, the heat is exhausted into the room from the
radiator heat exchanger by forced air with blowers. At the system level, these z15 T01 are still
air-cooled systems.
Unlike the radiator in air-cooled models, a WCU has two water loops: An internal closed water
loop and an external (chilled) water loop. The external water loop connects to the
client-supplied building’s chilled water. The internal water loop circulates between the WCU
heat exchanger and the PU SCMs cold plates. The loop takes heat away from the PU SCMs
and transfers it to the external water loop in the WCU’s heat exchanger. For more information,
see 2.9.7, “Water-cooling unit” on page 85.
In addition to the PU SCMs, the internal water loop circulates through two heat exchangers
that are in the path of the exhaust air in the rear of the frames. These heat exchangers
remove approximately 60% - 65% of the residual heat from the I/O drawers, PCIe I/O
drawers, the air-cooled logic in the CPC drawers, and the power enclosures. Almost
two-thirds of the total heat that is generated can be removed from the room by the chilled
water.
Air-cooled models or water-cooled models are chosen when ordering, and the corresponding
equipment is factory-installed. An MES (conversion) from an air-cooled model to a
water-cooled model and vice versa is not possible.
The RCU discharges heat from the internal frame water loop to the customer’s data center.
The new design uses a pump for filling the system and FRUs, and a compressor for draining
FRUs and discontinuance. The process is faster and more effective. The FDT is shown in
Figure 2-32.
The radiator cooling unit (RCU) contains three independent pump FRUs. The cooling
capability is a redundant N+2 design, so a single working pump and blower can support the
entire load. The replacement of one pump or blower can be done concurrently and does not
affect performance.
Each radiator cooling unit contains up to five independent fan assemblies that can be
concurrently serviced. The number of fans present depends on the number of CPC drawers
installed in the frame.
The closed water loop in the radiator unit is shown in Figure 2-34. The warm water that is
exiting from the PU SCMs cold plates enters pumps through a common manifold and is
pumped through a heat exchanger where heat is extracted by the air flowing across the heat
exchanger fins. The cooled water is then recirculated back into the PU SCMs cold plates.
The water in the closed loop within the system exchanges heat with the continuous supply of
building-provided chilled water. This water circulates between the PU SCMs cold plates and a
heat exchanger within the WCU. Heat from the PU SCMs is transferred to the cold plates,
where it is in turn transferred to the circulating system water (closed loop). The system water
then dissipates its heat to the building-provided chilled water within the WCU’s heat
exchanger. The PU SCMs are cooled efficiently in this manner.
z15 T01 servers operate with two fully redundant WCUs in the A-frame and two fully
redundant WCUs the B-frame (when present), each on separate loops. These water-cooling
units each have their own facility feed and return water connections. If water is interrupted to
one of the units, the other unit picks up the entire load, and the server continues to operate
without interruption. You must provide independent redundant water loops to the
water-cooling units to obtain full redundancy.
The internal circulating water is conditioned water that is added to the radiator unit during
system installation with the Fill and Drain Tool (FC 3393). The FDT is included with new z15
servers. The FDT is used to provide the internal water at the installation and for maintenance,
and to remove it at discontinuance.
In addition to the PU SCMs cold plates, the internal water loop circulates through heat
exchangers that are mounted on each frame in the configuration. These exchangers are in
the path of the exhaust air in the rear of the frames.
Depending on the configuration, these heat exchangers remove up to 93% of the residual
heat from the I/O drawers, PCIe I/O drawer, the air-cooled logic in the CPC drawer, and the
power enclosures. The goal is for two-thirds of the total heat that is generated to be removed
from the room by the chilled water.
If one client water supply or one WCU fails, the remaining feed maintains PU SCM cooling for
the frame. The WCUs and the associated drive card are concurrently replaceable. In addition,
the heat exchangers can be disconnected and removed from the system concurrently.
2.10 Summary
All aspects of the z15 T01 structure are listed in Table 2-14
Standard SAPs 4 8 12 16 22
Number of IFP 1 1 1 1 1
Standard spare 2 2 2 2 2
PUs
Enabled Memory 512 - 7936 512 - 16128 512 - 24320 512 - 32512 512 - 40704
sizes GB
Flexible memory N/A 512 - 7936 512 - 16128 512 - 24320 512 - 32512
sizes GB
Clock frequency 5.2 GHz 5.2 GHz 5.2 GHz 5.2 GHz 5.2 GHz
Maximum number 12 24 36 48 60
of PCIe fanouts
Number of support 2 2 2 2 2
elements
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, Features Max34, Max71, Max108,
Max145, and Max190, which is further identified as IBM z15 Model T01, or z15 T01,
unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, Features Max4, Max13, Max21, Max32,
and Max65, which is further identified as IBM z15 Model T02, or z15 T02, unless
otherwise specified.
In the remainder of this chapter, IBM z15 (z15) refers to both machine types.
For more information about the processor unit, see z/Architecture Principles of Operation,
SA22-7832.
z15 servers offer high levels of reliability, availability, serviceability (RAS), resilience, and
security. It fits into the IBM strategy in which mainframes play a central role in creating an
infrastructure for cloud, analytics, and mobile, underpinned by security. The z15 server is
designed so that everything around it, such as operating systems, middleware, storage,
security, and network technologies that support open standards, helps you achieve your
business goals.
The modular CPC drawer design aims to reduce, or in some cases even eliminate, planned
and unplanned outages. The design does so by offering concurrent repair, replace, and
upgrade functions for processors, memory, and I/O. For more information about the z15 RAS
features, see Chapter 9, “Reliability, availability, and serviceability” on page 383.
For more information about frames and configurations, see Chapter 2, “Central processor
complex hardware components” on page 35, and Appendix D, “Frame configurations” on
page 515.
z15 servers CPC continues the line of mainframe processors that are compatible with an
earlier version. Evolution® brings the following processor design enhancements:
Twelve cores per PU chip design
Each PU chip has 3 PCIe Generation 4 ports (x16@32GBps)
Pipeline optimization
Improved SMT and SIMD
It uses 24-, 31-, and 64-bit addressing modes, multiple arithmetic formats, and multiple
address spaces for robust inter-process security.
The remaining sections in this chapter describe the z15 system structure, showing a logical
representation of the data flow from PUs, caches, memory cards, and various interconnect
capabilities.
The following types of CPC drawer configurations are available for z15 T01:
One drawer: Max34
Two drawers: Max71
Three drawers: Max108
Four drawers: Max145
Five drawers: Max190
Note: Max145 and Max190 are factory build only. It is not possible to upgrade in the field to
Max145 or Max190.
The z15 T01 has up to 20 memory controller units (MCUs) for a Max190 feature (four MCUs
per CPC drawer). The MCU configuration uses five-channel redundant array of independent
memory (RAIM) protection, with dual inline memory modules (DIMM) bus cyclic redundancy
check (CRC) error retry.
The cache hierarchy (L1, L2, L3, and L4) is implemented with embedded dynamic random
access memory (eDRAM). Until recently, eDRAM was considered to be too slow for this use.
However, a breakthrough in technology that was made by IBM eliminated that limitation. In
addition, eDRAM offers higher density, less power utilization, fewer soft errors, and better
performance. Concurrent maintenance allows dynamic central processing complex (CPAC)
drawer add and repair.2
1
Federal Information Processing Standard (FIPS)140-2 Security Requirements for Cryptographic Modules
2 For configurations with two or more CPC drawers installed
The z15 T01 cache levels and memory hierarchy are shown in Figure 3-1.
Although L1, L2, and L3 caches are implemented on the PU SCM, the fourth cache level (L4)
is implemented within the system controller (SC) SCM. One L4 cache is present in each CPC
drawer, which is shared by all PU SCMs. The cache structure of the z15 has the following
characteristics:
Larger L1, L2, and L3 caches (more data closer to the core).
L1 and L2 caches use eDRAM, and are private for each PU core.
L2-L3 interface has a new Fetch cancel protocol, a revised L2 Least Recent Used (LRU)
Demote handling.
Considerations
Cache sizes are being limited by ever-diminishing cycle times because they must respond
quickly without creating bottlenecks. Access to large caches costs more cycles. Instruction
and data cache (L1) sizes must be limited because larger distances must be traveled to reach
long cache lines. This L1 access time generally occurs in one cycle, which prevents increased
latency.
Also, the distance to remote caches as seen from the microprocessor becomes a significant
factor. An example is an L4 cache that is not on the microprocessor (and might not even be in
the same CPC drawer). Although the L4 cache is rather large, several cycles are needed to
travel the distance to the cache. The node-cache topology of z15 servers is shown in
Figure 3-2.
Another approach is available for avoiding L4 cache access delays (latency). The L4 cache
straddles up to five CPC drawers. This configuration means that relatively long distances exist
between the higher-level caches in the processors and the L4 cache content.
To overcome the delays that are inherent in the SMP CPC drawer design and save cycles to
access the remote L4 content, the system keeps instructions and data as close to the
processors as possible. This configuration can be managed by directing as much work of a
particular LPAR workload to the processors in the same CPC drawer as the L4 cache. This
configuration is achieved by having the IBM Processor Resource/Systems Manager (PR/SM)
scheduler and the z/OS WLM and dispatcher work together. Have them keep as much work
as possible within the boundaries of as few processors and L4 cache space (which is best
within a CPC drawer boundary) without affecting throughput and response times.
The cache structures of z15 T01 systems are compared with the previous generation of
IBM Z (z14 M0x) in Figure 3-3.
Compared to z14, the z15 cache design has larger L4, L3, and L2 cache sizes. As in z14
servers, in z15 servers, more affinity exists between the memory of a partition, the L4 cache
in the SC, accessed by the two logical clusters in the same CPC drawer, and the cores in the
PU. The access time of the private cache usually occurs in one cycle.
HiperDispatch
To help avoid latency in a high-frequency processor design, PR/SM and the dispatcher must
be prevented from scheduling and dispatching a workload on any processor available, which
keeps the workload in as small a portion of the system as possible. The cooperation between
z/OS and PR/SM is bundled in a function called HiperDispatch. HiperDispatch uses the z15
cache topology, which features reduced cross-cluster “help” and better locality for multi-task
address spaces.
PR/SM can use dynamic PU reassignment to move processors (CPs, ZIIPs, IFLs, ICFs,
SAPs, and spares) to a different chip, node, and drawer to improve the reuse of shared
caches by processors of the same partition. It can use dynamic memory relocation (DMR) to
move a running partition’s memory to different physical memory to improve the affinity and
reduce the distance between the memory of a partition and the processors of the partition.
For more information about HiperDispatch, see 3.7, “Logical partitioning” on page 132.
The z15 T01 intra-CPC drawer communication structure is shown in Figure 3-4.
Inter-CPC drawer communication occurs at the L4 cache level, which is implemented on the
SC chips in each drawer. The SC function regulates coherent drawer-to-drawer traffic.
z15 T01 core frequency is 5.2 GHz (same as z14 M0x), but with increased number of
processors that share larger caches to have shorter access times and improved capacity and
performance.
Through innovative processor design (pipeline and cache management redesigns), the IBM Z
processor performance continues to evolve. Enhancements were made on the processor unit
design, such as on out-of-order execution, branch prediction mechanism, Floating point and
vector unit, Divide engine scheduler, Load/Store Unit and Operand Store Compare (OSC),
simultaneous multi-threading, and the relative nest intensity (RNI) redesigns. For more
information about RNI, see 12.4, “Relative Nest Intensity” on page 481.
The processing performance was enhanced by bringing the following changes on the z15
processor design:
Core optimization to enable performance and capacity growth.
New EDRAM macro design with 2x macro density (cache growths and L2-L3 Protocol
changes to reduce latency.
Because of those enhancements, the z15 processor full speed z/OS single-thread
performance is on average 1.12 - 1.14 times faster than the z14 at equal N-way. For more
information about performance, see Chapter 12, “Performance” on page 475.
z13 servers introduced architectural extensions with instructions that reduce processor
quiesce effects, cache misses, and pipeline disruption, and increase parallelism with
instructions that process several operands in a single instruction (SIMD). The processor
architecture was further developed for z14 and z15. z15 includes the following
enhancements:
Optimized second-generation SMT
Enhanced SIMD instructions set
Improved Out-of-Order core execution
Improvements in branch prediction and handling
Pipeline optimization
Secure Execution3
Co-processor compression enhancement
The z15 enhanced Instruction Set Architecture (ISA) includes a set of instructions that are
added to improve compiled code efficiency. These instructions optimize PUs to meet the
demands of various business and analytics workload types without compromising the
performance characteristics of traditional workloads.
An operating system with SMT support can be configured to dispatch work to a thread on a
zIIP (for eligible workloads in z/OS) or an IFL (for z/VM and Linux on Z) core in single thread
or SMT mode so that HiperDispatch cache optimization can be considered. For more
information about operating system support, see Chapter 7, “Operating system support” on
page 253.
SMT technology allows instructions from more than one thread to run in any pipeline stage at
a time. SMT can handle up to four pending translations.
Each thread has its own unique state information, such as Program Status Word - S/360
Architecture (PSW) and registers. The simultaneous threads cannot necessarily run
instructions instantly and must at times compete to use certain core resources that are
shared between the threads. In some cases, threads can use shared resources that are not
experiencing competition.
Figure 3-6 Two threads running simultaneously on the same processor core
The use of SMT provides more efficient use of the processors’ resources and helps address
memory latency, which results in overall throughput gains. The active thread shares core
resources in space, such as data and instruction caches, TLBs, branch history tables, and, in
time, pipeline slots, execution units, and address translators.
Although SMT increases the processing capacity, the performance in some cases might be
superior if a single thread is used. Enhanced hardware monitoring supports measurement
through CPUMF for thread usage and capacity.
For workloads that need maximum thread speed, the partition’s SMT mode can be turned off.
For workloads that need more throughput to decrease the dispatch queue size, the partition’s
SMT mode can be turned on.
SMT use is functionally transparent to middleware and applications, and no changes are
required to run them in an SMT-enabled partition.
The 32 vector registers feature 128 bits. The instructions include string operations, vector
integer, and vector floating point operations. Each register contains multiple data elements of
a fixed size. The following instructions code specifies which data format to use and the size of
the elements:
Byte (16 8-bit operands)
Halfword (eight 16-bit operands)
Word (four 32-bit operands)
Doubleword (two 64-bit operands)
Quadword (one 128-bit operand)
In addition to the instructions that were introduced in z14, the SIMD provides the following
enhancements for z15:
Double-bandwidth vector loads
Multiply/Divide speed ups
Conversion speed ups
New and enhanced vector instructions
Load/store reversed (to help with little endian conversion)
More vector shift operations
VECTOR STRING SEARCH, for fast string search, supporting different encodings
New vector FP converts
New and enhanced Vector Packed Decimal instructions
The collection of elements in a register is called a vector. A single instruction operates on all
of the elements in the register. Instructions include a non-destructive operand encoding that
allows the addition of the register vector A and register vector B and stores the result in the
register vector A (A = A + B).
A schematic representation of a SIMD instruction with 16-byte size elements in each vector
operand is shown in Figure 3-7.
For most operations, the condition code is not set. A summary condition code is used only for
a few instructions.
The z15 processor includes pipeline enhancements that benefit Out-of-Order execution. The
IBM Z processor design features advanced micro-architectural innovations that provide the
following benefits:
Maximized instruction-level parallelism (ILP) for a better cycles per instruction (CPI)
design.
Maximized performance per watt. Two cores are added (as compared to the z14 chip) at
slightly higher chip power.
Enhanced instruction dispatch and grouping efficiency.
Increased OoO resources (Global Completion Table entries, physical GPR entries, and
physical FPR entries).
Improved completion rate.
Reduced cache/TLB miss penalty.
Improved execution of D-Cache store and reload and new Fixed-point divide.
New Operand Store Compare (OSC) (load-hit-store conflict) avoidance scheme.
Enhanced branch prediction structure and sequential instruction fetching.
This implementation requires special circuitry to make execution and memory accesses
display in order to the software. The logical diagram of a z15 core is shown in Figure 3-9.
Memory address generation and memory accesses can occur out of (program) order. This
capability can provide a greater use of the z15 superscalar core, and improve system
performance.
The z15 T01 processor unit core is a superscalar, out-of-order, SMT processor with 12
execution units. Up to six instructions can be decoded per cycle, and up to 12 instructions or
operations can be started to run per clock cycle (0.192 ns). The execution of the instructions
can occur out of program order, and memory address generation and memory accesses can
also occur out of program order. Each core has special circuitry to display execution and
memory accesses in order to the software.
The z15 superscalar PU core can have up to 12 instructions or operations that are running
per cycle. This technology results in shorter workload runtime.
For this reason, various history-based branch prediction mechanisms are used, as shown on
the in-order part of the z15 PU core logical diagram in Figure 3-9 on page 103. The branch
target buffer (BTB) runs ahead of instruction cache pre-fetches to prevent branch misses in
an early stage. Furthermore, a branch history table (BHT), in combination with a pattern
history table (PHT) and the use of tagged multi-target prediction technology branch
prediction, offers a high branch prediction success rate.
The z15 microprocessor improves the branch prediction throughput by using a simplified and
larger BTB.
Many challenges exist in creating an efficient superscalar processor. The superscalar design
of the PU made significant strides in avoiding address generation interlock (AGI) situations.
Instructions that require information from memory locations can suffer multi-cycle delays to
get the needed memory content. Because high-frequency processors wait “faster” (spend
processor cycles more quickly while idle), the cost of getting the information might become
prohibitive.
One Nest Accelerator Unit (NXU) is used per processor chip, which is shared by all cores on
the chip and features the following benefits:
Brand new concept of sharing and operating an accelerator function in the nest
Supports DEFLATE compliant compression/decompression and GZIP CRC/ZLIB Adler
Low latency
Based on IBM benchmarks, the largest IBM z15 T01 with the Integrated Accelerator for zEDC
provides up to 17 times the total compression throughput (compress up to 260 GBps with the
Integrated Accelerator for zEDC) of a z14 (3906) configured with the maximum number of
zEDC Express cards. Also, the Integrated Accelerator for zEDC on z15 improves the
compression ratio by 5% over z14 zEDC Express.4
Sharing of zEDC cards is limited to 15 LPARs per adaptor. The On-Chip Compression
Accelerator removes this virtualization constraint because it is shared by all PUs on the
processors chip and is therefore available to all LPARs and guests.
Moving the compression function from the IO drawer to the processor chip means that
compression can operate directly on L3 cache and data does not need to be passed by using
I/O.
Synchronous execution occurs in problem state where the user application starts the
instruction in its virtual address space.
Asynchronous execution is optimized for Large Operations under z/OS for authorized
applications (for example, BSAM) and issues I/O using EADMF for asynchronous execution.
Asynchronous execution maintains the current user experience and provides a transparent
implementation for existing authorized users of zEDC.
4
Measurements were collected in a controlled environment running an IBM developed workload under z/OS.
Individual results can vary. Results are workload dependent.
For information about sizing, migration considerations, and software support, see
Appendix C, “IBM Integrated Accelerator for zEnterprise Data Compression” on page 509.
Coprocessor units
One coprocessor unit is available for compression and cryptography on each core in the chip.
The compression engine uses static dictionary compression and expansion. The
compression dictionary uses the L1-cache (instruction cache).
The cryptography engine is used for the CPACF, which offers a set of symmetric
cryptographic functions for encrypting and decrypting of clear key operations.
Compression enhancements
The compression features the following enhancements:
Huffman compression on top of CMPSC compression (embedded in dictionary, reuse of
generators)
Order Preserving compression in B-Trees and other index structures
Faster expansion algorithms
Reduced overhead on short data
CPACF
CPACF accelerates the encrypting and decrypting of SSL/TLS transactions, virtual private
network (VPN)-encrypted data transfers, and data-storing applications that do not require
FIPS 140-2 level 4 security. The assist function uses a special instruction set for symmetrical
clear key cryptographic encryption and decryption, and for hash operations. This group of
instructions is known as the Message-Security Assist (MSA). For more information about
these instructions, see z/Architecture Principles of Operation, SA22-7832.
For more information about cryptographic functions on z15 servers, see Chapter 6,
“Cryptographic features” on page 215.
IBM z15 processor adds special hardware to significantly accelerate frequently used
functions. The IBM Integrated Accelerator for Z sort has been implemented on each core and
it provides new architected instructions for sorting data to speed up sorting operations. It
supports multiple active sorts in parallel and it is designed to:
Base 10 arithmetic is used for most business and financial computation. Floating point
computation that is used for work that is typically done in decimal arithmetic involves frequent
data conversions and approximation to represent decimal numbers. This process makes
floating point arithmetic complex and error-prone for programmers who use it for applications
in which the data is typically decimal.
z15 servers have two DFP accelerator units per core, which improve the decimal floating
point execution bandwidth. The floating point instructions operate on newly designed vector
registers (32 new 128-bit registers).
z15 servers include new decimal floating point packed conversion facility support with the
following benefits:
Reduces code path length because extra instructions to format conversion are no longer
needed.
Packed data is operated in memory by all decimal instructions without general-purpose
registers, which were required only to prepare for decimal floating point packed conversion
instruction.
Converting from packed can now force the input packed value to positive instead of
requiring a separate OI, OILL, or load positive instruction.
Software support
DFP is supported in the following programming languages and products:
Release 4 and later of the High Level Assembler
C/C++, which requires supported z/OS version
Enterprise PL/I Release 3.7 and Debug Tool Release 8.1 or later
Java Applications that use the BigDecimal Class Library
SQL support as of Db2 Version 9 and later
The z15 core implements two other execution subunits for 2x throughput on BFP
(single/double precision) operations (see Figure 3-9 on page 103).
The key point is that Java and C/C++ applications tend to use IEEE BFP operations more
frequently than earlier applications. Therefore, the better the hardware implementation of this
set of instructions, the better the performance of applications.
Relocation under hardware control is possible because the R-unit has the full designed state
in its buffer. PU error detection and recovery are shown in Figure 3-12 on page 109.
The BHT offers significant branch performance benefits. The BHT allows each PU to take
instruction branches that are based on a stored BHT, which improves processing times for
calculation routines. In addition to the BHT, z15 servers use the following techniques to
improve the prediction of the correct branch to be run:
BTB
PHT
BTB data compression
The success rate of branch prediction contributes significantly to the superscalar aspects of
z15 servers. This success is because the architecture rules prescribe that, for successful
parallel execution of an instruction stream, the correctly predicted result of the branch is
essential.
The z15 branch prediction includes the following enhancements over z14:
Branch prediction search pipeline extended from five to six cycles to accommodate new
predictors for increased accuracy/performance.
Predictors enhancements:
– SKOOT enhanced BTB search (SKip Over OffesetT entries remember where the
branch search should continue). This enhancement saves BTB1 access cycles while
searching for the next branch.
– SSCRS (hardware-based super simple call-return stack). The major change is that it
does not necessarily have to return to the next sequential instruction (NSIA). This
feature supports return branches up to 8 bytes past the NSIA.
– New TAGE (TAgged GEometric) history length branch predicator.
Branch prediction is simplified by removing the BTBp ‘‘BTB1 victim cache’’, which is
replaced by a write buffer. The two independent read ports were replaced by one
double-bandwidth port, which simplifies physical design significantly:
– Level 1 Branch Target Buffer (BTB1): 2K rows x 4sets → 2 K rows x 8sets
– Level 2 Branch Target Buffer (BTB2) size remains unchanged
Better power efficiency: Several structures were redesigned to maintain their accuracy
while less power is used through smart access algorithms.
New static IBM IA regions expanded from four to eight. To conserve space, prediction
structures do not store full target addresses. Instead, they use the locality and limited
ranges of “4gig regions” of virtual instruction addresses - IA(0:31).
With the wild branch hardware facility, the last address from which a successful branch
instruction was run is kept. z/OS uses this information with debugging aids, such as the SLIP
command, to determine from where a wild branch came. It can also collect data from that
The size of the TLB is kept as small as possible because of its short access time
requirements and hardware space limitations. Because memory sizes recently increased
significantly as a result of the introduction of 64-bit addressing, a smaller working set is
represented by the TLB.
To increase the working set representation in the TLB without enlarging the TLB, large (1 MB)
page and giant page (2 GB) support is available and can be used when appropriate. For more
information, see “Large page support” on page 130.
With the enhanced DAT-2 (EDAT-2) improvements, the IBM Z servers support 2 GB page
frames.
The new translation engine allows up to four translations pending concurrently. Each
translation step is ~2x faster, which helps second level guests.
Instruction fetching
Instruction fetching normally tries to get as far ahead of instruction decoding and execution as
possible because of the relatively large instruction buffers available. In the microprocessor,
smaller instruction buffers are used. The operation code is fetched from the I-cache and put in
instruction buffers that hold prefetched data that is awaiting decoding.
Instruction decoding
The processor can decode up to six instructions per cycle. The result of the decoding process
is queued and later used to form a group.
Instruction grouping
From the instruction queue, up to 12 instructions can be completed on every cycle. A
complete description of the rules is beyond the scope of this publication.
The compilers and JVMs are responsible for selecting instructions that best fit with the
superscalar microprocessor. They abide by the rules to create code that best uses the
superscalar implementation. All IBM Z compilers and JVMs are constantly updated to benefit
from new instructions and advances in microprocessor designs.
The Transaction Execution Facility provides instructions, including declaring the beginning
and end of a transaction, and canceling the transaction. TX is expected to provide significant
performance benefits and scalability by avoiding most locks. This benefit is especially
important for heavily threaded applications, such as Java.
3.5.1 Overview
All PUs on a z15 server are physically identical. When the system is initialized, one integrated
firmware processor (IFP) is allocated from the pool of PUs that is available for the entire
system. The other PUs can be characterized to specific functions (CP, IFL, ICF, zIIP, or SAP).
The function that is assigned to a PU is set by the Licensed Internal Code (LIC). The LIC is
loaded when the system is initialized at power-on reset (POR) and the PUs are characterized.
This design brings outstanding flexibility to z15 servers because any PU can assume any
available characterization. The design also plays an essential role in system availability
because PU characterization can be done dynamically, with no system outage.
For more information about software level support of functions and features, see Chapter 7,
“Operating system support” on page 253.
Concurrent upgrades
For all z15 T01 features that have more processor units (PUs) installed (non-characterized)
than activated, concurrent upgrades can be done by LIC activation, which assigns a PU
function to a previously non-characterized PU. No hardware changes are required. The
upgrade can be done concurrently through the following facilities:
Customer Initiated Upgrade (CIU) for permanent upgrades
On/Off Capacity on Demand (On/Off CoD) for temporary upgrades
Capacity BackUp (CBU) for temporary upgrades
Capacity for Planned Event (CPE) for temporary upgrades
If the PU chips in the installed CPC drawers have no available remaining PUs, an upgrade
results in a feature upgrade and the installation of an extra CPC drawer. Field add (MES) of a
CPC drawer is possible for z15 Model T01 features Max34 and Max71 only. These features
can be upgraded to a Max108 provided initial order for the CPC Reserve features FC 2271 or
FC 2272. CPC drawer installation is nondisruptive, but takes more time than a simple LIC
upgrade. Features Max145 and Max190 are factory build only.
For more information about Capacity on Demand, see Chapter 8, “System upgrades” on
page 333.
PU sparing
If a PU failure occurs, the failed PU’s characterization is dynamically and transparently
reassigned to a spare PU. z15 servers have two spare PUs. PUs that are not characterized
on a CPC configuration can also be used as extra spare PUs. For more information about PU
sparing, see 3.5.10, “Sparing rules” on page 127.
All assigned PUs are grouped in the PU pool. These PUs are dispatched to online logical
PUs. As an example, consider a z15 server with 10 CPs, 2 IFLs, 5 zIIPs, and 1 ICF. This
system has a PU pool of 18 PUs, called the pool width. Subdivision defines the following
pools:
A CP pool of 10 CPs
An ICF pool of one ICF
An IFL pool of two IFLs
A zIIP pool of five zIIPs
PUs are removed from their pools when a concurrent downgrade occurs as the result of the
removal of a CBU. They are also removed through the On/Off CoD process and the
conversion of a PU. When a dedicated LPAR is activated, its PUs are taken from the correct
pools. This process is also the case when an LPAR logically configures a PU as on, if the
width of the pool allows for it.
For an LPAR, logical PUs are dispatched from the supporting pool only. The logical CPs are
dispatched from the CP pool, logical zIIPs from the zIIP pool, logical IFLs from the IFL pool,
and the logical ICFs from the ICF pool.
PU weighting
Because CPs, zIIPs, IFLs, and ICFs have their own pools from where they are dispatched,
they can be given their own weights. For more information about PU pools and processing
weights, see the IBM Z Processor Resource/Systems Manager Planning Guide, SB10-7175.
The z15 server can be initialized in LPAR (PR/SM) mode or in Dynamic Partition Manger
(DPM) mode.
CPs are defined as dedicated or shared. Reserved CPs can be defined to an LPAR to allow
for nondisruptive image upgrades. If the operating system in the LPAR supports the logical
processor add function, reserved processors are no longer needed. Regardless of the
installed model, an LPAR can have up to 190 logical CPs that are defined (the sum of active
and reserved logical CPs). In practice, define no more CPs than the operating system
supports.
The z15 T01 server recognizes four distinct capacity settings for CPs. Full-capacity CPs are
identified as CP7. In addition to full-capacity CPs, three subcapacity settings (CP6, CP5, and
CP4), each for up to 34 PUs, are offered.
Granular capacity adds 102 subcapacity settings to the 190 capacity settings that are
available with full capacity CPs (CP7). Each of the 102 subcapacity settings applies to up to
34 CPs only, independent of the model installed.
Information about CPs in the remainder of this chapter applies to all CP capacity settings,
unless indicated otherwise. For more information about granular capacity, see Chapter 2.3.3,
“PU characterization” on page 52.
Note: IFLs can be dedicated to a Linux, a z/VM, or LPAR, or can be shared by multiple
Linux guests, z/VM LPARs, or SSC that are running on the same z15 server. Only z/VM,
Linux on Z operating systems, SSC, and designated software products can run on IFLs.
IFLs are orderable by using FC 1945.
IFL pool
All PUs that are characterized as IFLs within a configuration are grouped into the IFL pool.
The IFL pool can be seen on the HMC workplace.
IFLs do not change the model capacity identifier of the z15 server. Software product license
charges that are based on the model capacity identifier are not affected by the addition of
IFLs.
Unassigned IFLs
An IFL that is purchased but not activated is registered as an unassigned IFL (FC 1948).
When the system is later upgraded with another IFL, the system recognizes that an IFL was
purchased and is present.
The allowable number of zIIPs for each model is listed in Table 3-2.
ICFs exclusively run CFCC. ICFs do not change the model capacity identifier of the z15
system. Software product license charges that are based on the model capacity identifier are
not affected by the addition of ICFs.
All ICFs within a configuration are grouped into the ICF pool. The ICF pool can be seen on the
HMC workplace.
The ICFs can be used by coupling facility LPARs only. ICFs are dedicated or shared. ICFs
can be dedicated to a CF LPAR, or shared by multiple CF LPARs that run on the same
system. However, having an LPAR with dedicated and shared ICFs at the same time is not
possible.
After the image is dispatched, “poll for work” logic in CFCC and z/OS can be used largely
as-is to locate and process the work. The new interrupt expedites the redispatching of the
partition.
LPAR presents these Coupling Thin Interrupts to the guest partition, so CFCC and z/OS both
require interrupt handler support that can deal with them. CFCC also changes to relinquish
control of the processor when all available pending work is exhausted, or when the LPAR
undispatches it off the shared processor, whichever comes first.
CF processor combinations
A CF image can have one of the following combinations that are defined in the image profile:
Dedicated ICFs
Shared ICFs
Dedicated CPs
Shared CPs
Shared ICFs add flexibility. However, running only with shared coupling facility PUs (ICFs or
CPs) is not a preferable production configuration. It is preferable for a production CF to
operate by using dedicated ICFs.
In Figure 3-13, the CPC on the left has two environments that are defined (production and
test), and each has one z/OS and one coupling facility image. The coupling facility images
share an ICF.
The LPAR processing weights are used to define how much processor capacity each CF
image can include. The capped option can also be set for a test CF image to protect the
production environment.
Connections between these z/OS and CF images can use internal coupling links to avoid the
use of real (external) coupling links, and get the best link bandwidth available.
DYNDISP allows more environments with multiple CF images to coexist in a server, and to
share CF engines with reasonable performance. For more information, see 3.9.3, “Dynamic
CF dispatching” on page 148.
To improve CF processor scaling for the customer’s CF images and to make effective use of
more processors as the sysplex workload increases, CF work management and dispatcher
provide the following improvements (z15):
Non-prioritized (FIFO-based) work queues, which avoids overhead of maintaining ordered
queues in the CF.
Streamlined system-managed duplexing protocol, which avoids costly latching deadlocks
that can occur between primary and secondary structure.
“Functionally specialized” ICF processors that operate for CF images with dedicated
processors defined under certain conditions that realizes the following benefits:
– One “functionally specialized” processor for inspecting suspended commands
– One “functionally specialized” processor for pulling in new commands
– The remaining processors are non-specialized for general CF request processing
– Avoids many inter-processor contentions that were associated with CF dispatching
A zIIP enables eligible z/OS workloads to have a portion of them directed for execution to a
processor that is characterized as a zIIP. The zIIPs do not increase the MSU value of the
processor and so do not affect the IBM software license changes.
z15 is the third generation of IBM Z processors to support SMT. z15 servers implement two
threads per core on IFLs and zIIPs. SMT must be enabled at the LPAR level and supported by
the z/OS operating system. SMT was enhanced for z15 and it is enabled for SAPs by default
(no customer intervention required).
New with z/OS 2.4, the z/OS Container Extensions6 allows deployment of Linux on Z software
components, such as Docker Containers in a z/OS system, in direct support of z/OS
workloads without requiring a separately provisioned Linux server, while maintaining overall
solution operational control within z/OS and with z/OS qualities of service. Workload deployed
in z/OS Container Extensions is zIIP eligible.
5
IBM z Systems® Application Assist Processors (zAAPs) are not available since z14 servers. A zAAP workload is
dispatched to available zIIPs (zAAP on zIIP capability).
6 z/OS Container Extensions require ordering the Container Hosting Foundation feature with Z systems (FC 0104).
This process reduces the CP time that is needed to run Java WebSphere applications, which
frees that capacity for other workloads.
The logical flow of Java code that is running on a z15 server that has a zIIP available is shown
in Figure 3-14. When JVM starts the execution of a Java program, it passes control to the
z/OS dispatcher that verifies the availability of a zIIP.
A zIIP runs only IBM authorized code. This IBM authorized code includes the z/OS JVM in
association with parts of system code, such as the z/OS dispatcher and supervisor services.
A zIIP cannot process I/O or clock comparator interruptions, and it does not support operator
controls, such as IPL.
Java application code can run on a CP or a zIIP. The installation can manage the use of CPs
so that Java application code runs only on CPs, only on zIIPs, or on both.
Two execution options for zIIP-eligible code execution are available. These options are
user-specified in IEAOPTxx and can be dynamically altered by using the SET OPT command.
The following options are supported for z/OS7:
Option 1: Java dispatching by priority (IIPHONORPRIORITY=YES)
This option is the default option and specifies that CPs must not automatically consider
zIIP-eligible work for dispatching on them. The zIIP-eligible work is dispatched on the zIIP
engines until Workload Manager (WLM) determines that the zIIPs are overcommitted.
WLM then requests help from the CPs. When help is requested, the CPs consider
dispatching zIIP-eligible work on the CPs themselves based on the dispatching priority
relative to other workloads. When the zIIP engines are no longer overcommitted, the CPs
stop considering zIIP-eligible work for dispatch.
This option runs as much zIIP-eligible work on zIIPs as possible. It also allows it to spill
over onto the CPs only when the zIIPs are overcommitted.
Option 2: Java dispatching by priority (IIPHONORPRIORITY=NO)
zIIP-eligible work runs on zIIPs only while at least one zIIP engine is online. zIIP-eligible
work is not normally dispatched on a CP, even if the zIIPs are overcommitted and CPs are
unused. The exception is that zIIP-eligible work can sometimes run on a CP to resolve
resource conflicts.
Therefore, zIIP-eligible work does not affect the CP utilization that is used for reporting
through the subcapacity reporting tool (SCRT), no matter how busy the zIIPs are.
If zIIPs are defined to the LPAR but are not online, the zIIP-eligible work units are processed
by CPs in order of priority. The system ignores the IIPHONORPRIORITY parameter in this
case and handles the work as though it had no eligibility to zIIPs.
7 z/OS V2R1 and later (older z/OS versions are out of support)
The following Db2 UDB for z/OS V8 or later workloads are eligible to run in Service Request
Block (SRB) mode:
Query processing of network-connected applications that access the Db2 database over a
TCP/IP connection by using IBM Distributed Relational Database Architecture™ (DRDA).
DRDA enables relational data to be distributed among multiple systems. It is native to Db2
for z/OS, which reduces the need for more gateway products that can affect performance
and availability. The application uses the DRDA requester or server to access a remote
database. IBM Db2 Connect is an example of a DRDA application requester.
Star schema query processing, which is mostly used in business intelligence work. A star
schema is a relational database schema for representing multidimensional data. It stores
data in a central fact table and is surrounded by more dimension tables that hold
information about each perspective of the data. For example, a star schema query joins
various dimensions of a star schema data set.
Db2 utilities that are used for index maintenance, such as LOAD, REORG, and REBUILD.
Indexes allow quick access to table rows, but over time, the databases become less
efficient and must be maintained as data in large databases is manipulated.
The zIIP runs portions of eligible database workloads, which helps to free computer capacity
and lower software costs. Not all Db2 workloads are eligible for zIIP processing. Db2 UDB for
z/OS V8 and later gives z/OS the information to direct portions of the work to the zIIP. The
result is that in every user situation, different variables determine how much work is redirected
to the zIIP.
On a z15 server, the following workloads can also benefit from zIIPs:
z/OS Communications Server uses the zIIP for eligible Internet Protocol Security (IPSec)
network encryption workloads. This configuration requires z/OS V1R10 or later. Portions
of IPSec processing take advantage of the zIIPs, specifically end-to-end encryption with
IPSec. The IPSec function moves a portion of the processing from the general-purpose
processors to the zIIPs. In addition, to run the encryption processing, the zIIP also handles
the cryptographic validation of message integrity and IPSec header processing.
z/OS Global Mirror, formerly known as Extended Remote Copy (XRC), also uses the zIIP.
Most z/OS Data Facility Storage Management Subsystem (DFSMS) system data mover
(SDM) processing that is associated with z/OS Global Mirror can run on the zIIP. This
configuration requires z/OS V1R10 or later releases.
The first IBM user of z/OS XML system services is Db2 V9. For Db2 V9 before the z/OS
XML System Services enhancement, z/OS XML System Services non-validating parsing
was partially directed to zIIPs when used as part of a distributed Db2 request through
DRDA. This enhancement benefits Db2 V9 by making all z/OS XML System Services
non-validating parsing eligible to zIIPs. This configuration is possible when processing is
used as part of any workload that is running in enclave SRB mode.
z/OS Communications Server also allows the HiperSockets Multiple Write operation for
outbound large messages (originating from z/OS) to be run by a zIIP. Application
workloads that are based on XML, HTTP, SOAP, and Java, and traditional file transfer can
benefit.
For business intelligence, IBM Scalable Architecture for Financial Reporting provides a
high-volume, high-performance reporting solution by running many diverse queries in
z/OS batch. It can also be eligible for zIIP.
For more information about zIIP and eligible workloads, see the IBM zIIP website.
zIIPs are orderable by using FC 1947. Up to two zIIPs can be ordered for each CP or marked
CP configured in the system. If the installed CPC drawer has no remaining unassigned PUs,
the assignment of the next zIIP might require the installation of another CPC drawer.
PUs that are characterized as zIIPs within a configuration are grouped into the zIIP pool. This
configuration allows zIIPs to have their own processing weights, independent of the weight of
parent CPs. The zIIP pool can be seen on the hardware console.
The number of permanent zIIPs plus temporary zIIPs cannot exceed twice the number of
purchased CPs plus temporary CPs. Also, the number of temporary zIIPs cannot exceed the
number of permanent zIIPs.
LPAR: In an LPAR, as many zIIPs as are available can be defined together with at least
one CP.
Standard SAPs 4 8 12 16 22
SAP configuration
A standard SAP configuration provides a well-balanced system for most environments.
However, some application environments feature high I/O rates, typically Transaction
Processing Facility (TPF) environments. In this case, more SAPs can be ordered. Assigning
more SAPs can increase the capability of the channel subsystem to run I/O operations. In z15
8
2:1 ratio can be exceeded (during boost periods) if System Recovery Boost Upgrade (FC 9930 and FC 6802) is
used for activating temoprary zIIP capacity.
9 Enabled by default, cannot be changed or altered by user
By using reserved processors, you can define more logical processors than the number of
available CPs, IFLs, ICFs, and zIIPs in the configuration to an LPAR. This process makes it
possible to nondisruptively configure online more logical processors after more CPs, IFLs,
ICFs, and zIIPs are made available concurrently. They can be made available with one of the
capacity on-demand options.
The maximum number of reserved processors that can be defined to an LPAR depends on
the number of logical processors that are defined. The maximum number of logical
processors plus reserved processors is 190. If the operating system in the LPAR supports the
logical processor add function, reserved processors are no longer needed.
Do not define more active and reserved processors than the operating system for the LPAR
can support. For more information about logical processors and reserved processors and
their definitions, see 3.7, “Logical partitioning” on page 132.
The z15 T01 PU assignment is based on CPC drawer plug order (not “ordering”). Feature
upgrade provides more processor (CPC) drawers. Max108 cannot be upgraded because the
supposed targeted features (Max145 and Max190) are factory built only.
The CPC drawers are populated from the bottom up. This process defines following the
low-order and the high-order CPC drawers:
CPC drawer 1 (CPC 0 at position A10): Plug order 1 (low-order CPC drawer)
CPC drawer 2 (CPC 1 at position A15): Plug order 2
CPC drawer 3 (CPC 2 at position A20): Plug order 3 (high-order CPC drawer
These rules are intended to isolate, as much as possible, on different CPC drawers and even
on different PU chips, processors that are used by different operating systems. This
configuration ensures that different operating systems do not use the same shared caches.
For example, CPs and zIIPs are all used by z/OS, and can benefit by using the same shared
caches. However, IFLs are used by z/VM and Linux, and ICFs are used by CFCC. Therefore,
for performance reasons, the assignment rules prevent them from sharing L3 and L4 caches
with z/OS processors.
This initial PU assignment, which is done at POR, can be dynamically rearranged by an LPAR
by swapping an active core to a core in a different PU chip in a different CPC drawer or cluster
to improve system performance. For more information, see “LPAR dynamic PU reassignment”
on page 138.
When a CPC drawer is added concurrently after POR and new LPARs are activated, or
processor capacity for active partitions is dynamically expanded, the extra PU capacity can
be assigned from the new CPC drawer. The processor unit assignment rules consider the
newly installed CPC drawer dynamically.
10 IBM zHyperLink Express1.1 and IBM zHyperLink Express are not managed by Resource Groups LIC
Systems with a failed PU for which no spare is available call home for a replacement. A
system with a failed PU that is spared and requires an SCM to be replaced (referred to as a
pending repair) can still be upgraded when sufficient PUs are available.
With transparent sparing, the status of the application that was running on the failed
processor is preserved. The application continues processing on a newly assigned CP, IFL,
ICF, zIIP, SAP, or IFP (allocated to one of the spare PUs) without client intervention.
Application preservation
If no spare PU is available, application preservation (z/OS only) is started. The state of the
failing processor is passed to another active processor that is used by the operating system.
Through operating system recovery services, the task is resumed successfully (in most
cases, without client intervention).
3.6.1 Overview
The z15 T01 memory design also provides flexibility, high availability, and the following
upgrades:
Concurrent memory upgrades if the physically installed capacity is not yet reached
z15 servers can have more physically installed memory than the initial available capacity.
Memory upgrades within the physically installed capacity can be done concurrently by LIC,
and no hardware changes are required. However, memory upgrades cannot be done
through CBU or On/Off CoD.
Concurrent memory upgrades if the physically installed capacity is reached
Physical memory upgrades require a processor drawer to be removed and reinstalled after
replacing the memory cards in the processor drawer. Except for the feature Max34, the
combination of enhanced drawer availability and the flexible memory option allows you to
concurrently add memory to the system. For more information, see 2.5.5, “Drawer
replacement and memory” on page 65, and 2.5.7, “Flexible Memory Option” on page 65.
When the total capacity that is installed has more usable memory than required for a
configuration, the LIC Configuration Control (LICCC) determines how much memory is used
from each processor drawer. The sum of the LICCC provided memory from each CPC drawer
is the amount that is available for use in the system.
Memory allocation
When the system is activated by using a POR, PR/SM determines the total installed memory
and the customer enabled memory. Later in the process, during LPAR activation, PR/SM
assigns and allocates each partition memory according to their image profile.
In older IBM Z processors, memory allocation was striped across the available CPC drawers
because relatively fast11 connectivity existed between the drawers. Splitting the work between
all of the memory controllers allowed a smooth performance variability.
The memory allocation algorithm changed starting with IBM z13. For z15, PR/SM tries to
allocate memory into a single CPC drawer. If memory does not fit into a single drawer, PR/SM
tries to allocate the memory into the CPC drawer with the most processor entitlement.
The PR/SM memory and logical processor resources allocation goal is to place all partition
resources on a single CPC drawer, if possible. The resources, such as memory and logical
processors, are assigned to the logical partitions at the time of their activation. Later on, when
all partitions are activated, PR/SM can move memory between CPC drawers to benefit the
performance of each LPAR, without operating system knowledge. This process was done on
the previous families of IBM Z servers only for PUs that use PR/SM dynamic PU reallocation.
With z15 servers, this process occurs whenever the configuration changes, such as in the
following circumstances:
Activating or deactivating an LPAR
Changing the LPARs processing weights
Upgrading the system through a temporary or permanent record
Downgrading the system through deactivation of a temporary record
PR/SM schedules a global reoptimization of the resources in use. It does so by looking at all
the partitions that are active and prioritizing them based on their processing entitlement and
weights, which creates a high and low priority rank. Then, the resources, such as logical
processors and memory, can be moved from one CPC drawer to another to address the
priority ranks that were created.
When partitions are activated, PR/SM tries to find a home assignment CPC drawer, home
assignment node, and home assignment chip for the logical processors that are defined to
them. The PR/SM goal is to allocate all the partition logical processors and memory to a
single CPC drawer (the home drawer for that partition).
If all logical processors can be assigned to a home drawer and the partition-defined memory
is greater than what is available in that drawer, the exceeding memory amount is allocated on
another CPC drawer. If all the logical processors cannot fit in one CPC drawer, the remaining
logical processors spill to another CPC drawer. When that overlap occurs, PR/SM stripes the
memory (if possible) across the CPC drawers where the logical processors are assigned.
The process of reallocating memory is based on the memory copy/reassign function, which is
used to allow enhanced drawer availability (EDA) and concurrent drawer replacement
(CDR)12. This process was enhanced starting with z13 and z13s to provide more efficiency
and speed to the process without affecting system performance.
z15 T01 implements a faster dynamic memory reallocation mechanism, which is especially
useful during service operations (EDA and CDR). PR/SM controls the reassignment of the
content of a specific physical memory array in one CPC drawer to a physical memory array in
another CPC drawer. To do accomplish this task, PR/SM uses all the available physical
memory in the system. This memory includes the memory that is not in use by the system
that is available but not purchased by the client, and the planned memory options, if installed.
11
Relatively fast to the processor clock frequency
12
In previous IBM Z generations (before z13), these service operations were known as enhanced book availability
(EBA) and concurrent book repair (CBR).
The TLB reduces the amount of time that is required to translate a virtual address to a real
address. This translation is done by dynamic address translation (DAT) when it must find the
correct page for the correct address space. Each TLB entry represents one page. As with
other buffers or caches, lines are discarded from the TLB on a least recently used (LRU)
basis.
The worst-case translation time occurs when a TLB miss occurs and the segment table
(which is needed to find the page table) and the page table (which is needed to find the entry
for the particular page in question) are not in cache. This case involves two complete real
memory access delays plus the address translation delay. The duration of a processor cycle
is much shorter than the duration of a memory cycle, so a TLB miss is relatively costly.
It is preferable to have addresses in the TLB. With 4 K pages, holding all of the addresses for
1 MB of storage takes 256 TLB lines. When 1 MB pages are used, it takes only one TLB line.
Therefore, large page size users have a much smaller TLB footprint.
Large pages allow the TLB to better represent a large working set and suffer fewer TLB
misses by allowing a single TLB entry to cover more address translations.
Users of large pages are better represented in the TLB and are expected to see performance
improvements in elapsed time and processor usage. These improvements are because DAT
and memory operations are part of processor busy time, even though the processor waits for
memory operations to complete without processing anything else in the meantime.
To overcome the processor usage that is associated with creating a 1 MB page, a process
must run for some time. It also must maintain frequent memory access to keep the pertinent
addresses in the TLB.
Short-running work does not overcome the processor usage. Short processes with small
working sets are expected to receive little or no improvement. Long-running work with high
memory-access frequency is the best candidate to benefit from large pages.
Long-running work with low memory-access frequency is less likely to maintain its entries in
the TLB. However, when it does run, few address translations are required to resolve all of the
memory it needs. Therefore, a long-running process can benefit even without frequent
memory access.
Weigh the benefits of whether something in this category must use large pages as a result of
the system-level costs of tying up real storage. A balance exists between the performance of
a process that uses large pages and the performance of the remaining work on the system.
It is easy to assume that increasing the TLB size is a feasible option to deal with TLB-miss
situations. However, this process is not as straightforward as it seems. As the size of the TLB
increases, so does the processor usage that is involved in managing the TLB’s contents.
Correct sizing of the TLB is subject to complex statistical modeling to find the optimal tradeoff
between size and performance.
Main storage can be accessed by all processors, but cannot be shared between LPARs. Any
system image (LPAR) must include a defined main storage size. This defined main storage is
allocated exclusively to the LPAR during partition activation.
The fixed size of the HSA eliminates planning for future expansion of the HSA because the
hardware configuration definition (HCD)/input/output configuration program (IOCP) always
reserves space for the following items:
Six channel subsystems (CSSs)
A total of 15 LPARs in CSSs 1 through 5, and 10 LPARs for the sixth CSS for a total of 85
LPARs
Subchannel set 0 with 63.75-K devices in each CSS
Subchannel set 1 with 64-K devices in each CSS
Subchannel set 2 with 64-K devices in each CSS
Subchannel set 3 with 64-K devices in each CSS
The HSA features sufficient reserved space to allow for dynamic I/O reconfiguration changes
to the maximum capability of the processor.
13 Virtual Flash Memory has replaced IBM zFlash Express. No carry forward of zFlash Express exists.
For z15 T01, IBM VFM provides up to 6.0 TB of virtual flash memory in 512 GB increments.
The minimum is 0, while the maximum is 12 features. The number of VFM features ordered
reduces the maximum orderable memory for the z15 server.
3.7.1 Overview
Logical partitioning is a function that is implemented by the PR/SM on z15. z15 can run in
LPAR mode, or in DPM mode. DPM provides a GUI for PR/SM to manage I/O resources
dynamically.
PR/SM is aware of the processor drawer structure on z15 servers. However, LPARs do not
feature this awareness. LPARs feature resources that are allocated to them from various
physical resources. From a systems standpoint, LPARs have no control over these physical
resources, but the PR/SM functions do have this control.
PR/SM manages and optimizes allocation and the dispatching of work on the physical
topology. Most physical topology that was handled by the operating systems is the
responsibility of PR/SM.
As described in 3.5.9, “Processor unit assignment” on page 126, the initial PU assignment is
done during POR by using rules to optimize cache usage. This step is the “physical” step,
where CPs, zIIPs, IFLs, ICFs, and SAPs are allocated on the processor drawers.
When an LPAR is activated, PR/SM builds logical processors and allocates memory for the
LPAR.
PR/SM assigns all logical processors to one CPC drawer that are packed into chips of that
drawer and cooperates with operating system use of HiperDispatch.
New for z15, all processor types can be dynamically reassigned except IFPs.
Memory allocation changed from the previous IBM Z servers. Partition memory is now
allocated based on processor drawer affinity and striped across processor clusters. For more
information, see “Memory allocation” on page 128.
Processor drawers and logical node level assignments are more important because they
optimize L4 cache usage. Therefore, logical processors from a specific LPAR are packed into
a processor drawer as much as possible.
PR/SM also tries to redispatch a logical processor on the same physical processor to
optimize private cache (L1 and L2) usage.
HiperDispatch
PR/SM and z/OS work in tandem to use processor resources more efficiently. HiperDispatch
is a function that combines the dispatcher actions and the knowledge that PR/SM has about
the topology of the system.
Performance can be optimized by redispatching units of work to the same processor group,
which keeps processes running near their cached instructions and data, and minimizes
transfers of data ownership among processors and processor drawers.
The nested topology is returned to z/OS by the Store System Information (STSI) instruction.
HiperDispatch uses the information to concentrate logical processors around shared caches
(L3 at PU chip level, and L4 at drawer level), and dynamically optimizes the assignment of
logical processors and units of work.
z/OS dispatcher manages multiple queues, called affinity queues, with a target number of
eight processors per queue, which fits well onto a single PU chip. These queues are used to
assign work to as few logical processors as are needed for an LPAR workload. Therefore,
even if the LPAR is defined with many logical processors, HiperDispatch optimizes this
number of processors to be near the required capacity. The optimal number of processors to
be used is kept within a processor drawer boundary, when possible.
Logical partitions
PR/SM enables z15 T01 systems to be initialized for a logically partitioned operation,
supporting up to 85 LPARs. Each LPAR can run its own operating system image in any image
mode, independently from the other LPARs.
An LPAR can be added, removed, activated, or deactivated at any time. Changing the number
of LPARs is not disruptive and does not require a POR. Certain facilities might not be
available to all operating systems because the facilities might have software corequisites.
Each LPAR has the following resources that are the same as a real CPC:
Processors
Called logical processors, they can be defined as CPs, IFLs, ICFs, or zIIPs. They can be
dedicated to an LPAR or shared among LPARs. When shared, a processor weight can be
defined to provide the required level of processor resources to an LPAR. Also, the capping
option can be turned on, which prevents an LPAR from acquiring more than its defined
weight and limits its processor consumption.
LPARs for z/OS can have CP and zIIP logical processors. The logical processor types can
be defined as all dedicated or all shared. The zIIP support is available in z/OS.
Memory
Memory (main storage) must be dedicated to an LPAR. The defined storage must be
available during the LPAR activation; otherwise, the LPAR activation fails.
Reserved storage can be defined to an LPAR, which enables nondisruptive memory
addition to and removal from an LPAR by using the LPAR dynamic storage reconfiguration
(z/OS and z/VM). For more information, see 3.7.5, “LPAR dynamic storage
reconfiguration” on page 142.
Modes of operation
The modes of operation are listed in Table 3-6. All available mode combinations, including
their operating modes and processor types, operating systems, and addressing modes, also
are listed. Only the currently supported versions of operating systems are considered.
CP z/VSE 64-bit
Linux on IBM Z
z/TPF
z/VM
The 64-bit z/Architecture mode has no special operating mode because the architecture
mode is not an attribute of the definable images operating mode. The 64-bit operating
systems are in 31-bit mode at IPL and change to 64-bit mode during their initialization. The
operating system is responsible for taking advantage of the addressing capabilities that are
provided by the architectural mode.
For information about operating system support, see Chapter 7, “Operating system support”
on page 253.
zIIP usage: zIIPs can be defined to General mode or z/VM mode image, as listed in
Table 3-6 on page 135. However, zIIPs are used only by z/OS. Other operating
systems cannot use zIIPs, even if they are defined to the LPAR. z/VM V6R4 and
later support real and virtual zIIPs to guest z/OS systems.
General mode is also used to run the z/TPF operating system on dedicated or shared CPs
CF mode, by loading the CFCC code into the LPAR that is defined as one of the following
types:
– Dedicated or shared CPs
– Dedicated or shared ICFs
Linux only mode to run the following systems:
– A Linux on Z operating system, on either of the following types:
• Dedicated or shared IFLs
• Dedicated or shared CPs
– A z/VM operating system, on either of the following types:
• Dedicated or shared IFLs
• Dedicated or shared CPs
z/VM mode to run z/VM on dedicated or shared CPs or IFLs, plus zIIPs and ICFs
Secure Service Container (SSC) mode LPAR can run on dedicated or shared:
– CPs
– IFLs
All LPAR modes, required characterized PUs, operating systems, and the PU
characterizations that can be configured to an LPAR image are listed in Table 3-7. The
available combinations of dedicated (DED) and shared (SHR) processors are also included.
For all combinations, an LPAR also can include reserved processors that are defined, which
allows for nondisruptive LPAR upgrades.
z/VM CPs, IFLs, z/VM (V6R4 and later) All PUs must be SHR or DED
zIIPs, or
ICFs
SSCa IFLs, or CPs z/VSE Network Appliance IFLs DED or IFLs SHR
or
CPs DED or CPs SHR
a. Secure Service Container
The extra channel subsystem and multiple image facility (MIF) image ID pairs (CSSID/MIFID)
can be later assigned to an LPAR for use (or later removed). This process can be done
through dynamic I/O commands by using the HCD. At the same time, required channels must
be defined for the new LPAR.
Partition profile: Cryptographic coprocessors are not tied to partition numbers or MIF IDs.
They are set up with Adjunct Processor (AP) numbers and domain indexes. These
numbers are assigned to a partition profile of a given name. The client assigns these AP
numbers and domains to the partitions and continues to have the responsibility to clear
them out when their profiles change.
14
In z/OS, this support is available since Version 1 Release 10 (z/OS V1.10), while z/VM supports this addition since
z/VM V5.4, and z/VSE since V4.3. However, z15 supports z/OS 2.1 and later, z/VSE 6.2 and z/VM 6.4 and later.
Swapping of specialty engines and general processors with each other, with spare PUs, or
with both, can occur as the system attempts to compact LPAR configurations into physical
configurations that span the least number of processor drawers.
LPAR dynamic PU reassignment can swap client processors of different types between
processor drawers. For example, reassignment can swap an IFL on processor drawer 1 with a
CP on processor drawer 2. Swaps can also occur between PU chips within a processor
drawer or a node and can include spare PUs. The goals are to pack the LPAR on fewer
processor drawers and also on fewer PU chips, based on the z15 processor drawers’
topology. The effect of this process is evident in dedicated and shared LPARs that use
HiperDispatch.
PR/SM and WLM work together to enforce the capacity that is defined for the group and the
capacity that is optionally defined for each individual LPAR.
Unlike traditional LPAR capping, absolute capping is designed to provide a physical capacity
limit that is enforced as an absolute (versus relative) value that is not affected by changes to
the virtual or physical configuration of the system.
Absolute capping provides an optional maximum capacity setting for logical partitions that is
specified in the absolute processors capacity (for example, 5.00 CPs or 2.75 IFLs). This
setting is specified independently by processor type (namely CPs, zIIPs, and IFLs) and
provides an enforceable upper limit on the amount of the specified processor type that can be
used in a partition.
Absolute capping is ideal for processor types and operating systems that the z/OS WLM
cannot control. Absolute capping is not intended as a replacement for defined capacity or
group capacity for z/OS, which are managed by WLM.
Absolute capping can be used with any z/OS, z/VM, or Linux on Z LPAR (that is running on an
IBM Z server). If specified for a z/OS LPAR, absolute capping can be used concurrently with
defined capacity or group capacity management for z/OS. When used concurrently, the
absolute capacity limit becomes effective before other capping controls.
The implementation provides built-in integrated capabilities that allow advanced virtualization
management on IBM Z servers. With DPM, you can use your Linux and virtualization skills
while taking advantage of the full value of IBM Z hardware, robustness, and security in a
workload optimized environment.
DPM provides facilities to define and run virtualized computing systems by using a
firmware-managed environment that coordinate the physical system resources that are
shared by the partitions. The partitions’ resources include processors, memory, network,
storage, crypto, and accelerators.
DPM provides a new mode of operation for IBM Z servers that provide the following services:
Facilitates defining, configuring, and operating PR/SM LPARs in a similar way to how
someone performs these tasks on another platform.
Lays the foundation for a general IBM Z new user experience.
DPM is not another hypervisor for IBM Z servers. DPM uses the PR/SM hypervisor
infrastructure and provides an intelligent interface on top of it that allows customers to define,
use, and operate the platform virtualization without IBM Z experience or skills.
Operating systems that run as guests of z/VM can use the z/VM capability of implementing
virtual memory to guest virtual machines. The z/VM dedicated real storage can be shared
between guest operating systems.
The z15 storage allocation and usage possibilities, depending on the image mode, are listed
in Table 3-8.
15
1 TB if an I/O drawer is installed in the z13 system (carry forward only). z15, z14 do not support I/O drawer. z/OS
V2.1 or later are supported at the time of this writing.
An LPAR must define an amount of main storage and optionally (if not a CF image), an
amount of expanded storage. Both main storage and expanded storage can feature the
following storage sizes defined:
The initial value is the storage size that is allocated to the partition when it is activated.
The reserved value is another storage capacity beyond its initial storage size that an LPAR
can acquire dynamically. The reserved storage sizes that are defined to an LPAR do not
have to be available when the partition is activated. They are predefined storage sizes to
allow a storage increase, from an LPAR point of view.
Without the reserved storage definition, an LPAR storage upgrade is a disruptive process that
requires the following steps:
1. Partition deactivation.
2. An initial storage size definition change.
3. Partition activation.
The extra storage capacity for an LPAR upgrade can come from the following sources:
Any unused available storage
Another partition that features released storage
A memory upgrade
A concurrent LPAR storage upgrade uses DSR. z/OS uses the reconfigurable storage unit
(RSU) definition to add or remove storage units in a nondisruptive way.
z/VM V6R4 and later releases support the dynamic addition of memory to a running LPAR by
using reserved storage. It also virtualizes this support to its guests. Removing storage from
the z/VM LPAR is disruptive. Removing memory from a z/VM guest is not disruptive to the
z/VM LPAR.
LPAR storage granularity information is required for LPAR image setup and for z/OS RSU
definition. On z15 T01 LPARs are limited to a maximum size of 16 TB of main storage.
However, the maximum amount of memory that is supported by z/OS V2.3 and z/OS V2.4 is
4 TB. For z/VM V6R4, V7R1, and V7R2 the limit is 2 TB.
With dynamic storage reconfiguration, the unused storage does not have to be continuous.
PR/SM dynamically takes offline a storage increment and makes it available to other
partitions when an operating system running on an LPAR releases a storage increment.
16 When defining an LPAR on the HMC, the 2G boundary should still be followed in PR/SM.
For more information about implementing LPAR processor management under IRD, see z/OS
Intelligent Resource Director, SG24-5952.
Figure 3-17 shows a z15 Model T01 system that contains multiple z/OS sysplex partitions. It
contains an internal CF (CF02), a z14 ZR1system that contains a stand-alone CF (CF01),
and a z14 M03 that contains multiple z/OS sysplex partitions.
Parallel Sysplex is an enabling technology that allows highly reliable, redundant, and robust
IBM Z technology to achieve near-continuous availability. A Parallel Sysplex consists of one or
more (z/OS) operating system images that are coupled through one or more Coupling
Facilities.
A correctly configured Parallel Sysplex cluster maximizes availability in the following ways:
Continuous availability: Changes can be introduced, such as software upgrades, one
image at a time, while the remaining images continue to process work. For more
information, see Parallel Sysplex Application Considerations, SG24-6523.
High capacity: 1- 32 z/OS images in a Parallel Sysplex operating as a single system.
Dynamic workload balancing: Because it is viewed as a single logical resource, work can
be directed to any operating system image in a Parallel Sysplex cluster that has available
capacity.
Systems management: The architecture defines the infrastructure to satisfy client
requirements for continuous availability. It also provides techniques for achieving simplified
systems management consistent with this requirement.
Resource sharing: Several base z/OS components use CF shared storage. This
configuration enables sharing of physical resources with significant improvements in cost,
performance, and simplified systems management.
Single logical system: The collection of system images in the Parallel Sysplex is displayed
as a single entity to the operator, user, and database administrator. A single system view
means reduced complexity from operational and definition perspectives.
N-2 support: Multiple hardware generations (normally three) are supported in the same
Parallel Sysplex. This configuration provides for a gradual evolution of the systems in the
Parallel Sysplex without changing all of them simultaneously. Software support for multiple
releases or versions is also supported.
Note: Parallel sysplex coupling and timing links connectivity for z15 (M/T 8561) is
supported to N-2 generation CPCs (z15, z14, z13/z13s).
Through state-of-the-art cluster technology, the power of multiple images can be harnessed
to work in concert on common workloads. The IBM Z Parallel Sysplex cluster takes the
commercial strengths of the platform to improved levels of system management, competitive
price performance, scalable growth, and continuous availability.
Consideration: z15, z14, z14 ZR1, z13, and z13s servers cannot coexist in the same
sysplex with System zEC12 or earlier generation systems.
For CFCC Level 24 enhancements see “Coupling Facility Enhancements with CFCC level
24” on page 118.
CFCC Level 23
CFCC level 23 is delivered on the z14 with driver level 36. CFCC Level 23 introduces the
following enhancements:
Asynchronous cross-invalidate (XI) of CF cache structures. Requires PTF support for
z/OS and explicit data manager support (Db2 V12 with PTFs). Consider the following
points:
– Instead of performing XI signals synchronously on every cache update request that
causes them, data managers can “opt in” for the CF to perform these XIs
asynchronously (and then sync them up with the CF at or before transaction
completion). Data integrity is maintained if all XI signals complete by the time
transaction locks are released.
– Results in faster completion of cache update CF requests, especially with cross-site
distance involved.
– Provides improved cache structure service times and coupling efficiency.
Coupling Facility hang detect enhancements provide a significant reduction in failure
scope and client disruption (CF-level to structure-level), with no loss of FFDC collection
capability:
– When a hang is detected, in most cases the CF confines the scope of the failure to
“structure damage” for the single CF structure the hung command was processing
against, capture diagnostics with a non-disruptive CF dump, and continue operating
without ending or restarting the CF image.
– Provides a significant reduction in failure scope and client disruption (CF-level to
structure-level), with no loss of FFDC collection capability.
Coupling Facility ECR granular latching:
– With this support, most CF list and lock structure ECR processing no longer use
structure-wide latching. It serializes its execution by using the normal structure object
latches that all mainline commands use.
– Eliminates the performance degradation that is caused by structure-wide latching.
– A few “edge conditions” in ECR processing still require structure-wide latching to be
used to serialize them.
– Cache structure ECR processing continues to require and use structure-wide latches
for its serialization.
z14 servers with CFCC Level 23 require z/OS V1R13 or later, and z/VM V6R4 or later for
virtual guest coupling.
CFCC Level 22
CFCC level 22 is delivered on the z14 servers with driver level D32. CFCC Level 22
introduces the following enhancements:
CF Enhancements:
– CF structure encryption
To support an upgrade from one CFCC level to the next, different levels of CFCC can be run
concurrently while the CF LPARs are running on different servers. CF LPARs that run on the
same server share the CFCC level.
z15 servers (CFCC level 24) can coexist in a sysplex with CFCC levels 23, 22, 21, and 20.
If the LPAR that is running the CFCC includes only dedicated processors (CPs or ICFs), the
use of all processor capacity (cycles) is not an issue. However, this configuration can be an
issue if the LPAR that is running the CFCC includes shared processors. We suggest that you
enable thin interrupts on the CF LPAR (Default for z15).
CF structure sizing changes are expected when moving to CFCC Level 24. Always review the
CF structure size by using the CFSizer tool when changing CFCC levels.
For more information about the recommended CFCC levels, see the current exception letter
that is published on IBM Resource Link®.
The interrupt causes a shared logical processor CF partition to be dispatched by PR/SM (if it
is not already dispatched), which allows the request or signal to be processed in a more
timely manner. The CF relinquishes control when work is exhausted or when PR/SM takes
the physical processor away from the logical processor.
You can experience CF response time improvements or more consistent CF response time
when using CFs with shared engines. This improvement can allow more environments with
multiple CF images to coexist in a server, and share CF engines with reasonable
performance.
The response time for asynchronous CF requests can also be improved as a result of the use
of Coupling Thin Interrupts on the z/OS host system, regardless of whether the CF is using
shared or dedicated engines.
This capability allows ICF engines to be shared by several CF images. In this environment, it
provides faster and far more consistent CF service times. It can also provide performance that
is reasonably close to dedicated-engine CF performance if the CF engines are not CF Control
Code Thin Interrupts.
The introduction of Thin Interrupts allows a CF to run by using a shared processor while
maintaining good performance. The shared engine is allowed to be undispatched when no
more work exists, as in the past. The new Thin Interrupt now gets the shared processor that is
dispatched when a command or duplexing signal is presented to the shared engine.
This function saves processor cycles and is an excellent option to be used by a production
backup CF or a testing environment CF. This function is activated by using the CFCC
command DYNDISP ON.
The CPs can run z/OS operating system images and CF images. For software charging
reasons, generally use only ICF processors to run CF images.
The “storage class memory” that is provided by Flash Express adapters is replaced with
memory allocated from main memory (VFM).
VFM is designed to help improve availability and handling of paging workload spikes when
running z/OS. With this support, z/OS is designed to help improve system availability and
responsiveness by using VFM across transitional workload events, such as market openings
and diagnostic data collection. z/OS is also designed to help improve processor performance
by supporting middleware use of pageable large (1 MB) pages.
VFM can also be used in CF images to provide extended capacity and availability for
workloads that use IBM WebSphere MQ Shared Queues structures. The use of VFM can
help availability by reducing latency from paging delays that can occur at the start of the
workday or during other transitional periods. It is also designed to eliminate delays that can
occur when collecting diagnostic data during failures.
Ordered VFM memory reduces the maximum orderable memory for the model.
VFM provides physical memory DIMMs that are needed to support activation of all customer
purchased memory and HSA on a multiple drawer z15 T01 with one drawer done for the
following features:
Scheduled concurrent drawer upgrade, such as memory add
Scheduled concurrent drawer maintenance, such N+1 repair
Concurrent repair of an out of service CPC drawer ‘‘fenced’’ during Activation (POR)
Note: All of these features can be done without VFM. However, all customer purchased
memory is not available for use in most cases. Some work might need to be shut down or
not restarted.
17
For more information about WP102400, see this web page:
http://www-03.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP102400
The information is relocated during CDR in a manner that is identical to the process that was
used for expanded storage. VFM is much simpler to manage (HMC task) and no hardware
repair and verify (no cables and no adapters) are needed. Also, because this feature is part of
internal memory, VFM is protected by RAIM and ECC and can provide better performance
because no I/O to an attached adapter occurs.
Note: Use cases for Flash did not change (for example, z/OS paging and CF shared queue
overflow). Instead, they transparently benefit from the changes in the hardware
implementation.
No option is available for VFM plan ahead. The only option is to always include zVFM plan
ahead when Flexible Memory option is selected.
Secure Service Container (SSC) is an integrated IBM Z appliance and is designed to host
most sensitive client workloads and applications, acting as a highly protected and secured
digital vault, enforcing security by encrypting the whole stack: memory, network and data
(both in-flight and at-rest). Applications running inside SSC are isolated and protected from
outsider and insider threats.
SSC combines hardware, software, and middleware and is unique to IBM Z platform. Though
it is called a container, it should not be confused with purely software open source containers
(such as Kubernetes or Docker).
SSC is a part of the Pervasive Encryption concept that was introduced with IBM z14, which is
aimed at delivering best IBM Security hardware and software enhancements, services, and
practices for 360 degree infrastructure protection.
SSC is a powerful IBM technology for providing the extra protection of the most sensitive
workloads.
IBM Hyper Protect Crypto Solutions family uses the SSC technology as a core layer to
provide hyper protected services. For more information, see the IBM Cloud Hyper Protect
Crypto Services website.
IBM Blockchain Platform can be deployed on IBM Z by using SSC to host the IBM Blockchain
network. For more information, see this web page.
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, Features Max34, Max71, Max108,
Max145, and Max190, which is further identified as IBM z15 Model T01, or z15 T01,
unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, Features Max4, Max13, Max21, Max32,
and Max65, which is further identified as IBM z15 Model T02, or z15 T02, unless
otherwise specified.
In the remainder of this chapter, IBM z15 (z15) refers to both machine types.
I/O cage, I/O drawer, and PCIe I/O drawer are not supported.
Note: Throughout this chapter, the terms adapter and card refer to a PCIe I/O feature that
is installed in a PCIe+ I/O drawer.
The PCIe I/O infrastructure in z15 T01 consists of the following components:
PCIe+ Gen3 dual port fanouts that support 16 GBps I/O bus for CPC drawer connectivity
to the PCIe+ I/O drawers. It connects to the PCIe Interconnect Gen3 in the PCIe+ I/O
drawers.
PCIe Gen3 feature that support coupling links, Integrated Coupling Adapter Short Reach
(ICA SR and ICA SR1.1). The ICA SR and ICA SR1.1 features have two ports, each port
supporting 8 GBps.
The 8U, 16-slot, and 2-domain PCIe+ I/O drawer for PCIe I/O features.
The PCIe standard uses a low-voltage differential serial bus. Two wires are used for signal
transmission, and a total of four wires (two for transmit and two for receive) form a lane of a
PCIe link, which is full-duplex. Multiple lanes can be aggregated into a larger link width. PCIe
supports link widths of 1, 2, 4, 8, 12, 16, and 32 lanes (x1, x2, x4, x8, x12, x16, and x32).
The data transmission rate of a PCIe link is determined by the link width (numbers of lanes),
the signaling rate of each lane, and the signal encoding rule. The signaling rate of one PCIe
Generation 3 lane is eight gigatransfers per second (GTps), which means that nearly 8
gigabits are transmitted per second (Gbps).
Note: I/O infrastructure for z15 is implemented as PCIe Generation 3. The PU chip PCIe
interface is PCIe Generation 4 (x16 @32 GBps), but the CPC I/O Fanout infrastructure
uses externally PCIe Generation 3 connectivity.
A PCIe Gen3 x16 link features the following data transmission rates:
The maximum theoretical data transmission rate per lane:
8 Gbps * 128/130 bit (encoding) = 7.87 Gbps=984.6 MBps
The maximum theoretical data transmission rate per link:
984.6 MBps * 16 (lanes) = 15.75 GBps
Considering that the PCIe link is full-duplex mode, the data throughput rate of a PCIe Gen3
x16 link is 31.5 GBps (15.75 GBps in both directions).
Link performance: The link speeds do not represent the performance of the link. The
performance depends on many factors, including latency through the adapters, cable
lengths, and the type of workload.
PCIe Gen3 x16 links are used in z15 servers for driving the PCIe+ I/O drawers, and for
coupling links for CPC to CPC communications.
Note: Unless specified otherwise, PCIe refers to PCIe Generation 3 in remaining sections
of this chapter.
4.2.1 Characteristics
The z15 T01 I/O subsystem is designed to provide great flexibility, high availability, and the
following excellent performance characteristics:
High bandwidth
Link performance: The link speeds do not represent the performance of the link. The
performance depends on many factors, including latency through the adapters, cable
lengths, and the type of workload.
z15 servers use PCIe Gen3 protocol to drive PCIe+ I/O drawers and CPC to CPC
(coupling) connections. The I/O bus infrastructure data rate of up to 128 GBps per system
(12 PCIe+ Gen3 fanout slots). For more information about coupling link connectivity, see
4.6.4, “Parallel Sysplex connectivity” on page 195.
Connectivity options:
– z15 servers can be connected to an extensive range of interfaces, such as FICON/FCP
for SAN connectivity, OSA features for LAN connectivity and zHyperLink Express for
storage connectivity (low latency compared to FICON).
– For CPC to CPC connections, z15 servers use Integrated Coupling Adapter (ICA SR)
and the Coupling Express Long Reach (CE LR). The Parallel Sysplex InfiniBand is not
supported.
– The 25GbE RoCE Express2.1, 10GbE RoCE Express2.1, 25GbE RoCE Express2,
10GbE RoCE Express2, and 10GbE RoCE Express features provide high-speed
memory-to-memory data exchange to a remote CPC by using the Shared Memory
Communications over RDMA (SMC-R) protocol for TCP (socket-based)
communications.
Concurrent I/O upgrade
You can concurrently add I/O features to z15 servers if unused I/O slot positions are
available.
Concurrent PCIe+ I/O drawer upgrade
Extra PCIe+ I/O drawers can be installed concurrently if free frame slots for the PCIe+ I/O
drawers and PCIe fanouts in the CPC drawer are available.
Dynamic I/O configuration
Dynamic I/O configuration supports the dynamic addition, removal, or modification of the
channel path, control units, and I/O devices without a planned outage.
Pluggable optics:
– The FICON Express16SA, FICON Express16S+ FICON Express16S and FICON
Express8S, OSA Express7S, OSA Express6S, OSA Express5S, RoCE Express2.1,
RoCE Express2, and RoCE Express features include Small Form-Factor Pluggable
(SFP) optics.1 These optics allow each channel to be individually serviced in a fiber
optic module failure. The traffic on the other channels on the same feature can
continue to flow if a channel requires servicing.
1 OSA-Express5S and 6S 1000BASE-T features do not have optics (copper only, RJ45 connectors).
2
Multifiber Termination Push-On.
3 For more information, see this web page: https://cw.infinibandta.org/document/dl/7157
PSU PSU
FSP FSP
PCIe Switch
PCIe Switch
The PCIe+ I/O drawer is a one-sided drawer (all I/O cards on one side, in the rear of the
drawer) that is 8U high. The PCIe+ I/O drawer contains the 16 I/O slots for PCIe features, two
switch cards, and two power supply units (PSUs) to provide redundant power, as shown in
Figure 4-1 on page 158.
The PCIe+ I/O drawer slots numbers are shown in Figure 4-2.
The I/O structure in a z15 CPC is shown in Figure 4-3 on page 160. The PCIe switch card
provides the fanout from the high-speed x16 PCIe host bus to eight individual card slots. The
PCIe switch card is connected to the CPC drawer through a single x16 PCIe Gen 3 bus from
a PCIe fanout card (PCIe+ fanout cards on z15 have two PCIe Gen3 x16 ports/busses/links).
In the PCIe+ I/O drawer, the eight I/O feature cards that directly attach to the switch card
constitute an I/O domain. The PCIe+ I/O drawer supports concurrent add and replace I/O
features with which you can increase I/O capability as needed, depending on the CPC
drawer.
The PCIe slots in a PCIe+ I/O drawer are organized into two I/O domains. Each I/O domain
supports up to eight features and is driven through a PCIe switch card. Two PCIe switch cards
always provide a backup path for each other through the passive connection in the PCIe+ I/O
drawer backplane. During a PCIe fanout card or cable failure, 16 I/O cards in two domains
can be driven through a single PCIe switch card. It is not possible to drive 16 I/O cards after
one of the PCIe switch cards is removed.
The two switch cards are interconnected through the PCIe+ I/O drawer board (Redundant I/O
Interconnect, or RII). In addition, switch cards in same PCIe+ I/O drawer are connected to
PCIe fanouts across clusters in CPC drawer for higher availability.
The RII design provides a failover capability during a PCIe fanout card failure. Both domains
in one of these PCIe+ I/O drawers are activated with two fanouts. The flexible service
processors (FSPs) are used for system control.
The domains and their related I/O slots are shown in Figure 4-2 on page 159.
Each I/O domain supports up to eight features (FICON, OSA, Crypto, and so on.) All I/O
cards connect to the PCIe switch card through the backplane board. The I/O domains and
slots are listed in Table 4-1.
Consideration: On a z15 server, only PCIe+ I/O drawers are supported. No older
generation drawers can be carried forward.
z15 T01 server supports the following PCIe I/O new features that are hosted in the PCIe+ I/O
drawers:
FICON Express16SA
OSA-Express7S 25GbE SR1.1
OSA-Express7S 10GbE
OSA-Express7S GbE
OSA-Express7S 1000BASE-T
25GbE RoCE Express2.1
10GbE RoCE Express2.1
Crypto Express7S (one or two ports)
Coupling Express Long Reach (CE LR)
zHyperLink Express1.1
The z15 CPC drawer I/O infrastructure consists of the following features:
The PCIe+ Generation 3 fanout cards: Two ports per card (feature) that connect to PCIe+
I/O drawers.
ICA SR (ICA SR and ICA SR1.1) fanout cards: Two ports per card (feature) that connect to
other (external) CPCs.
Note: IBM z15 does not support Parallel Sysplex InfiniBand (PSIFB) links.
Unless otherwise noted, ICA SR is used for ICA SR and ICA SR1.1 for the rest of the
chapter.
The PCIe fanouts cards are installed in the rear of the CPC drawers. Each CPC drawer
features 12 PCIe+ Gen3 fanout slots.
The following types of fanout cards are supported by z15 servers. Each CPC drawer fanout
slot can hold one of the following fanouts:
PCIe+ Gen3 fanout card: This dual-port copper fanout provides connectivity to the PCIe
switch card in the PCIe I/O drawer.
Integrated Coupling Adapter (ICA SR): This two-port adapter provides coupling
connectivity between z14 ZR1, z14, z13, and z13s servers, up to 150 meters (492 feet),
@8 GBps link rate.
A 16x PCIe copper cable of 1.5 meters (4.92 feet) to 4.0 meters (13.1 feet) is used for
connection to the PCIe switch card in the PCIe+ I/O drawer. PCIe fanout cards are always
plugged in pairs and provide redundancy for I/O domains within the PCIe+ I/O drawer.
PCIe fanout: The PCIe fanout is used exclusively for I/O and cannot be shared for any
other purpose.
The ICA SR (FC 0172) and ICA SR1.1 (FC 0176) use PCIe Gen3 technology, with x16 lanes
that are bifurcated into x8 lanes for coupling. No performance degradation is expected
compared to the coupling over InfiniBand 12x IFB3 protocol.
Both cards are designed to drive distances up to 150 meters (492 feet) with a link data rate of
8 GBps. ICA SR supports up to four channel-path identifiers (CHPIDs) per port and eight
subchannels (devices) per CHPID.
The coupling links can be defined as shared between images (z/OS) within a CSS. They also
can be spanned across multiple CSSs in a CPC. For ICA SR features, a maximum four
CHPIDs per port can be defined.
When STP (FC 1021) is available, ICA SR coupling links can be defined as timing-only links
to other z15, z14 ZR1, z14, and z13/z13s CPCs.
These two fanouts features are housed in the PCIe+ Gen3 I/O fanout slot on the z15 CPC
drawers. Up 48 ICA SR and ICA SR1.1 features (up to 96 ports) are supported on a z15 T01
system. This configuration enables greater connectivity for short distance coupling on a single
processor node compared to previous Z generations.
OM3 fiber optic can be used for distances up to 100 meters (328 feet). OM4 fiber optic cables
can be used for distances up to 150 meters (492 feet). For more information, see the following
manuals:
Planning for Fiber Optic Links, GA23-1408
IBM Z 8561 Installation Manual for Physical Planning, GC28-7002
Table 4-2 Fanout locations and their AIDs for the CPC drawer (z15 T01)
Fanout CPC0 CPC1 CPC2 CPC3 CPC4
locations Location A10 Location A15 Location A20 Location B10 Location B15
AID(Hex) AID(Hex) AID(Hex) AID(Hex) AID(Hex)
LG01 00 0C 18 24 30
LG02 01 0D 19 25 31
LG03 02 0E 1A 26 32
LG04 03 0F 1B 27 33
LG05 04 10 1C 28 34
LG06 05 11 1D 29 35
LG07 06 12 1E 2A 36
LG08 07 13 1F 2B 37
LG09 08 14 20 2C 38
LG10 09 15 21 2D 39
LG11 0A 16 22 2E 3A
LG12 0B 17 23 2F 3B
Fanout slots
The fanout slots are numbered LG01 - LG12, from left to right, as listed in Table 4-2. All fanout
locations and their AIDs for the CPC drawer are shown for reference only.
Important: The AID numbers that are listed in Table 4-2 are valid only for a new build
system. If a fanout is moved, the AID follows the fanout to its new physical location.
Fanout features that are supported by the z15 server are listed in Table 4-3, which includes
the feature type, feature code, and information about the link supported by the fanout feature.
ICA SR 0172 Coupling link OM4 MTP 150 m (492 ft) 8 GBps
ICA SR1.1 0176 Coupling link OM4 MTP 150 m (492 ft) 8 GBps
4 Certain I/O features do not have external ports, such as Crypto Express.
Coupling links connectivity support: zEC12 and zBC12 are not supported in same
Parallel Sysplex or STP CTN with z15.
A CHPID does not directly correspond to a hardware channel port. Instead, it is assigned to a
PCHID in the hardware configuration definition (HCD) or IOCP.
A PCHID REPORT is created for each new build server and for upgrades on servers. The
report lists all I/O features that are installed, the physical slot location, and the assigned
PCHID. A portion of a sample PCHID REPORT is shown in Example 4-2.
A resource group (RG) parameter is also shown in the PCHID REPORT for native PCIe
features. A balanced plugging of native PCIe features exists between four resource groups
(RG1, RG2, RG3, and RG4).
The preassigned PCHID number of each I/O port relates directly to its physical location (jack
location in a specific slot).
4.6 Connectivity
I/O channels are part of the CSS. They provide connectivity for data exchange between
servers, between servers and external control units (CUs) and devices, or between networks.
For more information about connectivity to external I/O subsystems (for example, disks), see
“Storage connectivity” on page 172.
For more information about communication to LANs, see “Network connectivity” on page 181.
Storage access
OSA-Express featuresc
Coupling features
Cryptographic featuresf
OSA-Express featuresc
Coupling features
Cryptographic featuresf
At least one I/O feature (FICON) or one coupling link feature (ICA SR or CE LR) must be
present in the minimum configuration.
The following features are plugged into a PCIe+ I/O drawer and do not require the definition of
a CHPID and CHPID type:
Each Crypto Express (7S/6S/5S) feature occupies one I/O slot, but does not include a
PCHID type. However, LPARs in all CSSs can access the features. Each Crypto Express
adapter can support up to 85 domains.
Each 25GbE RoCE Express2.1 feature occupies one I/O slot but does not include a
CHPID type. However, LPARs in all CSSs can access the feature. The 25GbE RoCE
Express2.1 can be defined to up to 126 virtual functions (VFs) per feature (port is defined
in z/OS Communications Server). The 25GbE RoCE Express2.1 features support up to 63
VFs per port (up to 126 VFs per feature).
Each 10 RoCE Express2.1 feature occupies one I/O slot but does not include a CHPID
type. The 10GbE RoCE Express2.1 can be defined to up to 126 virtual functions (VFs) per
feature (port is defined in z/OS Communications Server). The 10GbE RoCE Express2.1
features support up to 63 VFs per port (up to 126 VFs per feature).
Each RoCE Express/Express2 feature occupies one I/O slot but does not include a CHPID
type. However, LPARs in all CSSs can access the feature. The 10GbE RoCE Express can
be defined to up to 31 LPARs per feature (port is defined in z/OS Communications
Server). The 25GbE RoCE Express2 and the 10GbE RoCE Express2 features support up
to 63 LPARs per port (up to 126 LPARs per feature).
Each zHyperLink Express/zHyperlink Express2.1 feature occupies one I/O slot but does
not include a CHPID type. However, LPARs in all CSSs can access the feature. The
Cables: All fiber optic cables, cable planning, labeling, and installation are client
responsibilities for new z15 installations and upgrades. Fiber optic conversion kits and
mode conditioning patch cables are not orderable as features on z15 servers. All other
cables must be sourced separately.
The Enterprise Fiber Cabling Services use a proven modular cabling system, the fiber
transport system (FTS), which includes trunk cables, zone cabinets, and panels for servers,
directors, and storage devices. FTS supports Fiber Quick Connect (FQC). FQC feature code
is 7960. The FC 7961 is made of FQC feature, bracket and mounting hardware. The FC 7924
is made of FQC, bracket and mounting hardware and LC Duplex 2 meter (6.6 feet) harness.
They are optional in z15.
Whether you choose a packaged service or a custom service, high-quality components are
used to facilitate moves, additions, and changes in the enterprise to prevent the need to
extend the maintenance window.
The required connector and cable type for each I/O feature on z15 T01 servers are listed in
Table 4-6.
0433 CE LR LC Duplex 9 µm SM
The FICON features provide connectivity between any combination of servers, directors,
switches, and devices (control units, disks, tapes, and printers) in a SAN.
Each FICON Express feature occupies one I/O slot in the PCIe+ I/O drawer. Each feature
includes two ports, each supporting an LC Duplex connector, with one PCHID and one
CHPID that is associated with each port.
Each FICON Express feature uses SFP (SFP+ for FICON Express16SA) optics that allow for
concurrent repairing or replacement for each SFP. The data flow on the unaffected channels
on the same feature can continue. A problem with one FICON Express port does not require
replacement of a complete feature.
Each FICON Express feature also supports cascading, which is the connection of two FICON
Directors in succession. This configuration minimizes the number of cross-site connections
and helps reduce implementation costs for disaster recovery applications, IBM
Geographically Dispersed Parallel Sysplex (GDPS), and remote copy.
z15 servers support 32 K devices per FICON channel for all FICON features.
Each FICON Express channel can be defined independently for connectivity to servers,
switches, directors, disks, tapes, and printers, by using the following CHPID types:
CHPID type FC: The FICON, zHPF, and FCTC protocols are supported simultaneously.
CHPID type FCP: Fibre Channel Protocol that supports attachment to SCSI devices
directly or through Fibre Channel switches or directors.
FICON Express16SA
The FICON Express16SA feature is installed in the PCIe+ I/O drawer. Each of the two
independent ports is capable of 8 Gbps or 16 Gbps. The link speed depends on the capability
of the attached switch or device. The link speed is auto-negotiated, point-to-point, and is
transparent to users and applications.
The following types of FICON Express16SA optical transceivers are supported (no mix on
same card):
FICON Express16SA LX feature, FC 0436, with two ports per feature, supporting LC
Duplex connectors
FICON Express16SA SX feature, FC 0437, with two ports per feature, supporting LC
Duplex connectors
https://www.ibm.com/downloads/cas/US-ENUS120-013-CA/name/US-ENUS120-013-CA.PDF
FICON Express16S+
The FICON Express16S+ feature is installed in the PCIe+ I/O drawer. Each of the two
independent ports is capable of 4 Gbps, 8 Gbps, or 16 Gbps. The link speed depends on the
capability of the attached switch or device. The link speed is auto-negotiated, point-to-point,
and is transparent to users and applications.
The following types of FICON Express16S+ optical transceivers are supported (no mix on
same card):
FICON Express16S+ LX feature, FC 0427, with two ports per feature, supporting LC
Duplex connectors
FICON Express16S+ SX feature, FC 0428, with two ports per feature, supporting LC
Duplex connectors
For more information about supported distances, see Table 4-6 on page 171.
FICON Express16S
The FICON Express16S feature is installed in the PCIe+ I/O drawer. Each of the two
independent ports is capable of 4 Gbps, 8 Gbps, or 16 Gbps. The link speed depends on the
capability of the attached switch or device. The link speed is auto-negotiated, point-to-point,
and is transparent to users and applications.
For more information about supported distances, see Table 4-6 on page 171.
FICON Express8S
The FICON Express8S feature is installed in the PCIe I/O drawer. Each of the two
independent ports is capable of 2 Gbps, 4 Gbps, or 8 Gbps. The link speed depends on the
capability of the attached switch or device. The link speed is auto-negotiated, point-to-point,
and is transparent to users and applications.
For more information about supported distances, see Table 4-6 on page 171.
FICON enhancements
Together with the FICON Express16SA and FICON Express16S+, z15 servers provide
enhancements for FICON in functional and performance aspects with IBM Endpoint Security
solution.
With the IBM DS8870 or newer, z15 servers can extend the use of FEC to the fabric N_Ports
for a completed end-to-end coverage of 16 Gbps FC links. For more information, see the IBM
DS8884 and z13s: A new cost optimized solution, REDP-5327.
A static SAN routing policy normally assigns the ISL routes according to the incoming port
and its destination domain (port-based routing), or the source and destination ports pairing
(device-based routing).
The port-based routing (PBR) assigns the ISL routes statically that is based on “first-come,
first-served” when a port starts a fabric login (FLOGI) to a destination domain. The ISL is
round-robin that is selected for assignment. Therefore, I/O flow from same incoming port to
same destination domain always is assigned the same ISL route, regardless of the
destination port of each I/O. This setup can result in some ISLs overloaded while some are
under-used. The ISL routing table is changed whenever Z server undergoes a power-on-reset
(POR), so the ISL assignment is unpredictable.
Device-based routing (DBR) assigns the ISL routes statically that is based on a hash of the
source and destination port. That I/O flow from same incoming port to same destination is
assigned to same ISL route. Compared to PBR, the DBR is more capable of spreading the
load across ISLs for I/O flow from the same incoming port to different destination ports within
a destination domain.
When a static SAN routing policy is used, the FICON director features limited capability to
assign ISL routes based on workload. This limitation can result in unbalanced use of ISLs
(some might be overloaded, while others are under-used).
The dynamic routing ISL routes are dynamically changed based on the Fibre Channel
exchange ID, which is unique for each I/O operation. ISL is assigned at I/O request time, so
different I/Os from same incoming port to same destination port are assigned different ISLs.
With FIDR, z15 servers feature the following advantages for performance and management in
configurations with ISL and cascaded FICON directors:
Support sharing of ISLs between FICON and FCP (PPRC or distributed)
I/O traffic is better balanced between all available ISLs
Improved use of FICON director and ISL
Easier to manage with a predicable and repeatable I/O performance
z15, z14 ZR1, z14 M0x, z13s, and z13 servers now can read this extra diagnostic data for all
the ports that are accessed in the I/O configuration and make the data available to an LPAR.
For z/OS LPARs that use FICON channels, z/OS displays the data with a new message and
display command. For Linux on Z, z/VM, and z/VSE, and LPARs that use FCP channels, this
diagnostic data is available in a new window in the SAN Explorer tool.
By using the FICON Express16S (or newer) as an FCP channel with NPIV enabled, the
maximum numbers of the following aspects for one FCP physical channel are doubled:
Maximum number of NPIV hosts defined: 64
Maximum number of remote N_Ports communicated: 1024
Maximum number of addressable LUNs: 8192
Concurrent I/O operations: 1528
For more information about operating systems that support NPIV, see “N_Port ID
Virtualization” on page 304.
z15, z14 ZR1, z14 M0x, z13, and z13s servers allow for the modification of these default
assignments, which also allows FCP channels to keep previously assigned WWPNs, even
after being moved to a different slot position. This capability can eliminate the need for
reconfiguration of the SAN in many situations, and is especially helpful during a system
upgrade (FC 0099 - WWPN Persistence).
Note: For more information about the FICON enhancement, see Get More Out of Your IT
Infrastructure with IBM z13 I/O Enhancements, REDP-5134.
For more information about the FICON multi-hop, see the FICON Multihop: Requirements
and Configurations white paper at the IBM Techdocs Library website.
The zHyperLink Express1.1 feature (FC 0451) provides a low latency direct connection
between z15 and DS8k storage system.
The zHyperLink Express1.1 is the result of new business requirements that demand fast and
consistent application response times. It dramatically reduces latency by interconnecting the
z15 directly to I/O Bay of the DS8k by using PCIe Gen3 x 8 physical link (up to 150-meter
[492-foot] distance). A new transport protocol is defined for reading and writing IBM CKD data
records7, as documented in the zHyperLink interface specification.
On z15, zHyperLink Express1.1 card is a new PCIe Gen3 adapter, which installed in the
PCIe+ I/O drawer. HCD definition support was added for new PCIe function type with PORT
attributes.
7 CKD data records are handled by using IBM Enhanced Count Key Data (ECKD) command set.
The zHyperLink Express1.1 is virtualized as a native PCIe adapter and can be shared by
multiple LPARs. Each port can support up to 127 Virtual Functions (VFs), with one or more
VFs/PFIDs being assigned to each LPAR. This configuration gives a maximum of 254 VFs
per adapter. The zHyperLink Express requires the following components:
zHyperLink connector on DS8k I/O Bay
For DS8880 firmware R8.3 or newer, the I/O Bay planar is updated to support the
zHyperLink interface. This update includes the update of the PEX 8732 switch to
PEX8733 that includes a DMA engine for the zHyperLink transfers, and the upgrade from
a copper to optical interface by a CXP connector (provided).
Cable
The zHyperLink Express1.1 uses optical cable with MTP connector. Maximum supported
cable length is 150 meters (492 feet).
The zHyperLink Express feature (FC 0431) provides a low latency direct connection between
z15 and DS8k I/O Port.
The zHyperLink Express is the result of new business requirements that demand fast and
consistent application response times. It dramatically reduces latency by interconnecting the
z15 directly to I/O Bay of the DS8880 by using PCIe Gen3 x 8 physical link (up to 150-meter
[492-foot] distance). A new transport protocol is defined for reading and writing IBM CKD data
records8, as documented in the zHyperLink interface specification.
On z15, zHyperLink Express card is a new PCIe adapter, which installed in the PCIe+ I/O
drawer. HCD definition support was added for new PCIe function type with PORT attributes.
Requirements of zHyperLink
The zHyperLink Express feature is available on z15 servers, and includes the following
requirements:
8 CKD data records are handled by using IBM Enhanced Count Key Data (ECKD) command set.
The zHyperLink Express is virtualized as a native PCIe adapter and can be shared by
multiple LPARs. Each port can support up to 127 Virtual Functions (VFs), with one or more
VFs/PFIDs being assigned to each LPAR. This configuration gives a maximum of 254 VFs
per adapter. The zHyperLink Express requires the following components:
zHyperLink connector on DS8880 I/O Bay
For DS8880 firmware R8.4 or newer, the I/O Bay planar is updated to support the
zHyperLink interface. This update includes the update of the PEX 8732 switch to
PEX8733 that includes a DMA engine for the zHyperLink transfers, and the upgrade from
a copper to optical interface by a CXP connector (provided).
Cable
The zHyperLink Express uses optical cable with MTP connector. Maximum supported
cable length is 150 meters (492 feet).
OSA-Express7S 25 Gigabit Ethernet Short Reach1.1 (SR1.1) feature includes one PCIe
Gen3 adapter and one port per feature. The port supports CHPID types OSD.
The OSA-Express7S 25GbE SR1.1 feature supports the use of an industry standard small
form factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device has
an SR transceiver. The sending and receiving transceivers must be the same (SR to SR).
The OSA-Express7S 25GbE SR1.1 feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 µm multimode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express7S 25GbE SR feature supports the use of an industry standard small form
factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device has an
SR transceiver. The sending and receiving transceivers must be the same (SR-to-SR).
The OSA-Express7S 25GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 µm multimode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The supported OSA-Express7S features are listed in Table 4-5 on page 167.
The OSA-Express7S 10 GbE LR feature supports the use of an industry standard small form
factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device includes
an LR transceiver. The transceivers at both ends must be the same (LR-to-LR).
The OSA-Express7S 10 GbE LR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
OSA-Express7S 10 GbE SR
The OSA-Express7S 10 GbE Short Reach (SR) feature (FC 0445) includes one PCIe Gen3
adapter and one port per feature. The port supports CHPID types OSD. The 10 GbE feature
is designed to support attachment to a multimode fiber 10 Gbps Ethernet LAN or Ethernet
switch that is capable of 10 Gbps. The port can be defined as a spanned channel and shared
among LPARs within and across logical channel subsystems.
The OSA-Express7S 10 GbE SR feature supports the use of an industry standard small form
factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device has an
SR transceiver. The sending and receiving transceivers must be the same (SR to SR).
The OSA-Express7S 10 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
The OSA-Express7S GbE LX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an LX transceiver. The sending and receiving
transceivers must be the same (LX to LX).
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device. If multimode fiber optic cables are
being reused, a pair of Mode Conditioning Patch cables is required, with one cable for each
end of the link.
OSA-Express7S GbE SX
The OSA-Express7S GbE SX feature (FC 0443) includes one PCIe adapter and two ports.
The two ports share a channel path identifier (CHPID types OSD or OSC). The ports support
attachment to a 1 Gbps Ethernet LAN. Each port can be defined as a spanned channel and
shared among LPARs and across logical channel subsystems.
The OSA-Express7S GbE SX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an SX transceiver. The sending and receiving
transceivers must be the same (SX-to-SX).
A multi-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express7S 1000BASE-T Ethernet feature can be configured as CHPID type OSC,
OSD, or OSE. Non-QDIO operation mode requires CHPID type OSE.
OSA-Express6S
The OSA-Express6S feature is installed in the PCIe+ I/O drawer. The following
OSA-Express6S features can be installed on z15 servers (carry forward only):
OSA-Express6S 10 Gigabit Ethernet LR (FC 0424)
The supported OSA-Express6S features are listed in Table 4-5 on page 167.
OSA-Express6S 10 GbE LR
The OSA-Express6S 10 GbE LR feature (FC 0424) includes one PCIe adapter and one port
per feature. On z15, the port supports CHPID type OSD. The 10 GbE feature is designed to
support attachment to a single-mode fiber 10 Gbps Ethernet LAN or Ethernet switch that is
capable of 10 Gbps. The port can be defined as a spanned channel and can be shared
among LPARs within and across logical channel subsystems.
The OSA-Express6S 10 GbE LR feature supports the use of an industry standard small form
factor LC Duplex connector. Ensure that the attaching or downstream device includes an LR
transceiver. The transceivers at both ends must be the same (LR-to-LR).
The OSA-Express6S 10 GbE LR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
OSA-Express6S 10 GbE SR
The OSA-Express6S 10 GbE SR feature (FC 0416) includes one PCIe adapter and one port
per feature. On z15, the port supports CHPID type OSD. The 10 GbE feature is designed to
support attachment to a multimode fiber 10 Gbps Ethernet LAN or Ethernet switch that is
capable of 10 Gbps. The port can be defined as a spanned channel and shared among
LPARs within and across logical channel subsystems.
The OSA-Express6S 10 GbE SR feature supports the use of an industry-standard small form
factor LC Duplex connector. Ensure that the attaching or downstream device has an SR
transceiver. The sending and receiving transceivers must be the same (SR-to-SR).
The OSA-Express6S 10 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 or a 62.5 µm multimode fiber optic cable that ends with an LC Duplex connector is
required for connecting each port on this feature to the selected device.
OSA-Express6S GbE LX
The OSA-Express6S GbE LX feature (FC 0422) includes one PCIe adapter and two ports.
The two ports share a channel path identifier (CHPID type OSD). The ports support
attachment to a 1 Gbps Ethernet LAN. Each port can be defined as a spanned channel and
can be shared among LPARs and across logical channel subsystems.
The OSA-Express6S GbE LX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an LX transceiver. The sending and receiving
transceivers must be the same (LX-to-LX).
OSA-Express6S GbE SX
The OSA-Express6S GbE SX feature (FC 0423) includes one PCIe adapter and two ports.
The two ports share a channel path identifier (CHPID type OSD). The ports support
attachment to a 1 Gbps Ethernet LAN. Each port can be defined as a spanned channel and
shared among LPARs and across logical channel subsystems.
The OSA-Express6S GbE SX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an SX transceiver. The sending and receiving
transceivers must be the same (SX-to-SX).
A multi-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express6S 1000BASE-T Ethernet feature can be configured as CHPID type OSC,
OSD, or OSE. Non-QDIO operation mode requires CHPID type OSE.
The following settings are supported on the OSA-Express6S 1000BASE-T Ethernet feature
port:
Auto-negotiate
100 Mbps half-duplex or full-duplex
1000 Mbps full-duplex
For more information about supported distances, see Table 4-8 on page 188.
OSA-Express5S
The OSA-Express5S feature is installed in the PCIe I/O drawer. The following
OSA-Express5S features can be installed on z14 servers (carry forward only):
OSA-Express5S 10 Gigabit Ethernet LR, FC 0415
OSA-Express5S 10 Gigabit Ethernet SR, FC 0416
OSA-Express5S Gigabit Ethernet LX, FC 0413
OSA-Express5S Gigabit Ethernet SX, FC 0414
OSA-Express5S 1000BASE-T Ethernet, FC 0417
OSA-Express5S 10 GbE LR
The OSA-Express5S 10 GbE LR feature (FC 0415) includes one PCIe adapter and one port
per feature. On z15, the port supports CHPID type OSD.
The OSA-Express5S 10 GbE LR feature supports the use of an industry-standard small form
factor LC Duplex connector. Ensure that the attaching or downstream device includes an LR
transceiver. The transceivers at both ends must be the same (LR-to-LR).
The OSA-Express5S 10 GbE LR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
The 10 GbE feature is designed to support attachment to a multimode fiber 10 Gbps Ethernet
LAN or Ethernet switch that is capable of 10 Gbps. The port can be defined as a spanned
channel and shared among LPARs within and across logical channel subsystems.
The OSA-Express5S 10 GbE SR feature supports the use of an industry standard small form
factor LC Duplex connector. Ensure that the attaching or downstream device includes an SR
transceiver. The sending and receiving transceivers must be the same (SR-to-SR).
The OSA-Express5S 10 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 or a 62.5 µm multimode fiber optic cable that ends with an LC Duplex connector is
required for connecting each port on this feature to the selected device.
For more information about supported distances, see Table 4-8 on page 188.
The OSA-Express5S GbE LX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an LX transceiver. The sending and receiving
transceivers must be the same (LX-to-LX).
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device. If multimode fiber optic cables are
being reused, a pair of Mode Conditioning Patch cables is required, with one cable for each
end of the link.
For more information about supported distances, see Table 4-8 on page 188.
The OSA-Express5S GbE SX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an SX transceiver. The sending and receiving
transceivers must be the same (SX-to-SX).
A multi-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
For more information about supported distances, see Table 4-8 on page 188.
The OSA-Express5S 1000BASE-T Ethernet feature can be configured as CHPID type OSC,
OSD, or OSE. Non-QDIO operation mode requires CHPID type OSE.
The following settings are supported on the OSA-Express5S 1000BASE-T Ethernet feature
port:
Auto-negotiate
100 Mbps half-duplex or full-duplex
1000 Mbps full-duplex
If auto-negotiate is not used, the OSA-Express port attempts to join the LAN at the specified
speed and duplex mode. If this specified speed and duplex mode do not match the speed and
duplex mode of the signal on the cable, the OSA-Express port does not connect.
For more information about supported distances, see Table 4-8 on page 188.
Switch configuration for RoCE Express2.1: If the IBM 25GbE RoCE Express2.1
features are connected to 25GbE switches, the switches must meet the following
requirements:
Global Pause function enabled
Priority flow control (PFC) disabled
No firewalls, no routing, and no IEDN
The 25GbE RoCE Express2.1 feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
10GbE and 25GbE RoCE features should not be mixed in a z/OS SMC-R Link Group.
Mixing same speed RoCE features in the same z/OS SMC-R link group is allowed.
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 25GbE switch or to the 25GbE RoCE Express2.1 feature on the attached server:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
On IBM z15 servers, both ports are supported by z/OS and can be shared by up to 126
partitions (LPARs) per PCHID. The 25GbE RoCE Express2.1 feature uses SR optics and
supports the use of a multimode fiber optic cable that ends with an LC Duplex connector. Both
point-to-point connections and switched connections with an enterprise-class 25GbE switch
are supported.
Switch configuration for RoCE Express2.1: If the IBM 10GbE RoCE Express2.1
features are connected to 10GbE switches, the switches must meet the following
requirements:
Global Pause function enabled
Priority flow control (PFC) disabled
No firewalls, no routing, and no IEDN
The 10GbE RoCE Express2.1 feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
10GbE and 25GbE RoCE features should not be mixed in a z/OS SMC-R Link Group.
Mixing same speed RoCE features in the same z/OS SMC-R link group is allowed.
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 10GbE switch or to the 10GbE RoCE Express2 feature on the attached server:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The 10GbE RoCE Express2.1 feature uses SR optics and supports the use of a multimode
fiber optic cable that ends with an LC Duplex connector. Both point-to-point connections and
switched connections with an enterprise-class 10GbE switch are supported.
On z15, RoCE Express2 and 2.1 support 63 Virtual Functions per port (126 VFs per feature).
The 25GbE RoCE Express2 feature uses SR optics and supports the use of a multimode
fiber optic cable that ends with an LC Duplex connector. Both point-to-point connections and
switched connections with an enterprise-class 25GbE switch are supported.
On z15, RoCE Express2 and 2.1 features support 63 Virtual Functions per port (126 VFs per
feature).
The 25GbE RoCE Express2 feature does not support auto-negotiation to any other speed
and runs in full duplex mode only.
10GbE and 25GbE RoCE features should not be mixed in a z/OS SMC-R Link Group.
Mixing same speed RoCE features in the same z/OS SMC-R link group is allowed.
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 25GbE switch or to the 25GbE RoCE Express2 feature on the attached server:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
On z15 servers, both ports are supported by z/OS and can be shared by up to 126 partitions
(LPARs) per PCHID. The 10GbE RoCE Express2 feature uses SR optics and supports the
use of a multimode fiber optic cable that ends with an LC Duplex connector. Both
point-to-point connections and switched connections with an enterprise-class 10 GbE switch
are supported.
On z15, RoCE Express2 and 2.1 features support 63 Virtual Functions per port (126 VFs per
feature). The RAS was improved and ECC double bit correction added staring with FC 0412.
Switch configuration for RoCE Express2: If the IBM 10GbE RoCE Express2 features
are connected to 10GbE switches, the switches must meet the following requirements:
Global Pause function enabled
Priority flow control (PFC) disabled
No firewalls, no routing, and no IEDN
The 10GbE RoCE Express2 feature does not support auto-negotiation to any other speed
and runs in full duplex mode only.
10GbE and 25GbE RoCE features should not be mixed in a z/OS SMC-R Link Group.
Mixing same speed RoCE features in the same z/OS SMC-R link group is allowed.
The 10GbE RoCE Express is a native PCIe feature. It does not use a CHPID and is defined
by using the IOCP FUNCTION statement or in the hardware configuration definition (HCD).
Both ports are supported by z/OS and can be shared by up to 31 partitions (LPARs) per
PCHID on z15, z14 ZR1, z14 M0x, z13s, and z13.
The 10GbE RoCE Express feature uses SR optics and supports the use of a multimode fiber
optic cable that ends with an LC Duplex connector. Point-to-point connections and switched
connections with an enterprise-class 10 GbE switch are supported.
Switch configuration for RoCE: If the IBM 10GbE RoCE Express features are connected
to 10 GbE switches, the switches must meet the following requirements:
Global Pause function enabled
Priority flow control (PFC) disabled
No firewalls, no routing, and no IEDN
The 10GbE RoCE Express feature does not support auto-negotiation to any other speed
and runs in full duplex mode only.
10GbE and 25GbE RoCE features should not be mixed in a z/OS SMC-R Link Group.
Mixing same speed RoCE features in the same z/OS SMC-R link group is allowed.
The maximum supported unrepeated distance, point-to-point, is 300 meters (984 feet). A
client-supplied cable is required. The following types of cables can be used for connecting the
port to the selected 10 GbE switch or to the 10GbE RoCE Express feature on the attached
server:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector; supports 300 meters (984 feet)
OM2 50-micron multimode fiber optic cable that is rated at 500 MHz-km that ends with an
LC Duplex connector; supports 82 meters (269 feet)
OM1 62.5-micron multimode fiber optic cable that is rated at 200 MHz-km that ends with
an LC Duplex connector; supports 33 meters (108 feet)
SMC-R
SMC-R provides application transparent use of the RoCE-Express feature. This feature
reduces the network overhead and latency of data transfers, which effectively offers the
benefits of optimized network performance across processors.
SMC-D
SMC-D was used with the introduction of the Internal Shared Memory (ISM) virtual PCI
function. ISM is a virtual PCI network adapter that enables direct access to shared virtual
memory, which provides a highly optimized network interconnect for IBM Z intra-CPC
communications.
SMC-D maintains the socket-API transparency aspect of SMC-R so that applications that use
TCP/IP communications can benefit immediately without requiring any application software or
IP topology changes. SMC-D completes the overall SMC solution, which provides synergy
with SMC-R.
SMC-R and SMC-D use shared memory architectural concept, which eliminates the TCP/IP
processing in the data path, yet preserves TCP/IP Qualities of Service for connection
management purposes.
ISM is defined by the FUNCTION statement with a virtual CHPID (VCHID) in hardware
configuration definition (HCD)/IOCDS. Identified by the PNETID parameter, each ISM VCHID
defines an isolated, internal virtual network for SMC-D communication, without any hardware
component required. Virtual adapters are defined by virtual function (VF) statements. Multiple
LPARs can access the same virtual network for SMC-D data exchange by associating their
VF with same VCHID.
Applications that use HiperSockets can realize network latency and CPU reduction benefits
and performance improvement by using the SMC-D over ISM.
z15 servers support up to 32 ISM VCHIDs per CPC. Each VCHID supports up to 255 VFs,
with a total maximum of 8,000 VFs.
HiperSockets
The HiperSockets function of z15 servers provides up to 32 high-speed virtual LAN
attachments.
Traffic can be IPv4 or IPv6, or non-IP, such as AppleTalk, DECnet, IPX, NetBIOS, or SNA.
Layer 2 support helps facilitate server consolidation and can reduce complexity and simplify
network configuration. It also allows LAN administrators to maintain the mainframe network
environment similarly to non-mainframe environments.
Packet forwarding decisions are based on Layer 2 information instead of Layer 3. The
HiperSockets device can run automatic MAC address generation to create uniqueness within
and across LPARs and servers. The use of Group MAC addresses for multicast is supported,
and broadcasts to all other Layer 2 devices on the same HiperSockets networks.
Datagrams are delivered only between HiperSockets devices that use the same transport
mode. A Layer 2 device cannot communicate directly to a Layer 3 device in another LPAR
network. A HiperSockets device can filter inbound datagrams by VLAN identification, the
destination MAC address, or both.
HiperSockets Layer 2 is supported by Linux on Z, and by z/VM for Linux guest use.
z15 supports the HiperSockets Completion Queue function that is designed to allow
HiperSockets to transfer data synchronously (if possible) and asynchronously, if necessary.
This feature combines ultra-low latency with more tolerance for traffic peaks.
With the asynchronous support, data can be temporarily held until the receiver has buffers
that are available in its inbound queue during high volume situations. The HiperSockets
Completion Queue function requires the following minimum applications9:
z/OS V2.2 with PTFs
Linux on Z distributions:
– Red Hat Enterprise Linux (RHEL) 6.2
– SUSE Linux Enterprise Server (SLES) 11 SP2
– Ubuntu server 16.04 LTS
z/VSE V6.2
9
Minimum OS support for z15 can differ. For more information, see Chapter 7, “Operating system support” on
page 253.
In z/VM V6.4 and newer, the virtual switch function transparently bridges a guest virtual
machine network connection on a HiperSockets LAN segment. This bridge allows a single
HiperSockets guest virtual machine network connection to communicate directly with the
following systems:
Other guest virtual machines on the virtual switch
External network hosts through the virtual switch OSA UPLINK port
This section describes coupling link features supported in a Parallel Sysplex in which a z15
can participate.
Coupling links
The type of coupling link that is used to connect a CF to an operating system LPAR is
important. The link performance significantly affects response times and coupling processor
usage. For configurations that extend over large distances, the time that is spent on the link
can be the largest part of the response time.
Attention: Parallel Sysplex supports connectivity between systems that differ by up to two
generations (n-2). For example, an IBM z15 can participate in an IBM Parallel Sysplex
cluster with z14, z14 ZR1, z13, and z13s systems.
However, the IBM z15 and IBM z14 ZR1 do not support InfiniBand connectivity so these
servers support connectivity by using only Integrated Coupling Adapter Short Reach (ICA
SR) and Coupling Express Long Reach (CE LR) features. z15 can connect to z13 and
z13s only if these servers have ICA SR or CE LR coupling features.
Figure 4-4 shows the following supported Coupling Link connections for the z15:
InfiniBand links are supported between z13, z13s and z14 machines
Only ICA SR and CE LR links are supported on z15 and z14 ZR1 machines
The coupling link options that are listed in Table 4-10. Also listed are the coupling link support
for each IBM Z platform. Restrictions on the maximum numbers can apply, depending on the
configuration. Always check with your IBM support team for more information.
ICA SR1.1 Integrated 0176 8 GBps 150 meters 96 N/A N/A N/A N/A
Coupling (492 feet)
Adapter
The maximum number of combined external coupling links (active CE LR, ICA SR links) is
160 per z15 T01 system. z15 systems support up to 384 coupling CHPIDs per CPC. A z15
coupling link support summary is shown in Figure 4-4 on page 196. Consider the following
points:
The maximum supported links depends on the IBM Z model or capacity feature code and
the numbers are marked with an asterisk (*).
z15 ICA SR maximum depends on the number of CPU drawers. A total of 12 PCIe+
fanouts are used per CPU drawer, which gives a maximum of 24 ICA SR ports. The z15
machine maximum ICA SR1.1 and ICA SR ports combined is 96.
For more information about distance support for coupling links, see System z End-to-End
Extended Distance Guide, SG24-8047.
An IC link is a fast coupling link that uses memory-to-memory data transfers. IC links do not
have PCHID numbers, but do require CHPIDs.
IC links are enabled by defining CHPID type ICP. A maximum of 64 IC links can be defined on
an IBM z15 CPC.
The ICA SR & SR1.1 are designed to drive distances up to 150 m and supports a link data
rate of 8 GBps. It is designed to support up to four CHPIDs per port and seven subchannels
(devices) per CHPID.
For more information, see IBM Z Planning for Fiber Optic Links (FICON/FCP, Coupling Links,
and Open System Adapters), GA23-1407 as this web page.
The Coupling Express LR uses 10GbE RoCE technology and is designed to drive distances
up to 10 km (6.21 miles) unrepeated and support a link data rate of 10 Gigabits per second
(Gbps). For distance requirements greater than 10 km (6.21 miles), clients must use a
Wavelength Division Multiplexer (WDM). The WDM vendor must be qualified by IBM Z.
Coupling Express LR is designed to support up to four CHPIDs per port, 32 buffers (that is, 32
subchannels) per CHPID. The Coupling Express LR feature is in the PCIe+ I/O drawer on IBM
z15.
For more information, see IBM Z Planning for Fiber Optic Links (FICON/FCP, Coupling Links,
Open Systems Adapters, and zHyperLink Express), GA23-1408, which is available at this
web page.
10
PCIe+ I/O drawer (FC 4021) is introduced with z15 as-is and built in a 19-inch format. FC 4021 contains 16 I/O
slots. FC 4021 can host up to 16 PCIe I/O features (adapters). The PCIe I/O drawer (4032 on z14) cannot be
carried forward during and MES upgrade to z15. z15 support only PCIe+ I/O drawers.
The ICA SR fanout features provide short-distance connectivity to another z15, z14 ZR1, z14,
z13s, or z13 server.
The CE LR adapter provides long-distance connectivity to another z15, z14 ZR1, z14, z13s,
or z13 server.
The z15 server fanout slots in the CPC drawer provide coupling link connectivity through the
ICA SR fanout cards. In addition to coupling links for Parallel Sysplex, the fanout cards
provide connectivity for the PCIe+ I/O drawer (PCIe+ Gen3 fanout).
To migrate from an older generation machine to a z15 without disruption in a Parallel Sysplex
environment requires that the older machines are no more than n-2 generation (namely, at
least z13) and that they carry enough coupling links to connect to the existing systems while
also connecting to the new machine. N-2 generations rule and enough coupling links are
necessary to allow individual components (z/OS LPARs, CFs) to be shut down and moved to
the target machine and continue connect to the remaining systems.
It is beyond the scope of this book to describe all possible migration scenarios. Always
consult with subject matter experts to help you to develop your migration strategy.
The use of the coupling links to exchange STP messages has the following advantages:
STP can scale with distance by using the same links to exchange STP messages and CF
messages in a Parallel Sysplex. Servers that are exchanging messages over short
distances, such as IFB or ICA SR links, can meet more stringent synchronization
requirements than servers that exchange messages over long IFB LR links, with distances
up to 100 kilometers (62 miles). This advantage is an enhancement over the IBM Sysplex
Timer implementation, which does not scale with distance.
Coupling links also provide the connectivity that is necessary in a Parallel Sysplex.
Therefore, a potential benefit can be realized of minimizing the number of cross-site links
that is required in a multi-site Parallel Sysplex.
Between any two servers that are intended to exchange STP messages, configure each
server so that at least two coupling links exist for communication between the servers. This
configuration prevents the loss of one link from causing the loss of STP communication
between the servers. If a server does not have a CF LPAR, timing-only links can be used to
provide STP connectivity.
This feature provides a secure programming and hardware environment wherein crypto
processes are performed. Each cryptographic coprocessor includes general-purpose
processors, non-volatile storage, and specialized cryptographic electronics, which are all
contained within a tamper-sensing and tamper-responsive enclosure that eliminates all keys
and sensitive data on any attempt to tamper with the device. The security features of the HSM
are designed to meet the requirements of FIPS 140-2, Level 4, which is the highest security
level defined.
The Crypto Express7S (2 port), FC 0898 includes two IBM PCIe Cryptographic Coprocessors
(PCIeCC) per feature. The IBM PCIeCC is a hardware security module (HSM). The Crypto
Express7S (1 port), FC 0899 includes one IBM PCIe Cryptographic Coprocessors (PCIeCC)
per feature. For availability reasons, a minimum of two features is required for the one port
feature. Up to 30 Crypto Express7S (2 port) features are supported on z15 T01. The
maximum number of the one-port features is 16. The total number of HSMs supported on z15
T01 is 60 in a combination of Crypto Express7S (2 port), Crypto Express7S (1 port), Crypto
Express6S, or Crypto Express5S.
The Crypto Express7S feature occupies one I/O slot in a PCIe+ I/O drawer.
Each adapter can be configured as a Secure IBM CCA coprocessor, Secure IBM Enterprise
PKCS #11 (EP11) coprocessor, or accelerator.
The accelerator function is designed for maximum-speed Secure Sockets Layer and
Transport Layer Security (SSL/TLS) acceleration, rather than for specialized financial
applications for secure, long-term storage of keys or secrets. The Crypto Express7S can also
be configured as one of the following configurations:
The Secure IBM CCA coprocessor includes secure key functions with emphasis on the
specialized functions that are required for banking and payment card systems. It is
optionally programmable to add custom functions and algorithms by using User Defined
Extensions (UDX).
A new mode, called Payment Card Industry (PCI) PIN Transaction Security (PTS)
Hardware Security Module (HSM) (PCI-HSM), is available exclusively for Crypto
TKE feature: The Trusted Key Entry (TKE) Workstation feature is required for
supporting the administration of the Crypto Express6S when configured as an
Enterprise PKCS #11 coprocessor or managing the CCA mode PCI-HSM.
When the Crypto Express7S PCI Express adapter is configured as a secure IBM CCA
co-processor, it still provides accelerator functions. However, up to 3x better performance for
those functions can be achieved if the Crypto Express7S PCI Express adapter is configured
as an accelerator.
CCA enhancements include the ability to use triple-length (192-bit) Triple-DES (TDES) keys
for operations, such as data encryption, PIN processing, and key wrapping to strengthen
security. CCA also extended the support for the cryptographic requirements of the German
Banking Industry Committee Deutsche Kreditwirtschaft (DK).
Several features that support the use of the AES algorithm in banking applications also were
added to CCA. These features include the addition of AES-related key management features
and the AES ISO Format 4 (ISO-4) PIN blocks as defined in the ISO 9564-1 standard. PIN
block translation and the use of AES PIN blocks in other CCA callable services are supported.
IBM continues to add enhancements as AES finance industry standards are released.
Each Crypto Express6S feature holds one PCI Express cryptographic adapter. Each adapter
can be configured by the installation as a Secure IBM Common Cryptographic Architecture
(CCA) coprocessor, as a Secure IBM Enterprise Public Key Cryptography Standards (PKCS)
#11 (EP11) coprocessor, or as an accelerator.
The tamper-resistant hardware security module, which is contained on the Crypto Express6S
feature, conforms to the Federal Information Processing Standard (FIPS) 140-2 Level 4
Certification. It supports User Defined Extension (UDX) services to implement cryptographic
functions and algorithms (when defined as an IBM CCA coprocessor).
The following EP11 compliance levels are available (Crypto Express6S and Crypto
Express5S):
FIPS 2009 (default)
FIPS 2011
Each Crypto Express6S feature occupies one I/O slot in the PCIe I/O drawer, and features no
CHPID assigned. However, it includes one PCHID.
Each Crypto Express5S feature holds one PCI Express cryptographic adapter. Each adapter
can be configured by the installation as a Secure IBM CCA coprocessor, as a Secure IBM
Enterprise Public Key Cryptography Standards (PKCS) #11 (EP11) coprocessor, or as an
accelerator.
Each Crypto Express5S feature occupies one I/O slot in the PCIe I/O drawer, and features no
CHPID assigned. However, it includes one PCHID.
All native PCIe features should be ordered in pairs for redundancy. The features are assigned
to one of the four resource groups (RGs) that are running on the IFP according to their
physical location in the PCIe+ I/O drawer, which provides management functions and
virtualization functions.
If two features of the same type are installed, one always is managed by resource group 1
(RG 1) or resource group 3 (RG 3) while the other feature is managed by resource group 2
(RG 2) or resource group 4 (RG 4). This configuration provides redundancy if one of the
features or resource groups needs maintenance or fails.
The IFP and RGs support the following infrastructure management functions:
Firmware update of adapters and resource groups
Error recovery and failure data collection
Diagnostic and maintenance tasks
The channel subsystem directs the flow of information between I/O devices and main storage.
It allows data processing to proceeded concurrently with I/O processing, which relieves data
processors (central processor (CP) and Integrated Facility for Linux [IFL]) of the task of
communicating directly with I/O devices.
The channel subsystem includes subchannels, I/O devices that are attached through control
units, and channel paths between the subsystem and control unites. For more information
about the channel subsystem, see 5.1.1, “Multiple logical channel subsystems”.
The design of IBM Z servers offers considerable processing power, memory size, and I/O
connectivity. In support of the larger I/O capability, the CSS structure is scaled up by
introducing the multiple logical channel subsystem (LCSS) since z990, and multiple
subchannel sets (MSS) since z9.
An overview of the channel subsystem for z15 servers is shown in Figure 5-1. z15 T01
systems are designed to support up to six logical channel subsystems, each with four
subchannel sets and up to 256 channels.
All channel subsystems are defined within a single configuration, which is called I/O
configuration data set (IOCDS). The IOCDS is loaded into the hardware system area (HSA)
during a central processor complex (CPC) power-on reset (POR) to start all of the channel
subsystems.
On z15 T01 systems, the HSA is pre-allocated in memory with a fixed size of 256 GB, which
is in addition to the customer purchased memory. This fixed size memory for HSA eliminates
the requirement for more planning of the initial I/O configuration and pre-planning for future
I/O expansions.
CPC drawer repair: The HSA can be moved from one CPC drawer to a different drawer in
an enhanced availability configuration as part of a concurrent CPC drawer repair (CDR)
action.
The introduction of multiple LCSSs enabled an IBM Z server to have more than one channel
subsystems logically, while each logical channel subsystem maintains the same manner of
I/O processing. Also, a logical partition (LPAR) is now attached to a specific logical channel
subsystem, which makes the extension of multiple logical channel subsystems not apparent
to the operating systems and applications. The multiple image facility (MIF) in the structure
enables resource sharing across LPARs within a single LCSS or across the LCSSs.
The multiple LCSS structure extended the Z servers’ total number of I/O connectivity to
support a balanced configuration for the growth of processor and I/O capabilities.
Note: The phrase channel subsystem has same meaning as logical channel subsystem in
this section, unless otherwise stated.
Subchannels
A subchannel provides the logical appearance of a device to the program and contains the
information that is required for sustaining a single I/O operation. Each device is accessible by
using one subchannel in a channel subsystem to which it is assigned according to the active
IOCDS of the Z server.
In z/Architecture, the first subchannel set of an LCSS can have 63.75 K subchannels (with
0.25 K reserved), with a subchannel set ID (SSID) of 0. By enabling the multiple subchannel
sets, extra subchannel sets are available to increase the device addressability of a channel
subsystem. For more information about multiple subchannel sets, see 5.1.2, “Multiple
subchannel sets” on page 206.
Each channel path in a channel subsystem features a unique 2-digit hexadecimal identifier
that is known as a channel-path identifier (CHPID), which ranges 00 - FF. Therefore, a total of
256 CHPIDs are supported by a CSS, and a maximum of 1536 CHPIDs are available on a
z15 server with six logical channel subsystems.
A port on an I/O feature card features a unique physical channel identifier (PCHID) according
to the physical location of this I/O feature adapter, and the sequence of this port on the
adapter.
In addition, a port on a fanout adapter has a unique adapter identifier (AID), according to the
physical location of this fanout adapter, and the sequence of this port on the adapter.
A CHPID is assigned to a physical port by defining the corresponding PCHID or AID in the I/O
configuration definitions.
Control units
A control unit provides the logical capabilities that are necessary to operate and control an
I/O device. It adapts the characteristics of each device so that it can respond to the standard
form of control that is provided by the CSS.
A control unit can be housed separately or can be physically and logically integrated with the
I/O device, channel subsystem, or within the Z server.
I/O devices
An I/O device provides external storage, a means of communication between
data-processing systems, or a means of communication between a system and its
environment. In the simplest case, an I/O device is attached to one control unit and is
accessible through one or more channel paths that are connected to the control unit.
Each subchannel has a unique four-digit hexadecimal number 0x0000 - 0xFFFF. Therefore, a
single subchannel set can address and access up to 64 K I/O devices.
As with the z13 server, the z15 T01 systems support four subchannel sets for each logical
channel subsystem. It can access a maximum of 255.74 K devices for a logical channel
subsystem and a logical partition and the programs that are running on it.
Subchannel number
The subchannel number is a four-digit hexadecimal number 0x0000 - 0xFFFF, which is
assigned to a subchannel within a subchannel set of a channel subsystem. Subchannels in
each subchannel set are always assigned subchannel numbers within a single range of
contiguous numbers.
With the subchannel numbers, a program that is running on an LPAR (for example, an
operating system) can specify all I/O functions relative to a specific I/O device by designating
a subchannel that is assigned to the I/O devices.
Normally, subchannel numbers are used only in communication between the programs and
the channel subsystem.
Device number
A device number is an arbitrary number 0x0000 - 0xFFFF, which is defined by a system
programmer in an I/O configuration for naming an I/O device. The device number must be
unique within a subchannel set of a channel subsystem. It is assigned to the corresponding
subchannel by channel subsystem when an I/O configuration is activated. Therefore, a
subchannel in a subchannel set of a channel subsystem includes a device number together
with a subchannel number for designating an I/O operation.
The device number provide a means to identify a device, independent of any limitations that
are imposed by the system model, configuration, or channel-path protocols.
A device number also can be used to designate an I/O function to a specific I/O device.
Because it is an arbitrary number, it can easily be fit into any configuration management and
operating management scenarios. For example, a system administrator can set all of the
z/OS systems in an environment to device number 1000 for their system RES volumes.
With multiple subchannel sets, a subchannel is assigned to a specific I/O device by the
channel subsystem with an automatically assigned subchannel number and a device number
that is defined by user. An I/O device can always be identified by an SSID with a subchannel
number or a device number. For example, a device with device number AB00 of subchannel
set 1 can be designated as 1AB00.
Normally, the subchannel number is used by the programs to communicate with the channel
subsystem and I/O device, whereas the device number is used by a system programmer,
operator, and administrator.
For the extra subchannel sets enabled by the MSS facility, each has 65535 subchannels
(64 K minus one) for specific types of devices. These extra subchannel sets are referred as
alternative subchannel sets in z/OS. Also, a device that is defined in an alternative subchannel
set is considered a special device, which normally features a special device type in the I/O
configuration.
Currently, a z15 T01 system that is running z/OS defines the following types of devices in
another subchannel set, with proper APAP or PTF installed:
Alias devices of the parallel access volumes (PAV).
Secondary devices of GDPS Metro Mirror Copy Service (formerly Peer-to-Peer Remote
Copy [PPRC]).
FlashCopy SOURCE and TARGET devices with program temporary fix (PTF) OA46900.
Db2 data backup volumes with PTF OA24142.
The use of another subchannel set for these special devices helps reduce the number of
devices in the subchannel set 0, which increases the growth capability for accessing more
devices.
D IOS,CONFIG(ALL)
IOS506I 11.32.19 I/O CONFIG DATA 340
ACTIVE IODF DATA SET = SYS6.IODF39
CONFIGURATION ID = L06RMVS1 EDT ID = 01
TOKEN: PROCESSOR DATE TIME DESCRIPTION
SOURCE: SCZP501 14-10-31 08:51:47 SYS6 IODF39
ACTIVE CSS: 0 SUBCHANNEL SETS CONFIGURED: 0, 1, 2, 3
CHANNEL MEASUREMENT BLOCK FACILITY IS ACTIVE
LOCAL SYSTEM NAME (LSYSTEM): SCZP501
HARDWARE SYSTEM AREA AVAILABLE FOR CONFIGURATION CHANGES
PHYSICAL CONTROL UNITS 8099
CSS 0 - LOGICAL CONTROL UNITS 3996
SS 0 SUBCHANNELS 54689
SS 1 SUBCHANNELS 58862
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 1 - LOGICAL CONTROL UNITS 4088
SS 0 SUBCHANNELS 65280
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 2 - LOGICAL CONTROL UNITS 4088
SS 0 SUBCHANNELS 65280
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 3 - LOGICAL CONTROL UNITS 4088
SS 0 SUBCHANNELS 65280
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 4 - LOGICAL CONTROL UNITS 4088
SS 0 SUBCHANNELS 65280
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 5 - LOGICAL CONTROL UNITS 4088
SS 0 SUBCHANNELS 65280
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
Figure 5-2 Output for display ios,config(all) command with MSS
By assigning the same CHPID from different LCSSs to the same channel path (for example, a
PCHID), the channel path can be accessed by any LPARs from these LCSSs at the same
time. The CHPID is spanned across those LCSSs. The use of spanned channels paths
decreases the number of channels that are needed in an installation of Z servers.
A sample of channel paths that are defined as dedicated, shared, and spanned is shown in
Figure 5-3.
Partition Partition
1 2
... Partition Partition Partition Partition Partition
14 15 16 17 18
... Partition
5A
CSS0 CSS1
CHPID CHPID
CHPID CHPID CHPID CHPID CHPID CHPID CHPID CHPID CHPID CHPID
03 05
00 01 02 FF 06 00 01 22 FF
Share
... 04
SPAN SPAN ... Share
...
PCHID PCHID PCHID PCHID PCHID PCHID PCHID PCHID PCHID PCHID
10B 10C 10D 20E 20A 145 146 147 158 159
PCHID
120
Channel spanning is supported for internal links (HiperSockets and IC links) and for certain
types of external links. External links that are supported on z15 T01 systems include FICON
Express16SA, FICON Express16S+, FICON Express16S, FICON Express8S,
OSA-Express7S, OSA-Express6S, OSA-Express5S, and Coupling Links.
The definition of LPAR name, MIF image ID, and LPAR ID are used to identify an LPAR by the
channel subsystem to identify I/O functions from different LPARs of multiple LCSSs, which
support the implementation of these dedicated, shared, and spanned paths.
Specified in
CSS0 CSS1 CSS2 CSS3 CSS4 CSS5 HCD / IOCP
Logical Partition Name Logical Partition Name LPAR Name LPAR Name LPAR Name LPAR Name Specified in
HCD / IOCP
TST1 PROD1 PROD2 TST2 PROD3 PROD4 TST3 TST4 PROD5 PROD6 TST55 PROD7 PROD8 TST6
02 04 0A 14 16 1D 22 26 35 3A 44 47 56 5A
Specified in
Image Profile
2 4 A 4 6 D 2 6 5 A 4 7 6 A
Specified in
HCD / IOCP
LPAR name
The LPAR name is defined as partition name parameter in the RESOURCE statement of an I/O
configuration. The LPAR name must be unique across the server.
MIF image ID
The MIF image ID is defined as a parameter for each LPAR in the RESOURCE statement of an
I/O configuration. It ranges 1 - F, and must be unique within an LCSS. However, duplicates
are allowed in different LCSSs.
If a MIF image ID is not defined, an arbitrary ID is assigned when the I/O configuration
activated. The z15 server supports a maximum of six LCSSs, with a total of 85 LPARs that
can be defined.
Each LCSS of a z15 T01 system can support the following numbers of LPARs:
LCSS0 to LCSS4 support 15 LPARs each, and the MIF image ID is 1 - F.
LCSS5 supports 10 LPARs, and the MIF image IDs are 1 - A.
LPAR ID
The LPAR ID is defined by a user in an image activation profile for each LPAR. It is a 2-digit
hexadecimal number 00 - 7F. The LPAR ID must be unique across the server. Although it is
arbitrarily defined by the user, an LPAR ID often is the CSS ID concatenated to its MIF image
ID, which makes the value more meaningful for the system administrator. For example, an
LPAR with LPAR ID 1A defined in that manner means that the LPAR is defined in LCSS1, with
the MIF image ID A.
Note: Certain functions might require specific levels of an operating system, PTFs,
or both.
The configuration file for a new machine or upgrade is also available from IBM Resource
Link. Ask your IBM technical sales representative for the name of the file to download.
The z15 is designed for delivering a transparent and consumable approach that enables
extensive (pervasive) encryption of data in flight and at rest, with the goal of substantially
simplifying data security and reducing the costs that are associated with protecting data while
achieving compliance mandates.
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, which is further identified as IBM z15
Model T01, or z15 T01, unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, which is further identified as IBM z15
Model T02, or z15 T02, unless otherwise specified.
In the remainder of this chapter, IBM z15 (z15) refers to both machine types.
This chapter also introduces the principles of cryptography and describes the implementation
of cryptography in the hardware and software architecture of IBM Z. It also describes the
features that IBM z15 offers. Finally, the chapter summarizes the cryptographic features and
required software.
The functions support new standards and are designed to meet the following compliance
requirements:
Payment Card Industry (PCI) Hardware Security Module (HSM) certification to strength
the cryptographic standards for attack resistance in the payment card systems area.
PCI HSM certification is exclusive for Crypto Express7S and Crypto Express6S.
National Institute of Standards and Technology (NIST) through the Federal Information
Processing Standard (FIPS) standard to implement guidance requirements.
Common Criteria EP11 EAL4.
German Banking Industry Commission (GBIC).
Visa Format Preserving Encryption (VFPE) for credit card numbers.
Enhanced public key Elliptic Curve Cryptography (ECC) for users such as Chrome,
Firefox, and Apple’s iMessage.
Accredited Standards Committee X9 Inc Technical Report-34 (ASC X9 TR-34)
IBM z15 includes standard cryptographic hardware and optional cryptographic features for
flexibility and growth capability. IBM has a long history of providing hardware cryptographic
solutions. This history stretches from the development of the Data Encryption Standard (DES)
in the 1970s to the Crypto Express tamper-sensing and tamper-responding programmable
features.
Crypto Express is designed to meet the US Government’s highest security rating of FIPS
140-2 Level 41. It also meets several other security ratings, such as the Common Criteria for
Information Technology Security Evaluation, the PCI HSM criteria, and the criteria for German
Banking Industry Commission (formerly known as Deutsche Kreditwirtschaft evaluation).
The cryptographic functions include the full range of cryptographic operations that are
necessary for local and global business and financial institution applications. User Defined
Extensions (UDX) allow you to add custom cryptographic functions to the functions that z15
systems offer.
Also, it is necessary to ensure that a message cannot be corrupted (message integrity), while
ensuring that the sender and the receiver really are the persons who they claim to be. Over
time, several methods were used to achieve these objectives, with more or less success.
Many procedures and algorithms for encrypting and decrypting data were developed that are
increasingly complicated and time-consuming.
These goals should all be possible without unacceptable overhead to the communication. The
goal is to keep the system secure, manageable, and productive.
The basic data protection method is achieved by encrypting and decrypting the data, while
hash algorithms, message authentication codes (MACs), digital signatures, and certificates
are used for authentication, data integrity, and non-repudiation.
In other words, the security of a cryptographic system should depend on the security of the
key, so the key must be kept secret. Therefore, the secure management of keys is the primal
task of modern cryptographic systems.
6.2.3 Keys
The keys that are used for the cryptographic algorithms often are sequences of numbers and
characters, but can also be any other sequence of bits. The length of a key influences the
security (strength) of the cryptographic method. The longer the used key, the more difficult it
is to compromise a cryptographic algorithm.
For example, the DES (symmetric key) algorithm uses keys with a length of 56 bits,
Triple-DES (TDES) uses keys with a length of 112 bits, and Advanced Encryption Standard
(AES) uses keys of 128, 192, 256, or 512 bits. The asymmetric key RSA algorithm (named
after its inventors Rivest, Shamir, and Adleman) uses keys with a length of 1024 - 4096 bits.
secure
key
protected
key
clear
key
speed
Figure 6-1 Three levels of protection with three levels of speed
The most prominent one-way algorithms are the Secure Hash Algorithms (SHA).
The cryptographic hardware that is supported on IBM z15 is shown in Figure 6-2. These
features are described in this chapter.
PU SCM
CPC Drawer Each PU is
capable of
having the
CPACF
function
Crypto Express7S
(2port in this picture)
PCIe I/O
drawers
Implemented in every processor unit (PU) or core in a central processor complex (CPC) is a
cryptographic coprocessor that can be used2 for cryptographic algorithms that uses clear
keys or protected keys. For more information, see 6.4, “CP Assist for Cryptographic
Functions” on page 224.
The Crypto Express7S adapter is an HSM that is placed in the PCIe+ I/O drawer of z15. It
also supports cryptographic algorithms by using secret keys. For more information, see 6.5,
“Crypto Express7S” on page 230.
Finally, a TKE workstation is required for entering keys in a secure way into the Crypto
Express7S HSM, which often also is equipped with smart card readers. For more information,
see 6.6, “Trusted Key Entry workstation” on page 243.
This feature is a prerequisite to use CPACF (except for SHA-1, SHA-224, SHA-256,
SHA-384, and SHA-512) and the PCIe Crypto Express features.
These features are optional. The 2-port feature contains two IBM 4769 PCIe
Cryptographic Coprocessors (HSMs), which can be independently defined as
Coprocessor or Accelerator. New feature. Not supported on previous generations Z
systems.
These features are optional. The 1-port feature contains one IBM 4769 PCIe
Cryptographic Coprocessor (HSM), which can be defined as Coprocessor or
Accelerator. New feature. Not supported on previous generations Z systems
This feature is available as a carry forward MES from z14. This feature is optional. Each
feature one IBM 4768 PCIe Cryptographic Coprocessor (HSM). This feature is
supported in z15 and z14.
This feature is available as a carry forward MES from z14 or z13. This feature is optional
and each feature of which contains one IBM 4767 PCIe Cryptographic Coprocessor
(HSM). This feature is supported in z15, z14, z13, and z13s systems.
Included with the TKE tower workstation FC 0088 and the TKE rack-mounted
workstation FC 0087 for z15. Earlier versions of TKE features (feature codes: 0080,
0081, 0085, and 0086) can also be upgraded to TKE 9.2 LIC, if the TKE is assigned to
a z14 or later.
Access to information in the smart card is protected by a PIN. One feature code
includes two smart card readers, two cables to connect to the TKE workstation, and 20
smart cards.
This card allows the TKE to support zones with EC 521 key strength (EC 521 strength
for Logon Keys, Authority Signature Keys, and EP11 signature keys).
When one feature code is ordered, 10 smart cards are included. The order increment
is 1 - 99 (990 blank smart cards).
a. The maximum number of combined features of all types cannot exceed 60 HSMs on a z15 T01.
This means the maximum number for feature code 0898 is 30, for all other (single HSM) types
is 16 when installed exclusively.
A TKE includes support for the AES encryption algorithm with 256-bit master keys and key
management functions to load or generate master keys to the cryptographic coprocessor.
If the TKE workstation is chosen to operate the Crypto Express7S adapter in a z15, TKE
workstation with the TKE 9.2 LIC is required. For more information, see 6.6, “Trusted Key
Entry workstation” on page 243.
Important: Products that include any of the cryptographic feature codes contain
cryptographic functions that are subject to special export licensing requirements by the
United States Department of Commerce. It is your responsibility to understand and adhere
to these regulations when you are moving, selling, or transferring these products.
To access and use the cryptographic hardware devices that are provided by z15, the
application must use an application programming interface (API) that is provided by the
operating system. In z/OS, the Integrated Cryptographic Service Facility (ICSF) provides the
APIs and is managing the access to the cryptographic devices, as shown in Figure 6-3 on
page 224.
APPLICATION
APPLICATION
z/OS Software
LPAR x
LPAR y
LPAR z
ICSF
Software Crypto
(clear key)
TKE
LPAR Trusted Key Entry
(secure key)
Hypervisor HW/SW
HSM Crypto
(secure key & clear key )
ICSF is a software component of z/OS. ICSF works with the hardware cryptographic features
and the Security Server (IBM Resource Access Control Facility [IBM RACF®] element) to
provide secure, high-speed cryptographic services in the z/OS environment. ICSF provides
the APIs by which applications request the cryptographic services, and from the CPACF and
the Crypto Express features.
ICSF transparently routes application requests for cryptographic services to one of the
integrated cryptographic engines (CPACF or a Crypto Express feature), depending on
performance or requested cryptographic function. ICSF is also the means by which the
secure Crypto Express features are loaded with master key values, which allows the
hardware features to be used by applications.
The cryptographic hardware that is installed in z15 determines the cryptographic features and
services that are available to the applications.
The users of the cryptographic services call the ICSF API. Some functions are performed by
the ICSF software without starting the cryptographic hardware features. Other functions result
in ICSF going into routines that contain proprietary IBM Z crypto instructions. These
instructions are run by a CPU engine and result in a work request that is generated for a
cryptographic hardware feature.
This cryptographic coprocessor, which is known as the CPACF, is not qualified as an HSM;
therefore, it is not suitable for handling algorithms that use secret keys. However, the
coprocessor can be used for cryptographic algorithms that use clear keys or protected keys.
The CPACF works synchronously with the PU, which means that the owning processor is
busy when its coprocessor is busy. This setup provides a fast device for cryptographic
services.
CPACF supports pervasive encryption. Simple policy controls allow businesses to enable
encryption to protect data in mission-critical databases without the need to stop the database
or re-create database objects. Pervasive encryption includes z/OS Dataset Encryption, z/OS
Coupling Facility Encryption, z/VM encrypted hypervisor paging, and z/TPF transparent
database encryption, which use performance enhancements in the hardware.
The CPACF offers a set of symmetric cryptographic functions that enhances the encryption
and decryption performance of clear key operations. These functions are for SSL, virtual
private network (VPN), and data-storing applications that do not require FIPS 140-2 Level 4
security.
CPACF is designed to facilitate the privacy of cryptographic key material when used for data
encryption through key wrapping implementation. It ensures that key material is not visible to
applications or operating systems during encryption operations. For more information, see
6.4.2, “CPACF protected key” on page 227.
The CPACF feature provides hardware acceleration for the following cryptographic services:
DES
Triple-DES
AES-128
AES-192
AES-256 (all for clear and protected keys)
SHA-1
SHA-256 (SHA-2 or SHA-3 standard)
SHA-384 (SHA-2 or SHA-3 standard)
SHA-512 (SHA-2 or SHA-3 standard)
SHAKE-128
SHAKE-256
PRNG
DRNG
TRNG
These functions are provided as problem-state z/Architecture instructions that are directly
available to application programs. These instructions are known as Message-Security Assist
(MSA). When enabled, the CPACF runs at processor speed for every CP, IFL, and zIIP. For
more information about MSA instructions, see z/Architecture Principles of Operation,
SA22-7832.
For activating these functions, the CPACF must be enabled by using feature code (FC) 3863,
which is available for no extra charge. Support for hashing algorithms SHA-1, SHA-256,
SHA-384, and SHA-512 is always enabled.
The CPACF coprocessor in z14 was redesigned for improved performance compared to the
z13 and was further improved in z15, depending on the function that is being used. The
following tools might benefit from the throughput improvements:
Db2/IMS encryption tool
Db2 built-in encryption
z/OS Communication Server: IPsec/IKE/AT-TLS
z/OS System SSL
z/OS Network Authentication Service (Kerberos)
DFDSS Volume encryption
z/OS Java SDK
z/OS Encryption Facility
Linux on Z: Kernel, openSSL, openCryptoki, and GSKIT
For the SHA hashing algorithms and the random number generation algorithms, only clear
keys are used. For the symmetric encryption and decryption DES and AES algorithms and
clear keys, protected keys can also be used. Protected keys require a Crypto Express7S,
Crypto Express6S, or a Crypto Express5S adapter that is running in CCA mode. For more
information, see 6.5.2, “Crypto Express7S as a CCA coprocessor” on page 233.
The hashing algorithms SHA-1, SHA-2, and SHA-3 support for SHA-224, SHA-256,
SHA-384, and SHA-512, are enabled on all systems and do not require the CPACF
enablement feature. For all other algorithms, the no-charge CPACF enablement feature (FC
3863) is required.
The CPACF functions are implemented as processor instructions and require operating
system support for use. Operating systems that use the CPACF instructions include z/OS,
z/VM, z/VSE, z/TPF, and Linux on Z.
Clear keys process faster than secure keys because the process is done synchronously on
the CPACF. Protected keys blend the security of Crypto Express7S, Crypto Express6S, or
Crypto Express5S coprocessors and the performance characteristics of the CPACF. This
process allows it to run closer to the speed of clear keys.
CPACF facilitates the continued privacy of cryptographic key material when used for data
encryption. In Crypto Express7S, Crypto Express6S, or Express5S coprocessors, a secure
key is encrypted under a master key. However, a protected key is encrypted under a wrapping
key that is unique to each LPAR.
Because the wrapping key is unique to each LPAR, a protected key cannot be shared with
another LPAR. By using key wrapping, CPACF ensures that key material is not visible to
applications or operating systems during encryption operations.
Wrapping keys are generated during the clear reset each time an LPAR is activated or reset.
No customizable option is available at Support Element (SE) or Hardware Management
Console (HMC) that permits or avoids the wrapping key generation. This function flow for the
Crypto Express7S and Crypto Express6S adapters is shown in Figure 6-5.
Secure Key
8 ICSF 1 DKCCAMK CKDS
Software
Hardware + Firmware
CPACF
CPACF Wrapping Key (CPACFWK)
Figure 6-5 CPACF key wrapping for Crypto Express7S and Crypto Express6S
The key wrapping for Crypto Express5S is similar to Crypto Express7S or Crypto Express6S;
however, the Data Key that is exchanged between the Crypto Express5S and the CPACF is
not wrapped by way of a Transport Key.
The CPACF Wrapping Key and the Transport Key for use with Crypto Express7S or Crypto
Express6S are in a protected area of the HSA that is not visible to operating systems or
applications.
Secure Key
8 ICSF 1 DKCCAMK CKDS
CPACF
CPACF Wrapping Key (CPACFWK)
6
DK => DKCPACFWK
A new segment in the profiles of the CSFKEYS class in IBM RACF restricts which secure
keys can be used as protected keys. By default, all secure keys are considered not eligible to
be used as protected keys. The process that is shown in Figure 6-5 on page 228 considers a
secure key as the source of a protected key.
The source key in this case is stored in the ICSF Cryptographic Key Data Set (CKDS) as a
secure key, which was encrypted under the master key. This secure key is sent to CEX7C,
CEX6C, or CEX5C to be deciphered and then, sent to the CPACF in clear text. At the CPACF,
the key is wrapped under the LPAR wrapping key, and is then returned to ICSF. After the key
is wrapped, ICSF can keep the protected value in memory. It then passes it to the CPACF,
where the key is unwrapped for each encryption or decryption operation.
The protected key is designed to provide substantial throughput improvements for a large
volume of data encryption and low latency for encryption of small blocks of data. A
high-performance secure key solution, also known as a protected key solution, requires the
ICSF HCR7770 as a minimum release.
Each Crypto Express7S PCI Express adapter (HSM) is available in one of the following
configurations:
Secure IBM CCA coprocessor (CEX7C) for FIPS 140-2 Level 4 certification. This
configuration includes secure key functions. It is optionally programmable to deploy more
functions and algorithms by using UDX. For more information, see 6.5.2, “Crypto
Express7S as a CCA coprocessor” on page 233.
A TKE workstation is required to support the administration of the Crypto Express7S when
it is configured in CCA mode when in full PCI-compliant mode for the necessary certificate
management in this mode. The TKE is optional in all other use cases for CCA.
Secure IBM Enterprise PKCS #11 (EP11) coprocessor (CEX7P) implements an
industry-standardized set of services that adheres to the PKCS #11 specification V2.20
and more recent amendments. It was designed for extended FIPS and Common Criteria
evaluations to meet public sector requirements. This new cryptographic coprocessor
mode introduced the PKCS #11 secure key function. For more information, see 6.5.3,
“Crypto Express7S as an EP11 coprocessor” on page 239.
A TKE workstation is always required to support the administration of the Crypto
Express7S when it is configured in EP11 mode.
Accelerator (CEX7A) for acceleration of public key and private key cryptographic
operations that are used with SSL/TLS processing. For more information, see 6.5.4,
“Crypto Express7S as an accelerator” on page 240.
These modes can be configured by using the SE. The PCIe adapter must be configured
offline to change the mode.
Attention: Switching between configuration modes erases all adapter secrets. The
exception is when you are switching from Secure CCA to accelerator, and vice versa.
The Crypto Express7S feature is released for enhanced cryptographic performance. Clients
who migrated to variable-length AES key tokens cannot take advantage of faster encryption
speeds by using CPACF. Support is being added to translate a secure variable-length AES
CIPHER token to a protected key token (protected by the system wrapping key). This support
allows for faster AES encryption speeds when variable-length tokens are used while
maintaining strong levels of security.
The Crypto Express7S feature does not include external ports and does not use optical fiber
or other cables. It does not use channel path identifiers (CHPIDs), but requires one slot in the
PCIe I/O drawer and one physical channel ID (PCHID) for each PCIe cryptographic adapter.
Removal of the feature or adapter zeroizes its content. Access to the PCIe cryptographic
adapter is controlled through the setup in the image profiles on the SE.
Each z15 T01 supports up to 60 HSMs in total (combination of Crypto Express7S (2port)
Crypto Express7S (1 port), Crypto Express6S, and Crypto Express5S). Crypto Express6S
and Crypto Express5S features have one HSM (port) and are not orderable for a new build
z15 T01 system, but can be carried forward from a z14 or z13 by using an MES. Configuration
information for Crypto Express7S is listed in Table 6-2.
The concept of dedicated processor does not apply to the PCIe cryptographic adapter.
Whether configured as a coprocessor or an accelerator, the PCIe cryptographic adapter is
made available to an LPAR. It is made available as directed by the domain assignment and
the candidate list in the LPAR image profile. This availability is not changed by the shared or
dedicated status that is given to the PUs in the partition.
The definition of domain indexes and PCIe cryptographic adapter numbers in the candidate
list for each LPAR must be planned to allow for nondisruptive changes. Consider the following
points:
Operational changes can be made by using the Change LPAR Cryptographic Controls
task from the SE, which reflects the cryptographic definitions in the image profile for the
The same PCIe cryptographic adapter number and usage domain index combination can be
defined for more than one LPAR (up to 85 for z15). For example, you might define a
configuration for backup situations. However, only one of the LPARs can be active at a time.
For more information, see 6.5.5, “Managing Crypto Express7S” on page 240.
Several of these algorithms require a secure key and must run on an HSM. Some of these
algorithms can also run with a clear key on the CPACF. Many standards are supported only
when Crypto Express7S is running in CCA mode. Others are supported only when the
adapter is running in EP11 mode.
The three modes for Crypto Express6S are described next. For more information, see 6.7,
“Cryptographic functions comparison” on page 249.
UDX is supported under a special contract through an IBM or approved third-party service
offering. The Crypto Cards website directs your request to an IBM Global Services location
for your geographic location. A special contract is negotiated between IBM Global Services
and you for the development of the UDX code by IBM Global Services according to your
specifications and an agreed-upon level of the UDX.
A UDX toolkit for IBM Z is tied to specific versions of the CCA code and the related host code.
UDX is available for the Crypto Express7S and Crypto Express6S (Secure IBM CCA
coprocessor mode only) features. An UDX migration is no more disruptive than a normal
Microcode Change Level (MCL) or ICSF release migration.
In z15, up to four UDX files can be imported. These files can be imported from a USB media
stick or an FTP server. The UDX configuration window is updated to include a Reset to IBM
Default button.
Consideration: CCA features a new code level starting with z13 systems, and the UDX
clients require a new UDX.
On z15 T01, Crypto Express7S is delivered with CCA Level 6.3 firmware. A new set of
cryptographic functions and callable services is provided by the IBM CCA LIC to enhance the
functions that secure financial transactions and keys. The Crypto Express7S includes the
following features:
Greater than 16 domains support up to 85 LPARs on z15 T01.
Payment Card Industry (PCI) PIN Transaction Security (PTS) HSM Certification that is
exclusive to z15 in combination with CEX7S or CEX6S features, and z14 and CEX6S
features.
VFPE support, which was introduced with z13/z13s systems.
AES PIN support for the German banking industry.
PKA Translate UDX function into CCA.
Verb Algorithm Currency.
Starting with z14, the IBM Z crypto architecture can support up to 256 domains in an adjunct
processor (AP) with the AP extended addressing (APXA) facility that is installed. As such, the
Crypto Express adapters are enhanced to handle 256 domains. The IBM Z firmware provides
up to 85 domains for z15 to customers (to match the current LPAR maximum). Customers can
map individual LPARs to unique crypto domains or continue to share crypto domains across
LPARs.
Compliance with the PCI-HSM standard is valuable for customers, particularly those
customers who are in the banking and finance industry. This certification is important to
clients for the following fundamental reasons:
Compliance is increasingly becoming mandatory.
The requirements in PCI-HSM make the system more secure.
If you are a bank, acquirer, processor, or other participant in the payment card systems, the
card brands can impose requirements on you if you want to process their cards. One set of
requirements they are increasingly enforcing is the PCI standards.
The card brands work with PCI in developing these standards, and they focused first on the
standards they considered most important, particularly the PCI Data Security Standard
(PCI-DSS). Some of the other standards were written or required later, and PCI-HSM is one
of the last standards to be developed. In addition, the standards themselves were increasing
the strength of their requirements over time. Some requirements that were optional in earlier
versions of the standards are now mandatory.
In general, the trend is for the card brands to enforce more of the PCI standards and to
enforce them more rigorously. The trend in the standards is to impose more and stricter
requirements in each successive version. The net result is that companies subject to these
requirements can expect that they eventually must comply with all of the requirements.
The result of these requirements is that applications and procedures often must be updated
because they used some of the things that are now prohibited. Although this issue is
inconvenient and imposes some costs, it does increase the resistance of the systems to
attacks of various kinds. Updating a system to use PCI-HSM compliant HSMs is expected to
reduce the risk of loss for the institution and its clients.
VFPE allows customers to add encryption to their applications in such a way that the
encrypted data can flow through their systems without requiring a massive redesign of their
application. In our example, if the credit card number is VFPE-encrypted at the point of entry,
the cipher text still behaves as a credit card number. It can flow through business logic until it
meets a back-end transaction server that can VFPE-decrypt it to get the original credit card
number to process the transaction.
Note: VFPE technology forms part of Visa, Inc.’s, Data Secure Platform (DSP). The use of
this function requires a service agreement with Visa. You must maintain a valid service
agreement with Visa when you use DSP/FPE.
This support includes PIN method APIs, PIN administration APIs, new key management
verbs, and new access control points support that is needed for DK-defined functions.
UDX is integrated into the base CCA code to support translating an external RSA CRT key
into new formats. These formats use tags to identify key components. Depending on which
new rule array keyword is used with the PKA Key Translate callable service, the service TDES
encrypts those components in CBC or ECB mode. In addition, AES CMAC support is
delivered.
7
CCA 5.4 and 6.1 enhancements are also supported for z/OS V2R1 with ICSF HCR77C1 (WD17) with Small
Program Enhancements (SPEs) (z/OS continuous delivery model).
Note: Although older IBM Z and operating systems also are supported, they are
beyond the scope of this IBM Redbooks publication.
The secure IBM Enterprise PKCS #11 (EP11) coprocessor runs the following tasks:
Encrypt and decrypt (AES, DES, TDES, and RSA)
Sign and verify (DSA, RSA, and ECDSA)
Generate keys and key pairs (DES, AES, DSA, ECC, and RSA)
HMAC (SHA1, SHA2 or SHA3 [SHA224, SHA256, SHA384, and SHA512])
Digest (SHA1, SHS2 or SHA3 [SHA224, SHA256, SHA384, and SHA512])
Wrap and unwrap keys
Random number generation
Get mechanism list and information
Attribute values
Key Agreement (Diffie-Hellman)
When defined in EP11 mode, the TKE workstation is required to manage the Crypto
Express6S feature.
z/OS V2.2 and V2.3 require ICSF Web Deliverable WD18 (HCR77D0) to support the
following new features:
EP11 Stage 4:
– New elliptic curve algorithms for PKCS#11 signature, key derivation operations:
• Ed448 elliptic curve
• EC25519 elliptic curve
– EP11 Concurrent Patch Apply: Allows service to be applied to the EP11 coprocessor
dynamically without taking the crypto adapter offline (already available for CCA
coprocessors).
– eIDAS compliance: eIDAS: Cross-border EU regulation for portable recognition of
electronic identification.
FIPS 140-2 certification is not relevant to the accelerator because it operates with clear keys
only. The function extension capability through UDX is not available to the accelerator.
The functions that remain available when the Crypto Express6S feature is configured as an
accelerator are used for the acceleration of modular arithmetic operations. That is, the RSA
cryptographic operations are used with the SSL/TLS protocol. The following operations are
accelerated:
PKA Decrypt (CSNDPKD) with PKCS-1.2 formatting
PKA Encrypt (CSNDPKE) with zero-pad formatting
Digital Signature Verify
The RSA encryption and decryption functions support key lengths of 512 - 4,096 bits in the
Modulus-Exponent (ME) and Chinese Remainder Theorem (CRT) formats.
Each Crypto Express7S FC 0898 includes two IBM 4769 PCIe Cryptographic Coprocessors
(PCIeCC - which is a hardware security module - HSM); FC 0899 includes one IBM 4769
PCIeCC. The adapters are available in the following configurations:
IBM Enterprise Common Cryptographic Architecture (CCA) Coprocessor (CEX7C)
IBM Enterprise Public Key Cryptography Standards #11 (PKCS) Coprocessor (CEX7P)
IBM Crypto Express7S Accelerator (CEX7A)
During the feature installation, the PCI-X adapter is configured by default as the CCA
coprocessor.
The Crypto Express7S feature does not use CHPIDs from the channel subsystem pool.
However, the Crypto Express7S feature requires one slot in a PCIe I/O drawer, and one
PCHID for each PCIe cryptographic adapter.
For enabling an LPAR to use a Crypto Express7S adapter, the following cryptographic
resources in the image profile must be defined for each partition:
Usage domain index
Control domain index
PCI Cryptographic Coprocessor Candidate List
PCI Cryptographic Coprocessor Online List
This task is accomplished by using the Customize/Delete Activation Profile task, which is in
the Operational Customization Group, from the HMC or from the SE. Modify the
cryptographic initial definition from the Crypto option in the image profile, as shown in
Figure 6-7 on page 242.
Important: After this definition is modified, any change to the image profile requires a
DEACTIVATE and ACTIVATE of the logical partition for the change to take effect.
Therefore, this cryptographic definition is disruptive to a running system.
After they are activated, the active partition cryptographic definitions can be viewed from the
HMC. Select the CPC, and click View LPAR Cryptographic Controls in the CPC
Operational Customization window. The resulting window displays the definition of Usage and
Control domain indexes, and PCI Cryptographic candidate and online lists, as shown in
Figure 6-8. Information is provided for active logical partitions only.
Operational changes can be made by using the Change LPAR Cryptographic Controls task,
which reflects the cryptographic definitions in the image profile for the partition. With this
function, the cryptographic feature can be added and removed dynamically, without stopping
a running operating system.
For more information about the management of Crypto Express7S, see IBM z15
Configuration Setup, SG24-8860.
A TKE workstation is part of a customized solution for the use of the Integrated Cryptographic
Service Facility for z/OS (ICSF for z/OS) or Linux for IBM Z. This program provides a basic
key management system for the cryptographic keys of a z15 system that has Crypto Express
features installed.
The TKE provides a secure, remote, and flexible method of providing Master Key Part Entry,
and to remotely manage PCIe cryptographic coprocessors. The cryptographic functions on
the TKE run by one PCIe cryptographic coprocessor. The TKE workstation communicates
with the IBM Z system through a TCP/IP connection. The TKE workstation is available with
Ethernet LAN connectivity only. Up to 10 TKE workstations can be ordered.
TKE FCs 0087 and 0088 can be used to control any supported Crypto Express feature
supported on z15. They also can be used to control the Crypto Express6S, Crypto Express5S
on z14, Crypto Express5S on z13 and z13s systems, and the Crypto adapters on older, still
supported systems.
The TKE 9.2 LIC (FC 0881) features the following enhancements over the described
functions of LIC 9.1 and 9.0:
TKE 9.2 LIC is enhanced with functions to support the management of the Crypto
Express7S 4769 host crypto module.
TKE 9.2 allow Host Transaction Programs to run over a TLS connection for z/OS LPARs.
The TKE is checking whether TLS is configured on the Host Transaction Program port and
automatically uses TLS when communicating with the host.
TKE 9.2 can handle a combination of AT-TLS configured hosts with those hosts that are
not TLS capable.
The TKE 9.2. supports AES Operational Key parts to be tagged as PCI-compliant. With
CCA 6.3, support for marking some AES operational key parts as being PCI-compliant is
introduced. The TKE is mandatory to support the PCI-compliant environment for
CCA-mode.
A new Stronger encryption is used when negotiating the session key to an EP11 Domain.
When loading key parts into an EP11 host domain, a session key is derived by the smart
card and the target domain. The BLUE TKE smart cards (00RY790) includes 521-bit EC
capability. The 521-bit EC strength is used during the EP11 session key derivation
process. For CCA-mode, this stronger encryption was made available with TKE 9.1 LIC.
TKE 9.2 supports 32 character user IDs when the TKE Host Transaction Program is on a
LINUX system.
The following enhancements were introduced with TKE 9.1 LIC (FC 0880):
TKE 9.1 License Internal Code enhancements for support EC521 strength TKE and
Migration zones. An EC521 Migration zone is required if you want to use the migration
wizard to collect and apply PCI-compliant domain information.
TKE 9.1 also has a new family of wizards that makes it easy to create EC521 zones on all
of its smart cards. This feature simplifies the process of deploying a TKE for the first time
or moving data from a weaker TKE zone to a new EC521 zone.
A new smart card for the TKE allows stronger Elliptic Curve Cryptography (ECC) levels.
Other TKE Smart Cards (FC 0900, packs of 10, FIPS certified blanks) require TKE 9.1
LIC.
The following features are related to support for the Crypto Express6S with CCA 6.0. The
Crypto Express6S with CCA 6.0 is designed to meet the PCI-HSM PIN Transaction Security
v3.0, 2016 standard:
Domain mode management
With CCA 6.0, individual domains are in one of the following modes:
– Normal Mode
– Imprint Mode
– Compliant Mode
Tip: For more information about handling a TKE, see the TKE Introduction video.
Each LPAR in the same system that uses a domain that is managed through a TKE
workstation connection is a TKE host or TKE target. An LPAR with a TCP/IP connection to the
TKE is referred to as the TKE host; all other partitions are TKE targets.
The cryptographic controls that are set for an LPAR through the SE determine whether the
workstation is a TKE host or a TKE target.
If smart cards with applets that are not supported by the new smart card reader are reused,
new smart cards on TKE 8.1 or later must be created and the content from the old smart
cards to the new smart cards must be copied. The new smart cards can be created and
copied on a TKE 8.1 system. If the copies are done on TKE 9.0, the source smart card must
be placed in an older smart card reader from feature code 0885 or 0891.
A new smart card for the Trusted Key Entry (TKE) allows stronger Elliptic Curve Cryptography
(ECC) levels. More TKE Smart Cards (FC 0900, packs of 10, FIPS certified blanks) require
TKE 9.1 LIC or up.
Note: Several options for ordering the TKE with or without ordering Keyboard, Mouse, and
Display are available. Ask your IBM Representative for more information about which
option is the best option for you.
The TKE 9.2 LIC requires the 4768 crypto adapter. The TKE 8.x and TKE 7.3 workstations
can be upgraded to the TKE 9.2 tower workstation by purchasing a 4768 crypto adapter.
The Omnikey Cardman 3821 smart card readers can be carried forward to any TKE 9.2
workstation. Smart cards 45D3398, 74Y0551, and 00JA710 can be used on TKE 9.2.
When performing a MES upgrade from TKE 7.3, TKE 8.0, or TKE 8.1 to a TKE 9.2
installation, the following steps must be completed:
1. Save Upgrade Data on an old TKE to USB memory to save client data.
2. Replace the 4767 crypto adapter with the 4768 crypto adapter.
3. Upgrade the firmware to TKE 9.2.
4. Install the Frame Roll to apply Save Upgrade Data (client data) to the TKE 9.2 system.
5. Run the TKE Workstation Setup wizard.
Note: A workstation that was upgraded to TKE V8.x includes the 4767 cryptographic
adapter that is required to manage Crypto Express5S; however, it cannot be used to
manage the Crypto Express7S or Crypto Express6S.
If your z15 includes Crypto Express7S or Crypto Express6S, you must upgrade to TKE
V9.2, which requires the 4768 cryptographic adapter.
Important: TKE workstations that are at FC 0841 or older do not support the 4767 or 4768
cryptographic adapters.
For more information about TKE hardware support, see Table 6-3. For some functionality,
requirements must be considered; for example, the characterization of a Crypto Express
adapter in EP 11 mode always requires the use of a TKE.
Attention: The TKE is unaware of the CPC type where the host crypto module is installed.
That is, the TKE does not consider whether a Crypto Express is running on z15, 14, or z13,
or z13s system. Therefore, the LIC can support any CPC where the coprocessor is
supported, but the TKE LIC must support the specific crypto module.
Offers UDX - X - -
RSA functions - X X X
For more information about the minimum required and recommended distribution levels, see
the IBM Z website.
For more information about the software support levels for cryptographic functions, see
Chapter 7, “Operating system support” on page 253.
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, Features Max34, Max71, Max108,
Max145, and Max190, which is further identified as IBM z15 Model T01, or z15 T01,
unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, Features Max4, Max13, Max21, Max32,
and Max65, which is further identified as IBM z15 Model T02, or z15 T02, unless
otherwise specified.
In the remainder of this chapter, IBM z15 (z15) refers to both machine types.
Because this information is subject to change, see the following hardware fix categories for
most current information:
IBM.Device.Server.z15-8561.* for z15 T01, and
IBM.Device.Server.z15-8562.* for z15 T02.
Support of z15 functions depends on the operating system, its version, and release.
End of service operating systems: Operating system levels that are no longer in service
are not covered in this publication. These older levels might support some features.
z/OS V2R1c
z/VM V6R4
z/VSE V6.2d,e
z/TPF V1R1
The use of certain features depends on the operating system. In all cases, program
temporary fixes (PTFs) might be required with the operating system level that is indicated.
Check the z/OS fix categories, or the subsets of the 8561DEVICE (z15 T01) and
8652DEVICE (z15 T02) PSP buckets for z/VM and z/VSE. The fix categories and the PSP
buckets are continuously updated, and contain the latest information about maintenance:
Hardware and software buckets contain installation information, hardware and software
service levels, service guidelines, and cross-product dependencies.
For more information about Linux on Z distributions and KVM hypervisor, see the
distributor’s support information.
For more information about supported functions that are based on operating systems, see
7.3, “z15 features and function support overview” on page 257. Tables are built by function
and feature classification to help you determine, by a quick scan, what is supported and the
minimum operating system level that is required.
For more information about supported functions and their minimum required support levels,
see 7.3, “z15 features and function support overview” on page 257.
7.2.2 z/VM
z/VM V6R4, z/VM V7R1, and z/VM V7R2 provide support that enables guests to use the
following features that are supported by z/VM on IBM z15™:
z/Architecture support
New hardware facilities
ESA/390-compatibility mode for guests
Crypto Clear Key ECC operations
RoCE Express2 support
Dynamic I/O support
Provided for managing the configuration of OSA-Express7S and OSA-Express6S OSD
CHPIDs, FICON Express16SA2 and FICON Express16S+ (FC and FCP CHPIDs), and
RoCE Express2 features.
Improved memory management
For more information about supported functions and their minimum required support levels,
see 7.3, “z15 features and function support overview” on page 257.
7.2.3 z/VSE
z15 support is provided by z/VSE V6R2 and later, with the following considerations:
z/VSE runs in z/Architecture mode only.
z/VSE supports 64-bit real and virtual addressing.
For more information about supported functions and their minimum required support levels,
see 7.3, “z15 features and function support overview” on page 257.
7.2.4 z/TPF
z15 support is provided by z/TPF V1R1 with PTFs. For more information about supported
functions and their minimum required support levels, see 7.3, “z15 features and function
support overview” on page 257.
1
Use support for select features by way of PTFs. Toleration support for new hardware might also require PTFs.
2 FICON Express16SA supported on z15 T01 only.
The service levels of SUSE, Red Hat, and Ubuntu releases that are supported at the time of
this writing are listed in Table 7-2.
For more information about supported Linux distributions on IBM Z servers, see the Tested
platforms for Linux page of the IBM IT infrastructure website.
IBM is working with Linux distribution Business Partners to provide further use of selected
z15 functions in future Linux on Z distribution releases.
Note: The tables in this section list but do not explicitly mark all the features that require
fixes that are required by the corresponding operating system for toleration or exploitation.
For more information, see the PSP buckets for their respective devices and subsets:
z15 T01 - PSP bucket 8561DEVICE
z15 T02 - PSP bucket 8562DEVICE
Note: z/OS V2R1 support has ended on as of September 2018. No new function is
provided for exploiting the new HW features (toleration support only). Although extended
(fee-based) support for z/OS 2.1 can be obtained, support for z/OS 2.1 is not covered
extensively in this document.
z15 servers Y Y Y Y Y Y
Maximum processor unit (PUs) per system image 190 190b 190b 80c 80c 64d
Dynamic PU add Y Y Y Y Y Y
Program-directed re-IPL - - - Y Y Y
HiperDispatch Y Y Y Y Y Y
Out-of-order execution Yl Y Y Y Y Y
The supported base CPC functions for z/VSE, z/TPF, and Linux on Z are listed in Table 7-4.
Table 7-4 Supported base CPC functions for z/VSE, z/TPF, and Linux on Z
Functiona z/VSE z/TPF Linux on
V6R2 V1R1 Zb
z15 servers Y Y Y
z15 T01 Maximum processor unit (PUs) per system image 10 86 190c
Dynamic PU add Y N Y
Program-directed re-IPL Y N Y
HiperDispatch N N Y
Transactional Execution N N Y
Out-of-order execution Y Y Y
Note: z/OS V2R1 support has ended on as of September 2018. No new function is
provided for exploiting the new HW features (toleration support only). Although extended
(fee-based) support for z/OS 2.1 can be obtained, support for z/OS 2.1 is not covered
extensively in this document.
Table 7-5 Supported coupling and clustering functions for z/OS and z/VM
Functiona z/OS z/OS z/OS z/VM z/VM z/VM
V2R4 V2R3 V2R2 V7R2 V7R1 V6R4
CF Monopolization Avoidance Yc Yc Yc Yg Yg Yg
In addition to operating system support that is listed in Table 7-5 on page 261, Server Time
Protocol is supported on z/TPF V1R1 and Linux on Z. Also, CFCC Level 22, Level 23, and
Level 24 are supported for z/TPF V1R1.
Storage connectivity
The supported storage connectivity functions for z/OS and z/VM are listed Table 7-6.
Table 7-6 Supported storage connectivity functions for z/OS and z/VM
Functiona z/OS z/OS z/OS z/VM z/VM z/VM
V2R4 V2R3 V2R2 V7R2 V7R1 V6R4
The supported storage connectivity functions for z/VSE, z/TPF, and Linux on Z are listed in
Table 7-7.
Table 7-7 Supported storage connectivity functions for z/VSE, z/TPF, and Linux on Z
Functiona z/VSE z/TPF Linux
V6R2 V1R1 on Zb
zHyperLink Express - - -
Note: z/OS V2R1 is End of Support as of September 2018. No new function is provided for
exploiting HW features and functions introduced with IBM z15 (toleration support only).
Although extended (fee-based) support for z/OS 2.1 can be obtained, support for z/OS 2.1
is not covered extensively in this document.
Table 7-8 Supported network connectivity functions for z/OS and z/VM
Functiona z/OS z/OS z/OS z/VM z/VM z/VM
V2R4 V2R3 V2R2 V7R2 V7R1 V6R4
Hipersockets
HiperSocketse Y Y Y Y Y Y
32 HiperSockets Y Y Y Y Y Y
The supported network connectivity functions for z/VSE, z/TPF, and Linux on Z are listed in
Table 7-9.
Table 7-9 Supported network connectivity functions for z/VSE, z/TPF, and Linux on Z
Functiona z/VSE z/TPF Linux
V6R2 V1R1 on Zb
Hipersockets
HiperSocketsd Y N Y
32 HiperSockets Y N Y
Note: z/OS V2R1 is End of Support as of September 2018. No new function is provided for
exploiting HW features and functions introduced with IBM z15 (toleration support only).
Although extended (fee-based) support for z/OS 2.1 can be obtained, support for z/OS 2.1
is not covered extensively in this document.
Crypto Express7S Y Y Yc Yb Yb Yb
Crypto Express6S Y Y Yc Yb Yb Yb
Crypto Express5S Y Y Y Yb Yb Yb
The z15 supported cryptography functions for z/VSE, z/TPF, and Linux on Z are listed in
Table 7-11.
Table 7-11 Supported cryptography functions for z/VSE, z/TPF, and Linux on Z
Functiona z/VSE z/TPF Linux
V6R2 V1R1 on Zb
Crypto Express7S Y Y Y
Crypto Express6S Y Y Y
Crypto Express5S Y Y Y
Note: z/OS V2R1 is End of Support as of September 2018. No new function is provided for
exploiting HW features and functions introduced with IBM z15 (toleration support only).
Although extended (fee-based) support for z/OS 2.1 can be obtained, support for z/OS 2.1
is not covered extensively in this document.
z/VSE V6.2 and later z/VSE Turbo Dispatcher can use up to 4 CPs, and tolerates up to
10-way LPARs
Linux on Z SUSE Linux Enterprise Server 12 and later: 256 CPs or IFLs.
SUSE Linux Enterprise Server 11: 64 CPs or IFLs.
Red Hat RHEL 7 and later: 256 CPs or IFLs.
Red Hat RHEL 6: 64 CPs or IFLs.
Ubuntu 16.04 LTS and 18.04 LTS: 256 CPs or IFLs.
KVM Hypervisor The KVM hypervisor is offered with the following Linux distributions
-- 256CPs or IFLs--:
SLES 12 SP4 and later.
RHEL 7.5 with kernel-alt package (kernel 4.14).
Ubuntu 16.04 LTS and Ubuntu 18.04 LTS.
The supported operating systems are listed in Table 7-3 and Table 7-4 on page 259.
The supported operating systems are listed in Table 7-3 and Table 7-4 on page 259.
The supported operating systems are listed in Table 7-3 and Table 7-4 on page 259.
The dynamic PU add function enhances this support by allowing you to dynamically define
and change the number and type of reserved PUs in an LPAR profile, which removes any
planning requirements. The new resources are immediately made available to the operating
system and in the case of z/VM, to its guests.
The supported operating systems are listed in Table 7-3 and Table 7-4 on page 259.
z/OS can take advantage of this support and nondisruptively acquire and release memory
from the reserved area. z/VM V6R4 and later can acquire memory nondisruptively and
immediately make it available to guests. z/VM virtualizes this support to its guests, which now
also can increase their memory nondisruptively if supported by the guest operating system.
Currently, releasing memory from z/VM is not supported3. Releasing memory from the z/VM
guest depends on the guest’s operating system support.
Linux on Z also supports acquiring and releasing memory nondisruptively. This feature is
enabled for SUSE Linux Enterprise Server 11 and RHEL 6 and later releases.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
The Capacity Provisioning Manager, which is a feature that is first available with z/OS V1R9,
interfaces with z/OS Workload Manager (WLM) and implements capacity provisioning
policies. Several implementation options are available, from an analysis mode that issues
only guidelines, to an autonomic mode that provides fully automated operations.
3
z/VM Dynamic Memory Downgrade (releasing memory from z/VM LPAR) will be made available in the future with
PTFs for APAR VM66271. For more information, see: http://www.vm.ibm.com/newfunction/#dmd
Program-directed re-IPL
First available on System z9®, program directed re-IPL allows an operating system on a z15
to IPL again without operator intervention. This function is supported for SCSI and IBM
extended count key data (IBM ECKD) devices.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
IOCP
All IBM Z servers require a description of their I/O configuration. This description is stored in
I/O configuration data set (IOCDS) files. The I/O configuration program (IOCP) allows for the
creation of the IOCDS file from a source file that is known as the I/O configuration source
(IOCS).
The IOCS file contains definitions of LPARs and channel subsystems. It also includes detailed
information for each channel and path assignment, control unit, and device in the
configuration.
IOCP required level for z15 servers: The required level of IOCP for the z15 is IOCP 5.5.0
with PTFs. For more information, see the following publications:
IBM Z Stand-Alone Input/Output Configuration Program User's Guide, SB10-7173-01
IBM Z Input/Output Configuration Program User’s Guide for ICP IOCP, SB10-7172-04
Dynamic Partition Manager V4.0: Dynamic Partition Manager V4.0 is available for
managing IBM Z servers that are running Linux. DPM 4.0 is available with HMC Driver
Level 41 (HMC Version 2.15.0). IOCP does not need to configure a server that is running in
DPM mode. For more information, see IBM Dynamic Partition Manager (DPM) Guide,
SB10-7176-02.
HiperDispatch
The HIPERDISPATCH=YES/NO parameter in the IEAOPTxx member of SYS1.PARMLIB and on
the SET OPT=xx command controls whether HiperDispatch is enabled or disabled for a z/OS
image. It can be changed dynamically, without an IPL or any outage.
4 FICON Express16SA and FICON Express16S+ do not allow a mixture of CHPID types on the same card.
The use of SMT on z15 systems requires that HiperDispatch is enabled on the operating
system. For more information, see “Simultaneous multithreading” on page 282.
Additionally, any LPAR that is running with more than 64 logical processors is required to
operate in HiperDispatch Management Mode.
HiperDispatch on z15 systems uses a new chip and CPC drawer configuration to improve the
access cache performance. Beginning with z/OS V2R1, HiperDispatch was changed to use
the PU Chip/cluster/drawer cache structure of z15 servers. The base support is provided by
PTFs that are identified by:
For z15 T01 - IBM.device.server.z15-8561.requiredservice
For z15 T02 - IBM.device.server.z15-8562.requiredservice
PR/SM on z15 servers seeks to assign all memory in one CPC drawer that is striped across
the clusters of that drawer to take advantage of the lower latency memory access in a drawer.
Also, PR/SM tries to consolidate storage onto drawers with the most processor entitlement.
PR/SM on z15 servers seeks to assign all logical processors of a partition to one CPC
drawer, packed into PU chips of that CPC drawer in cooperation with operating system
HiperDispatch optimize shared cache usage.
PR/SM automatically keeps a partition’s memory and logical processors on the same CPC
drawer. This arrangement looks simple for a partition, but it is a complex optimization for
multiple logical partitions because some must be split among processors drawers.
To use HiperDispatch effectively, WLM goal adjustment might be required. Review the WLM
policies and goals and update them as necessary. WLM policies can be changed without
turning off HiperDispatch. A health check is provided to verify whether HiperDispatch is
enabled on a system image.
z/TPF
z/TPF on z15 can use more processors immediately without reactivating the LPAR or IPLing
the z/TPF system.
In installations older than z14, z/TPF workload is evenly distributed across all available
processors, even in low-utilization situations. This configuration causes cache and core
contention with other LPARs. When z/TPF is running in a shared processor configuration, the
achieved MIPS is higher when z/TPF is using a minimum set of processors.
In low-utilization periods, z/TPF now minimizes the processor footprint by compressing TPF
workload onto a minimal set of I-streams (engines), which reduces the effect on other LPARs
and allows the entire CPC to operate more efficiently.
As a consequence, z/OS and z/VM experience less contention from the z/TPF system when
the z/TPF system is operating at periods of low demand.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
zIIP support
zIIPs do not change the model capacity identifier of z15 servers. IBM software product license
charges that are based on the model capacity identifier are not affected by the addition of
zIIPs. On a z15 server, z/OS Version 2 Release 1 is the minimum level for supporting zIIPs.
No changes to applications are required to use zIIPs. They can be used by the following
applications:
Db2 V8 and later for z/OS data serving for applications that use data Distributed Relational
Database Architecture (DRDA) over TCP/IP, such as data serving, data warehousing, and
selected utilities.
z/OS XML services.
z/OS CIM Server.
z/OS Communications Server for network encryption (Internet Protocol Security [IPSec])
and for large messages that are sent by HiperSockets.
IBM GBS Scalable Architecture for Financial Reporting.
IBM z/OS Global Mirror (formerly XRC) and System Data Mover.
IBM z/OS Container Extensions.
IBM OMEGAMON® XE on z/OS, OMEGAMON XE on Db2 Performance Expert, and Db2
Performance Monitor.
Any Java application that uses the current IBM SDK.
WebSphere Application Server V5R1 and later, and products that are based on it, such as
WebSphere Portal, WebSphere Enterprise Service Bus (WebSphere ESB), and
WebSphere Business Integration (WBI) for z/OS.
CICS/TS V2R3 and later.
Db2 UDB for z/OS Version 8 and later.
IMS Version 8 and later.
On z15 servers, the zIIP processor is designed to run in SMT mode, with up to two threads
per processor. This function is designed to help improve throughput for zIIP workloads and
provide appropriate performance measurement, capacity planning, and SMF accounting
data. This support is available for z/OS V2.1 with PTFs and higher.
Use the PROJECTCPU option of the IEAOPTxx parmlib member to help determine whether zIIPs
can be beneficial to the installation. Setting PROJECTCPU=YES directs z/OS to record the
amount of eligible work for zIIPs in SMF record type 72 subtype 3. The field APPL% IIPCP of
the Workload Activity Report listing by WLM service class indicates the percentage of a
processor that is zIIP eligible. Because of the zIIP’s lower price as compared to a CP, even a
utilization as low as 10% can provide cost benefits.
Transactional Execution
The IBM zEnterprise EC12 introduced an architectural feature called Transactional Execution
(TX). This capability is known in academia and industry as hardware transactional memory.
Transactional execution is also implemented on subsequent IBM Z servers.
This feature enables software to indicate to the hardware the beginning and end of a group of
instructions that must be treated in an atomic way. All of their results occur or none occur, in
true transactional style. The execution is optimistic.
The hardware provides a memory area to record the original contents of affected registers
and memory as the instruction’s execution occurs. If the transactional execution group is
canceled or must be rolled back, the hardware transactional memory is used to reset the
values. Software can implement a fallback capability.
This capability increases the software’s efficiency by providing a way to avoid locks (lock
elision). This advantage is of special importance for speculative code generation and highly
parallelized applications.
TX is used by IBM Java virtual machine (JVM) and might be used by other software. The
supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on page 259.
The temporary processing capacity boost is intended to help workloads catch-up after an IPL
(planned or unplanned) to work faster through a backlog after a downtime, which improves
overall system availability by reducing the elapsed time required to recover service.
Automation
The client’s automation product can be used to automate and control the following System
Recovery Boost activities:
To activate and deactivate the eBod temporary capacity record to provide more physical
zIIPs for an IPL or Shutdown Boost.
To dynamically modify LPAR weights, as might be needed to modify the sharing of
physical zIIP capacity during a Boost period.
To drive the invocation of the PROC that indicates the beginning of a shutdown process
(and the start of the shut-down Boost).
To take advantage of new composite HW API reconfiguration actions.
5
System Recovery Boost Upgrade using temporary zIIP capacity is NOT available for z15 T02. System Recovery
boost is available with existing hardware only (subcapacity CP speed boost and zIIP boost with existing zIIPs.
Simultaneous multithreading
SMT is the hardware capability to process up to two simultaneous threads in a single core,
sharing the resources of the superscalar core. This capability improves the system capacity
and efficiency in the usage of the processor, which increases the overall throughput of the
system.
The z15 can run up two threads simultaneously in the same processor, which dynamically
shares resources of the core, such as cache, translation lookaside buffer (TLB), and
execution resources. It provides better utilization of the cores and more processing capacity.
Note: For zIIPs and IFLs, SMT must be enabled on z/OS, z/VM, or Linux on Z instances.
An operating system with SMT support can be configured to dispatch work to a thread on a
zIIP (for eligible workloads in z/OS) or an IFL (for z/VM) core in single-thread or SMT
mode.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
An operating system that uses SMT controls each core and is responsible for maximizing
their throughput and meeting workload goals with the smallest number of cores. In z/OS,
consider HiperDispatch cache optimization when you must choose the two threads to be
dispatched in the same processor.
HiperDispatch attempts to dispatch guest virtual CPUs on the same logical processor on
which they ran. PR/SM attempts to dispatch a vertical low logical processor in the same
physical processor. If that process is not possible, it attempts to dispatch it in the same node,
or then the same CPC drawer where it was dispatched before to maximize cache reuse.
From the point of view of an application, SMT is transparent and no changes are required in
the application for it to run in an SMT environment, as shown in Figure 7-1.
MT Ignorant
z/OS z/VM
PR/SM Hypervisor MT Aware
z/OS
The following APARs must be applied to z/OS V2R1 to use SMT:
6 SMT is also enabled (not user configurable) by default for SAPs.
The use of SMT on z/OS V2R1 requires enabling HiperDispatch, and defining the processor
view (PROCVIEW) control statement in the LOADxx parmlib member and the MT_ZIIP_MODE
parameter in the IEAOPTxx parmlib member.
The PROCVIEW statement is defined for the life of IPL, and can have the following values:
CORE: This value specifies that z/OS should configure a processor view of core, in which a
core can include one or more threads. The number of threads is limited by z15 to two
threads. If the underlying hardware does not support SMT, a core is limited to one thread.
CPU: This value is the default. It specifies that z/OS should configure a traditional processor
view of CPU and not use SMT.
CORE,CPU_OK: This value specifies that z/OS should configure a processor view of core (as
with the CORE value) but the CPU parameter is accepted as an alias for applicable
commands.
When PROCVIEW CORE or CORE,CPU_OK are specified in z/OS that is running in z15,
HiperDispatch is forced to run as enabled, and you cannot disable HiperDispatch. The
PROCVIEW statement cannot be changed dynamically; therefore, you must run an IPL after
changing it to make the new setting effective.
The MT_ZIIP_MODE parameter in the IEAOPTxx controls zIIP SMT mode. It can be 1 (the
default), where only one thread can be running in a core, or 2, where up two threads can be
running in a core. If PROCVIEW CPU is specified, the MT_ZIIP_MODE is always 1. Otherwise, the
use of SMT to dispatch two threads in a single zIIP logical processor (MT_ZIIP_MODE=2) can be
changed dynamically by using the SET OPT=xx setting in the IEAOPTxx parmlib. Changing the
MT mode for all cores can take some time to complete.
PROCVIEW CORE requires DISPLAY M=CORE and CONFIG CORE to display the core states and
configure an entire core.
With the introduction of Multi-Threading support for SAPs, a maximum of 88 logical SAPs can
be used. RMF is updated to support this change by implementing page break support in the
I/O Queuing Activity report that is generated by the RMF Post processor.
The default in z/VM is multithreading disabled. With the addition of dynamic SMT capability to
z/VM V6R4 through an SPE, the number of active threads per core can be changed without a
system outage and potential capacity gains going from SMT-1 to SMT-2 (one to two threads
per core) can now be achieved dynamically. Dynamic SMT requires applying PTFs that are
running in SMT enabled mode and enables dynamically varying the active threads per core.
z/VM V7R2 and V7R2 support up to 40 multithreads cores (80 threads), while V6R4 supports
up to 32 multithreaded cores (64 threads) for IFLs, and each thread is treated as an
independent processor. z/VM dispatches virtual IFLs on the IFL logical processor so that the
same or different guests can share a core. Each core has a single dispatch vector, and z/VM
attempts to place virtual sibling IFLs on the same dispatch vector to maximize cache reuses.
The following minimum releases of Linux on Z distributions are supported on z15 (native SMT
support):
SUSE:
– SLES 15 SP1 with service
– SUSE SLES 12 SP4 with service
– SUSE SLES 11 SP4 with service
Red Hat:
– Red Hat RHEL 8.0 with service
– Red Hat RHEL 7.7 with service
– Red Hat RHEL 6.10 with service
Ubuntu:
– Ubuntu 18.04.1 LTS with service
– Ubuntu 16.04.5 LTS with service
The KVM hypervisor is supported on the same Linux on Z distributions in this list.
For most current support, see the Linux on IBM Z Tested platforms website.
Single-instruction multiple-data
The SIMD feature introduces a new set of instructions to enable parallel computing that can
accelerate code with string, character, integer, and floating point data types. The SIMD
instructions allow a larger number of operands to be processed with a single complex
instruction.
z15 is equipped with new set of instructions to improve the performance of complex
mathematical models and analytic workloads through vector processing and new complex
instructions, which can process numerous data with a single instruction. This new set of
instructions, which is known as SIMD, enables more consolidation of analytic workloads and
business transactions on Z servers.
SIMD on z15 has support for enhanced math libraries that provide performance
improvements for analytical workloads by processing more information with a single CPU
instruction.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259. Operating System support includes the following features7:
Enablement of vector registers.
Use of vector registers that use XL C/C++ ARCH(11) and TUNE(11).
7 The features that are listed here might not be available on all operating systems that are listed in the tables.
MASS and ATLAS can reduce the time and effort for middleware and application developers.
IBM provides compiler built-in functions for SMID that software applications can use as
needed, such as for using string instructions.
The use of new hardware instructions require the z/OS V2R4 XL C/C++ compiler with
ARCH(13) and TUNE(13) options for targeting z15 instructions. The ARCH(13) compiler
option allows the compiler to use any new z15 instructions where appropriate. The TUNE(13)
compiler option allows the compiler to tune for any z15 micro-architecture.
Vector programming support is extended for z15 to provide access to the new instructions
that were introduced by the VEF 28 specification.
Older levels of z/OS XL C/C++ compilers do not provide z15 exploitation; however, the z/OS
V2R4 XL C/C++ compiler can be used to generate code for the older levels of z/OS running
on z15.
Code must be developed to take advantage of the SIMD functions. Applications with SIMD
instructions abend if they run on a lower hardware level system. Some mathematical function
replacement can be done without code changes by including the scalar MASS library before
the standard math library.
The MASS and standard math library include different accuracies, so assess the accuracy of
the functions in the context of the user application before deciding whether to use the MASS
and ATLAS libraries.
The SIMD functions can be disabled in z/OS partitions at IPL time by using the MACHMIG
parameter in the LOADxx member. To disable SIMD code, use the MACHMIG VEF
hardware-based vector facility. If you do not specify a MACHMIG statement, which is the default,
the system is unlimited in its use of the Vector Facility for z/Architecture (SIMD).
Decimal floating point support was introduced with z9 EC. z15 servers inherited the decimal
floating point accelerator feature that was introduced with z10 EC.
8 Hardware-based Vector-extension facility 2.
Out-of-order execution
Out-of-order (OOO) execution yields significant performance benefits for compute-intensive
applications by reordering instruction execution, which allows later (newer) instructions to be
run ahead of a stalled instruction, and reordering storage accesses and parallel storage
accesses. OOO maintains good performance growth for traditional applications.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259. For more information, see “3.4.3, “Out-of-Order execution” on page 102.
For more information about this function, see The Load-Program-Parameter and the
CPU-Measurement Facilities.
For more information about the CPU Measurement Facility, see the CPU MF - Update and
WSC Experiences page of the IBM Techdocs Library website.
For more information, see “12.2, “z15 Large System Performance Reference ratio” on
page 478.
IBM Virtual Flash Memory (FC 0643) offers up to 6.0 TB of memory for z15 T01, and up to 2.0
TB for z15 T02 in 0.5 TB Increments. VFM is provided for improved application availability and
to handle paging workload spikes.
IBM Virtual Flash Memory is designed to help improve availability and handling of paging
workload spikes when running z/OS V2.1, V2.2, V2.3, or V2.4. With this support, z/OS is
designed to help improve system availability and responsiveness by using VFM across
transitional workload events, such as market openings, and diagnostic data collection. z/OS is
also designed to help improve processor performance by supporting middleware exploitation
of pageable large (1 MB) pages.
Therefore, VFM can help organizations meet their most demanding service level agreements
and compete more effectively. VFM is easily configurable, and to provide rapid time to value.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
z/OS
GSF support allows an area of storage to be identified such that an Exit routine assumes
control if a reference is made to that storage. GSF is managed by new instructions that define
Guarded Storage Controls and system code to maintain that control information across
undispatch and redispatch.
z/VM
With the PTF for APAR VM65987, z/VM V6.4 provides support for guest exploitation of the
guarded storage facility. This facility is designed to improve the performance of
garbage-collection processing by various languages, in particular Java.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
Through enhanced hardware features (based on DAT table entry bit) and explicit software
requests to obtain memory areas as non-executable, areas of memory can be protected from
unauthorized execution. A Protection Exception occurs if an attempt is made to fetch an
instruction from an address in such an element or if an address in such an element is the
target of an execute-type instruction.
z/OS
To use IEP, Real Storage Manager (RSM) is enhanced to request non-executable memory
allocation. Use new keyword EXECUTABLE=YES|NO on STORAGE OBTAIN or IARV64 to indicate
whether memory to be used contains executable code. Recovery Termination Manager
(RTM) writes LOGREC record of any program-check that results from IEP.
IEP support is for z/OS 2.2 and later running on z15 with APARs OA51030 and OA51643
installed.
z/VM
Guest exploitation support for the Instruction Execution Protection Facility is provided with
APAR VM65986.
The supported operating systems are listed in Table 7-3 on page 258 and Table 7-4 on
page 259.
Each PU chip has one on-chip compression unit, which is designed to replace the
zEnterprise Data Compression (zEDC) Express PCIe feature.
The zEDC Express feature available on older systems is NOT carried forward to z15.
The IBM Integrated Accelerator for zEDC maintains software compatibility with existing zEDC
Express use cases. For more information, see Integrated Acceleratorfor zEnterprise Data
Compression.
All data interchange with existing (zEDC) compressed data remains compatible as z15 and
zEDC capable machines coexist (accessing same data). Data that is compressed and written
with zEDC will be read and decompressed by z15 well into the future.
Functionality support for the IBM Integrated Accelerator for zEDC is listed in Table 7-3 on
page 258 and Table 7-4 on page 259.
For more information, see Appendix C, “IBM Integrated Accelerator for zEnterprise Data
Compression” on page 509.
Consideration: Because coupling link connectivity with z14 does not support InfiniBand,
introducing z15 into an installation requires extra planning. Consider the level of CFCC. For
more information, see “Migration considerations” on page 199.
CFCC Level 24
CFCC Level 24 is delivered on z15 servers with driver level 41. CFCC Level 24 introduces the
following enhancements:
CFCC Fair Latch Manager
This enhancement to the internals of the Coupling Facility (CFCC) dispatcher provides CF
work management efficiency and processor scalability improvements, and improve the
“fairness” of arbitration for internal CF resource latches across tasks
CFCC Message Path Resiliency enhancement
CF Message Paths use a z/OS-provided system identifier (SYID) to uniquely identify
which z/OS system image, and instance of that system image, is sending requests over a
message path to the CF. With z15, we are providing a new resiliency mechanism that
transparently recovers for this “missing” message path deactivate (if and when that
deactivation ever occurs).
During path initialization, the CF provides more information to z/OS about every message
path that appears active, including the SYID for the path. Whenever z/OS interrogates the
state of the message paths to the CF, z/OS checks this SYID information for currency and
correctness, and if incorrect, gather diagnostic information and reactivates the path to
correct the problem.
CF monopolization avoidance
z/OS will take advantage of current CF support in CFLEVEL 24 (z15 T01/T02) to deliver
improved z/OS support for handling CF monopolization.
With z15 T01/T02, the CF dispatcher will monitor in real-time the number of CF tasks that
have a command assigned to them for a given structure, on a structure-by-structure basis.
When the number of CF tasks being used by any given structure exceeds a
model-dependent CF threshold, and a global threshold on the number of active tasks is
also exceeded, the structure will be considered to be “monopolizing” the CF, and z/OS will
be informed of this monopolization.
New support in z/OS will observe the monopolization state for a structure, and start to
selectively queue and throttle incoming requests to the CF, on a structure-specific basis –
while other requests, for other “non-monopolizing” structures and workloads, are
completely unaffected.
z/OS will dynamically manage the queue of requests for the “monopolizing” structures to
limit the number of active CF requests (parallelism) to them, and will monitor the CF’s
monopolization state information so as to observe the structure becoming
CFCC Level 23
CFCC Level 23 is delivered on z14 servers with driver level 36. In addition to CFCC Level 22
enhancements, it introduces the following enhancements:
Asynchronous cross invalidation (XI) for CF cache structures
This enhancement requires z/OS fixes for APARs OA54688 (exploitation) and OA54985
(toleration). It also requires explicit data manager support (Db2 V12 with PTFs).
Coupling Facility hang detection
These enhancements provide a significant reduction in failure scope and client disruption
(CF-level to structure-level), with no loss of FFDC collection capability. With this support,
the CFCC dispatcher significantly reduces the CF hang detection interval to only
2 seconds, which allows more timely detection and recovery from such events.
When a hang is detected, in most cases the CF confines the scope of the failure to
“structure damage” for the single CF structure the hung command was processing
against, capture diagnostics with a nondisruptive CF dump, and continue operating
without stopping or rebooting the CF image.
Coupling Facility granular latching
This enhancement eliminates the performance degradation that is caused by
structure-wide latching. With this support, most CF list and lock structure ECR processing
no longer uses structure-wide latching. It serializes its execution by using the normal
structure object latches that all mainline commands use. However, a few “edge conditions”
in ECR processing still require structure-wide latching.
Before you begin the migration process, install the compatibility and coexistence PTFs. A
planned outage is required when you upgrade the CF or CF LPAR to CFCC Level 23.
The supported operating systems are listed in Table 7-5 on page 261.
For more information about the latest CFCC code levels, see the current exception letter that
is published on Resource Link website (login is required).
CF structure sizing changes are expected when upgrading from a previous CFCC Level to
CFCC Level 21. Review the CF LPAR size by using the available CFSizer tool, which is
available for download at the IBM Systems support website.
The Sizer Utility, which is an authorized z/OS program download, is useful when you are
upgrading a CF. The tool is available for download at the IBM Systems support website.
Before you begin the migration process, install the compatibility and coexistence PTFs. A
planned outage is required when you upgrade the CF or CF LPAR to CFCC Level 22.
For more information, see “Coupling Thin Interrupts (default CF LPAR setting with z15)”
on page 116. The supported operating systems are listed in Table 7-5 on page 261.
The supported operating systems are listed in Table 7-5 on page 261.
Instead of performing XI signals synchronously on every cache update request that causes
them, data managers can “opt in” for the CF to perform these XIs asynchronously (and then
sync them up with the CF at or before transaction completion). Data integrity is maintained if
all XI signals complete by the time transaction locks are released.
The feature enables faster completion of cache update CF requests, especially with cross-site
distance involved and provides improved cache structure service times and coupling
efficiency. It requires explicit data manager exploitation/participation, which is not transparent
to the data manager. No SMF data changes were made for CF monitoring and reporting.
This function refers exclusively to the z/VM dynamic I/O support of InfiniBand and ICA
coupling links. Support is available for the CIB and CS5 CHPID type in the z/VM dynamic
commands, including the change channel path dynamic I/O command.
Specifying and changing the system name when entering and leaving configuration mode are
also supported. z/VM does not use InfiniBand or ICA, and does not support the use of
InfiniBand or ICA coupling links by guests. The supported operating systems are listed in
Table 7-5 on page 261.
zHyperLink Express is designed for up to 5x lower latency than High-Performance FICON for
Z (zHPF) by directly connecting the Z central processor complex (CPC) to the I/O Bay of the
DS8000 (DS8880 or later). This short distance (up to 150 m [492.1 feet]), direct connection is
intended to speed Db2 for z/OS transaction processing and improve active log throughput.
The improved performance of zHyperLink Express allows the Processing Unit (PU) to make a
synchronous request for the data that is in the DS8000 cache. This feature eliminates the
undispatch of the running request, the queuing delays to resume the request, and the PU
cache disruption.
Support for zHyperLink Writes can accelerate Db2 log writes to help deliver superior service
levels by processing high-volume Db2 transactions at speed. IBM zHyperLink Express
requires compatible levels of DS8000/F hardware, firmware R8.5.1 or later, and Db2 12 with
PTFs.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
FICON Express16SA
Important: FICON Express16SA is only available on z15 T01 (new build system). z15 T02
supports FICON Express16S+ for new build and as carry forward.
FICON Express16SA supports a link data rate of 16 gigabits per second (Gbps) and
autonegotiation to 8 Gbps for synergy with switches, directors, and storage devices. With
support for native FICON, High-Performance FICON for Z (zHPF), and Fibre Channel
Protocol (FCP), the IBM z15™ server enables you to position your SAN for even higher
performance, which helps you to prepare for an end-to-end 16 Gbps infrastructure to meet
the lower latency and increased bandwidth demands of your applications.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
https://www.ibm.com/downloads/cas/US-ENUS120-013-CA/name/US-ENUS120-013-CA.PDF
The new FICON Express16S+ channel works with your fiber optic cabling environment
(single mode and multimode optical cables). The FICON Express16S+ feature running at
end-to-end 16 Gbps link speeds provides reduced latency for large read/write operations and
increased bandwidth compared to the FICON Express8S feature.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
FICON Express16S
FICON Express16S supports a link data rate of 16 Gbps and autonegotiation to 4 or 8 Gbps
for synergy with existing switches, directors, and storage devices. With support for native
FICON, zHPF, and FCP, the z14 server enables SAN for even higher performance, which
helps to prepare for an end-to-end 16 Gbps infrastructure to meet the increased bandwidth
demands of your applications.
The new features for the multimode and single mode fiber optic cabling environments reduce
latency for large read/write operations and increase bandwidth compared to the FICON
Express8S features.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
FICON Express8S
The FICON Express8S provides a link rate of 8 Gbps, with auto negotiation to 4 or 2 Gbps for
compatibility with previous devices and investment protection. Both 10 km (6.2 miles) LX and
SX connections are offered (in a feature, all connections must include the same type).
FICON Express8S introduced a hardware data router for more efficient zHPF data transfers.
It is the first channel with hardware that is designed to support zHPF, as compared to FICON
Express8, FICON Express4, and FICON Express2, which include a firmware-only zHPF
implementation.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
To use this enhancement, the control unit must support the new IU pacing protocol. IBM
System Storage™ DS8000 series supports extended distance FICON for IBM Z
environments. The channel defaults to current pacing values when it operates with control
units that cannot use extended distance FICON.
High-performance FICON
High-performance FICON (zHPF) was first provided on System z10®, and is a FICON
architecture for protocol simplification and efficiency. It reduces the number of information
units (IUs) that are processed. Enhancements were made to the z/Architecture and the
FICON interface architecture to provide optimizations for online transaction processing
(OLTP) workloads.
Important: FICON Express16SA is only available on z15 T01 (new build system). z15 T02
supports FICON Express16S+ for new build and as carry forward.
zHPF is available on z15, z14, z13, z13s, zEC12, and zBC12 servers. The FICON
Express16SA, FICON Express16S+ FICON Express16S, and FICON Express8S (CHPID
type FC) concurrently support the existing FICON protocol and the zHPF protocol in the
server LIC.
When used by the FICON channel, the z/OS operating system, and the DS8000 control unit
or other subsystems, the FICON channel processor usage can be reduced and performance
improved. Appropriate levels of Licensed Internal Code (LIC) are required.
Also, the changes to the architectures provide end-to-end system enhancements to improve
reliability, availability, and serviceability (RAS).
For example, the zHPF channel programs can be used by the z/OS OLTP I/O workloads,
Db2, VSAM, the partitioned data set extended (PDSE), and the z/OS file system (zFS).
At the zHPF announcement, zHPF supported the transfer of small blocks of fixed size data
(4 K) from a single track. This capability was extended, first to 64 KB, and then to multitrack
operations. The 64 KB data transfer limit on multitrack operations was removed by z196. This
improvement allows the channel to fully use the bandwidth of FICON channels, which results
in higher throughputs and lower response times.
zHPF is enhanced to allow all large write operations (greater than 64 KB) at distances up to
100 km (62.13 miles) to be run in a single round trip to the control unit. This process does not
elongate the I/O service time for these write operations at extended distances. This
enhancement to zHPF removes a key inhibitor for clients adopting zHPF over extended
distances, especially when the IBM HyperSwap capability of z/OS is used.
From the z/OS perspective, the FICON architecture is called command mode and the zHPF
architecture is called transport mode. During link initialization, the channel node and the
control unit node indicate whether they support zHPF.
The mode that is used for an I/O operation depends on the control unit that supports zHPF
and its settings in the z/OS operating system. For z/OS use, a parameter is available in the
IECIOSxx member of SYS1.PARMLIB (ZHPF=YES or NO) and in the SETIOS system command
to control whether zHPF is enabled or disabled. The default is ZHPF=NO.
Support is also added for the D IOS,ZHPF system command to indicate whether zHPF is
enabled, disabled, or not supported on the server.
Similar to the existing FICON channel architecture, the application or access method provides
the channel program (CCWs). How zHPF (transport mode) manages channel program
operations is different from the CCW operation for the existing FICON architecture (command
mode). While in command mode, each CCW is sent to the control unit for execution. In
transport mode, multiple channel commands are packaged together and sent over the link to
the control unit in a single control block. Fewer processors are used compared to the existing
FICON architecture. Certain complex CCW chains are not supported by zHPF.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
For more information about FICON channel performance, see the performance technical
papers that are available at the IBM Z I/O connectivity page of the IBM IT infrastructure
website.
The MIDAW facility is a system architecture and software feature that is designed to improve
FICON performance. This facility was first made available on System z9 servers, and is used
by the Media Manager in z/OS.
The MIDAW facility provides a more efficient CCW/IDAW structure for certain categories of
data-chaining I/O operations.
MIDAW can improve FICON performance for extended format data sets. Non-extended data
sets can also benefit from MIDAW.
MIDAW can improve channel utilization and I/O response time. It also reduces FICON
channel connect time, director ports, and control unit processor usage.
IBM laboratory tests indicate that applications that use EF data sets, such as Db2, or long
chains of small blocks can gain significant performance benefits by using the MIDAW facility.
MIDAW is supported on FICON channels that are configured as CHPID type FC. The
supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on page 264.
Figure 7-2 shows a single CCW that controls the transfer of data that spans non-contiguous
4 K frames in main storage. When the IDAW flag is set, the data address in the CCW points to
a list of words (IDAWs). Each IDAW contains an address that designates a data area within
real storage.
The number of required IDAWs for a CCW is determined by the following factors:
IDAW format as specified in the operation request block (ORB)
Count field of the CCW
Data address in the initial IDAW
For example, three IDAWS are required when the following events occur:
The ORB specifies format-2 IDAWs with 4 KB blocks.
The CCW count field specifies 8 KB.
The first IDAW designates a location in the middle of a 4 KB block.
CCWs with data chaining can be used to process I/O data blocks that have a more complex
internal structure, in which portions of the data block are directed into separate buffer areas.
This process is sometimes known as scatter-read or scatter-write. However, as technology
evolves and link speed increases, data chaining techniques become less efficient because of
switch fabrics, control unit processing and exchanges, and other issues.
10
Exceptions are made to this statement, and many details are omitted in this description. In this section, we
assume that you can merge this brief description with an understanding of I/O operations in a virtual memory
environment.
The use of MIDAWs is indicated by the MIDAW bit in the CCW. If this bit is set, the skip flag
cannot be set in the CCW. The skip flag in the MIDAW can be used instead. The data count in
the CCW must equal the sum of the data counts in the MIDAWs. The CCW operation ends
when the CCW count goes to zero or the last MIDAW (with the last flag) ends.
The combination of the address and count in a MIDAW cannot cross a page boundary.
Therefore, the largest possible count is 4 K. The maximum data count of all the MIDAWs in a
list cannot exceed 64 K, which is the maximum count of the associated CCW.
The scatter-read or scatter-write effect of the MIDAWs makes it possible to efficiently send
small control blocks that are embedded in a disk record to separate buffers from those that
are used for larger data areas within the record. MIDAW operations are on a single I/O block,
in the manner of data chaining. Do not confuse this operation with CCW command chaining.
VSAM and non-VSAM (DSORG=PS) sets can be defined as EF data sets. For non-VSAM
data sets, a 32-byte suffix is appended to the end of every physical record (that is, block) on
disk. VSAM appends the suffix to the end of every control interval (CI), which normally
corresponds to a physical record.
A 32 K CI is split into two records to span tracks. This suffix is used to improve data reliability,
and facilitates other functions that are described next. Therefore, for example, if the DCB
BLKSIZE or VSAM CI size is equal to 8192, the actual block on storage consists of 8224
bytes. The control unit does not distinguish between suffixes and user data. The suffix is
transparent to the access method and database.
EA is useful for creating large Db2 partitions (larger than 4 GB). Striping can be used to
increase sequential throughput, or to spread random I/Os across multiple logical volumes.
DFSMS striping is useful for using multiple channels in parallel for one data set. The Db2 logs
are often striped to optimize the performance of Db2 sequential inserts.
Processing an I/O operation to an EF data set normally requires at least two CCWs with data
chaining. One CCW is used for the 32-byte suffix of the EF data set. With MIDAW, the
additional CCW for the EF data set suffix is eliminated.
MIDAWs benefit EF and non-EF data sets. For example, to read 12 4 K records from a
non-EF data set on a 3390 track, Media Manager chains together 12 CCWs by using data
chaining. To read 12 4 K records from an EF data set, 24 CCWs are chained (two CCWs per
4 K record). By using Media Manager track-level command operations and MIDAWs, an
entire track can be transferred by using a single CCW.
Performance benefits
z/OS Media Manager includes I/O channel program support for implementing EF data sets,
and automatically uses MIDAWs when appropriate. Most disk I/Os in the system are
generated by using Media Manager.
Users of the Executing Fixed Channel Programs in Real Storage (EXCPVR) instruction can
construct channel programs that contain MIDAWs. However, doing so requires that they
construct an IOBE with the IOBEMIDA bit set. Users of the EXCP instruction cannot construct
channel programs that contain MIDAWs.
The MIDAW facility removes the 4 K boundary restrictions of IDAWs and for EF data sets,
reduces the number of CCWs. Decreasing the number of CCWs helps to reduce the FICON
channel processor utilization. Media Manager and MIDAWs do not cause the bits to move any
faster across the FICON link. However, they reduce the number of frames and sequences that
flow across the link, and therefore use the channel resources more efficiently.
Media Manager with the MIDAW facility can provide significant performance benefits when
used in combination applications that use EF data sets (such as Db2) or long chains of small
blocks.
For more information about FICON and MIDAW, see the following resources:
The I/O Connectivity page of the IBM IT infrastructure website includes information about
FICON channel performance
DS8000 Performance Monitoring and Tuning, SG24-7146
ICKDSF
Device Support Facilities, ICKDSF, Release 17 is required on all systems that share disk
subsystems with a z15 processor.
ICKDSF supports a modified format of the CPU information field that contains a two-digit
LPAR identifier. ICKDSF uses the CPU information field instead of CCW reserve/release for
concurrent media maintenance. It prevents multiple systems from running ICKDSF on the
same volume, and at the same time allows user applications to run while ICKDSF is
processing. To prevent data corruption, ICKDSF must determine all sharing systems that
might run ICKDSF. Therefore, this support is required for z15.
Remember: The need for ICKDSF Release 17 also applies to systems that are not part of
the same sysplex, or are running an operating system other than z/OS, such as z/VM.
The zDAC function is integrated into the hardware configuration definition (HCD). Clients can
define a policy that can include preferences for availability and bandwidth that include parallel
access volume (PAV) definitions, control unit numbers, and device number ranges. When new
controllers are added to an I/O configuration or changes are made to existing controllers, the
system discovers them and proposes configuration changes that are based on that policy.
zDAC provides real-time discovery for the FICON fabric, subsystem, and I/O device resource
changes from z/OS. By exploring the discovered control units for defined logical control units
(LCUs) and devices, zDAC compares the discovered controller information with the current
system configuration. It then determines delta changes to the configuration for a proposed
configuration.
All added or changed logical control units and devices are added into the proposed
configuration. They are assigned proposed control unit and device numbers, and channel
paths that are based on the defined policy. zDAC uses channel path chosen algorithms to
minimize single points of failure. The zDAC proposed configurations are created as work I/O
definition files (IODFs) that can be converted to production IODFs and activated.
zDAC applies to all FICON features that are supported on z15 when configured as CHPID
type FC. The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7
on page 264.
Important: FICON Express16SA is only available on z15 T01 (new build system). z15 T02
supports FICON Express16S+ for new build and as carry forward.
Information about the channels that are connected to a fabric (if registered) allows other
nodes or storage area network (SAN) managers to query the name server to determine what
is connected to the fabric.
The platform and name server registration service are defined in the Fibre Channel Generic
Services 4 (FC-GS-4) standard.
The informal name, 63.75-K subchannels, represents 65280 subchannels, as shown in the
following equation:
63 x 1024 + 0.75 x 1024 = 65280
This equation is applicable for subchannel set 0. For subchannel sets 1, 2 and 3, the available
subchannels are derived by using the following equation:
(64 X 1024) -1=65535
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
z/VM V6R3 MSS support for mirrored direct access storage device (DASD) provides a subset
of host support for the MSS facility to allow the use of an alternative subchannel set for
Peer-to-Peer Remote Copy (PPRC) secondary volumes.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264. For more information about channel subsystem, see Chapter 5, “Central processor
complex channel subsystem” on page 203.
Subchannel sets
z15 T01 supports four subchannel sets (SS0, SS1, SS2, SS3), while z15 T02 support three
subchannel sets (SS0, SS1, SS2).
Subchannel sets SS1, SS2, and SS3 (SS3 for z15 T01 only) can be used for disk alias
devices of primary and secondary devices, and as Metro Mirror secondary devices. This set
helps facilitate storage growth and complements other functions, such as extended address
volume (EAV) and Hyper Parallel Access Volumes (HyperPAV).
See Table 7-6 on page 262 and Table 7-7 on page 264 for list of supported operating
systems.
See Table 7-6 on page 262 and Table 7-7 on page 264 for list of supported operating
systems. For more information, refer to “IPL from an alternative subchannel set” on page 303.
32 K subchannels
To help facilitate growth and continue to enable server consolidation, the z15 supports up to
32 K subchannels per FICON ExpressSE, FICON Express16S+ and FICON Express16S
channels (CHPID). More devices can be defined per FICON channel, which includes primary,
secondary, and alias devices. The maximum number of subchannels across all device types
that are addressable within an LPAR remains at 63.75 K for subchannel set 0 and 64 K (64 X
1024)-1 for subchannel sets 1, 2, and 3.
Important: FICON Express16SA is only available on z15 T01 (new build system). z15 T02
supports FICON Express16S+ for new build and as carry forward.
This support is available to the z15, z14, z13, and z13s servers and applies to the FICON
ExpressSE, FICON Express16S+, and FICON Express16S features (defined as CHPID type
FC). FICON Express8S remains at 24 subchannel support when defined as CHPID type FC.
The supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on
page 264.
No action is required on z/OS to enable the health check; it is automatically enabled at IPL
and reacts to changes that might cause problems. The health check can be disabled by using
the PARMLIB or SDSF modify commands.
The supported operating systems are listed in Table 7-6 on page 262. For more information
about FICON Dynamic Routing (FIDR), see Chapter 4, “Central processor complex I/O
structure” on page 153.
The supported operating systems are listed in Table 7-6 on page 262.
For more information about FCP channel performance, see the performance technical papers
that are available at the IBM Z I/O connectivity page of the IBM IT infrastructure website.
The FCP protocol is supported by z/VM, z/VSE, and Linux on Z. The supported operating
systems are listed in Table 7-6 on page 262 and Table 7-7 on page 264.
T10-DIF support
American National Standards Institute (ANSI) T10 Data Integrity Field (DIF) standard is
supported on IBM Z for SCSI end-to-end data protection on fixed block (FB) LUN volumes.
IBM Z provides added end-to-end data protection between the operating system and the
DS8870 unit. This support adds protection information that consists of Cyclic Redundancy
Checking (CRC), Logical Block Address (LBA), and host application tags to each sector of FB
data on a logical volume.
IBM Z support applies to FCP channels only. The supported operating systems are listed in
Table 7-6 on page 262 and Table 7-7 on page 264.
N_Port ID Virtualization
N_Port ID Virtualization (NPIV) allows multiple system images (in LPARs or z/VM guests) to
use a single FCP channel as though each were the sole user of the channel. First introduced
with z9 EC, this feature can be used with supported FICON features on z14 servers. The
supported operating systems are listed in Table 7-6 on page 262 and Table 7-7 on page 264.
The capabilities of the WWPN are extended to calculate and show WWPNs for virtual and
physical ports ahead of system installation.
The tool assigns WWPNs to each virtual FCP channel or port by using the same WWPN
assignment algorithms that a system uses when assigning WWPNs for channels that use
NPIV. Therefore, the SAN can be set up in advance, which allows operations to proceed
much faster after the server is installed. In addition, the SAN configuration can be retained
instead of altered by assigning the WWPN to physical FCP ports when a FICON feature is
replaced.
The WWPN tool takes a .csv file that contains the FCP-specific I/O device definitions and
creates the WWPN assignments that are required to set up the SAN. A binary configuration
file that can be imported later by the system is also created. The .csv file can be created
manually or exported from the HCD/HCM. The supported operating systems are listed in
Table 7-6 on page 262 and Table 7-7 on page 264.
The WWPN tool is applicable to all FICON channels that are defined as CHPID type FCP (for
communication with SCSI devices) on z15. It is available for download at the Resource Link at
the following website (log in is required).
Note: An optional feature can be ordered for WWPN persistency before shipment to keep
the same I/O serial number on the new CPC. Current information must be provided during
the ordering process.
The 25GbE RoCE Express2 has one PCHID and the same virtualization characteristics and
the 10GbE RoCE Express2 (FC 0412 and FC 0432); that is, 126 Virtual Functions per
PCHID.
z/OS requires fixes for APAR OA55686. RMF 2.2 and later is also enhanced to recognize the
CX4 card type and properly display CX4 cards in the PCIe Activity reports.
25GbE RoCE Express2 feature also are used by Linux on Z for applications that are coded to
the native RoCE verb interface or use Ethernet (such as TCP/IP). This native exploitation
does not require a peer OSA.
Support for select Linux on Z distributions is now provided for Shared Memory
Communications over Remote Direct Memory Access (SMC-R) by using RoCE Express
features. For more information, see this Linux on Z Blogspot web page.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
z/OS Communications Server (CS) provides a new software device driver ConnectX4 (CX4)
for RoCE Express2. The device driver is not apparent to both upper layers of the CS (the
SMC-R and TCP/IP stack) and application software (using TCP sockets). RoCE Express2
introduces a minor change in how the physical port is configured.
RMF 2.2 and later is also enhanced to recognize the new CX4 card type and properly display
CX4 cards in the PCIE Activity reports.
The 10GbE RoCE Express2 feature also is used by Linux on Z for applications that are coded
to the native RoCE verb interface or use Ethernet (such as TCP/IP). This native use does not
require a peer OSA.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The 10GbE RoCE Express feature reduces consumption of CPU resources for applications
that use the TCP/IP stack (such as WebSphere accessing a Db2 database). Use of the
10GbE RoCE Express feature also can help reduce network latency with memory-to-memory
transfers by using Shared Memory Communications over Remote Direct Memory Access
(SMC-R) in z/OS V2R1 or later.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Support for this function is required by the sending operating system. For more information,
see “HiperSockets” on page 193. The supported operating systems are listed in Table 7-8 on
page 266.
Layer 2 support can help facilitate server consolidation. Complexity can be reduced, network
configuration is simplified and intuitive, and LAN administrators can configure and maintain
the mainframe environment the same way as they do a non-mainframe environment.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Linux on Z tools can be used to format, edit, and process the trace records for analysis by
system programmers and network administrators.
Note: OSA Express7S features are available on z15 T01 only, except for OSA-Express7S
25GbE SR (FC 0429).
New build z15 T02 supports OSA-Express6S (all features) and OSA-Express7S 25GbE
SR (FC 0429).
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Note: OSA Express7S features are available on z15 T01 only, except for OSA-Express7S
25GbE SR (FC 0429).
New build z15 T02 supports OSA-Express6S (all features) and OSA-Express7S 25GbE
SR (FC 0429).
Note: Operating system support is required to recognize and use the second port on the
OSA-Express6S 1000BASE-T Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express6S Gigabit Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express5S Gigabit Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express6S 1000BASE-T Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express5S 1000BASE-T Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
With the OSA-ICC function, 3270 emulation for console session connections is integrated in
the z15 through a port on the OSA-Express7S 1000BASE-T, OSA-Express7S GbE,
OSA-Express6S 1000BASE-T, or OSA-Express5S 1000BASE-T.
Note: OSA-ICC supports up to 48 secure sessions per CHPID (the overall maximum of
120 connections is unchanged).
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Checksum offload provides checksum offload for several types of traffic and is supported by
the following features when configured as CHPID type OSD (QDIO mode only):
OSA-Express7S GbE
OSA-Express7S 1000BASE-T Ethernet
OSA-Express6S GbE
OSA-Express6S 1000BASE-T Ethernet
OSA-Express5S GbE
OSA-Express5S 1000BASE-T Ethernet
When checksum is offloaded, the OSA-Express feature runs the checksum calculations for
Internet Protocol version 4 (IPv4) and Internet Protocol version 6 (IPv6) packets. The
checksum offload function applies to packets that go to or come from the LAN.
When multiple IP stacks share an OSA-Express, and an IP stack sends a packet to a next
hop address that is owned by another IP stack that is sharing the OSA-Express,
OSA-Express sends the IP packet directly to the other IP stack. The packet does not have to
be placed out on the LAN, which is termed LPAR-to-LPAR traffic. Checksum offload is
enhanced to support the LPAR-to-LPAR traffic, which was not originally available.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
The use of display OSAINFO (z/OS) or NETSTAT OSAINFO (z/VM) allows the operator to monitor
and verify the current OSA configuration and helps improve the overall management,
serviceability, and usability of OSA-Express cards.
These commands apply to CHPID type OSD. The supported operating systems are listed in
Table 7-8 on page 266.
QDIO data connection isolation allows disabling internal routing for each QDIO connected. It
also provides a means for creating security zones and preventing network traffic between the
zones.
QDIO data connection isolation is supported by all OSA-Express features on z15. The
supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on page 268.
QDIO interface isolation is supported on all OSA-Express features on z15. The supported
operating systems are listed in Table 7-8 on page 266 and Table 7-9 on page 268.
The supported operating systems are listed in Table 7-8 on page 266.
In extending the use of adapter interruptions to OSD (QDIO) channels, the processor
utilization to handle a traditional I/O interruption is reduced. This configuration benefits
OSA-Express TCP/IP support in z/VM, z/VSE, and Linux on Z. The supported operating
systems are listed in Table 7-8 on page 266 and Table 7-9 on page 268.
Each input queue is a unique type of workload, and has unique service and processing
requirements. The IWQ function allows z/OS to preassign the appropriate processing
resources for each input queue. This approach allows multiple concurrent z/OS processing
threads to process each unique input queue (workload), which avoids traditional resource
contention.
IWQ reduces the conventional z/OS processing that is required to identify and separate
unique workloads. This advantage results in improved overall system performance and
scalability.
11 Only OSA-Express6S and OSA-Express5S cards are supported on z15 as carry forward.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
Link aggregation is applicable to CHPID type OSD (QDIO). The supported operating systems
are listed in Table 7-8 on page 266 and Table 7-9 on page 268.
Large send support for IPv6 packets applies to the OSA-Express7S, OSA-Express6S, and
OSA-Express5S11 features (CHPID type OSD) on z15, z14, z13, and z13s.
z13 added support of large send for IPv6 packets (segmentation offloading) for
LPAR-to-LPAR traffic. OSA-Express6S added TCP checksum on large send, which reduces
the cost (CPU time) of error detection for large send.
The supported operating systems are listed in Table 7-8 on page 266 and Table 7-9 on
page 268.
In all cases, the TCP/IP stack determines the best setting based on the current system and
environmental conditions, such as inbound workload volume, processor utilization, and traffic
patterns. It can then dynamically update the settings.
Supported OSA-Express features adapt to the changes, which avoids thrashing and frequent
updates to the OAT. Based on the TCP/IP settings, OSA holds the packets before presenting
them to the host. A dynamic setting is designed to avoid or minimize host interrupts.
The minimum software support levels are described in the following sections. Review the
current PSP buckets to ensure that the latest support levels are known and included as part
of the implementation plan.
In addition, the z15 core implements a Modulo Arithmetic unit in support of Elliptic Curve
Cryptography.
CPACF also is used by several IBM software product offerings for z/OS, such as IBM
WebSphere Application Server for z/OS. For more information, see 6.4, “CP Assist for
Cryptographic Functions” on page 224.
The supported operating systems are listed in Table 7-10 on page 271 and Table 7-11 on
page 272.
12 CPACF hardware is implemented on each z15 core. CPACF functionality is enabled with FC 3863.
Support of Crypto Express6S functions varies by operating system and release and by the
way the card is configured as a coprocessor or an accelerator. For more information, see 6.5,
“Crypto Express7S” on page 230. The supported operating systems are listed in Table 7-10
on page 271 and Table 7-11 on page 272.
Support of Crypto Express6S functions varies by operating system and release and by the
way the card is configured as a coprocessor or an accelerator. For more information, see 6.5,
“Crypto Express7S” on page 230. The supported operating systems are listed in Table 7-10
on page 271 and Table 7-11 on page 272.
Web deliverables
For more information about web-deliverable code on z/OS, see the z/OS downloads website.
For Linux on Z, support is delivered through IBM and the distribution partners. For more
information, see Linux on Z on the IBM developerWorks website.
Despite being a z/OS base component, ICSF functions are generally made available through
web deliverable support a few months after a new z/OS release. Therefore, new functions are
related to an ICSF function modification identifier (FMID) instead of a z/OS version.
ICSF HCR77D1 - Cryptographic Support for z/OS V2R2, V2R3, and V2R4
z/OS V2.2, V2.3, and V2.4 require ICSF Web Deliverable WD19 (HCR77D1) to support the
following features:
Support for CCA 5.5 and CCA 6.3:
– PCI HSM Phase 4 (AES and RSA) and ANSI TR-34
ICSF HCR77D0 - Cryptographic Support for z/OS V2R2 and z/OS V2R3
z/OS V2.2 and V2.3 require ICSF Web Deliverable WD18 (HCR77D0) to support the
following features:
Support for the updated German Banking standard (DK):
– CCA 5.4 & 6.113:
• ISO-4 PIN Blocks (ISO-9564-1)
• Directed keys: A key can either encrypt or decrypt data, but not both.
• Allow AES transport keys to be used to export/import DES keys in a standard ISO
20038 key block. This feature helps with interoperability between CCA and
non-CCA systems.
• Allow AES transport keys to be used to export/import a small subset of AES keys in
a standard ISO 20038 key block. This feature helps with interoperability between
CCA and non-CCA systems.
• Triple-length TDES keys with Control Vectors for increased data confidentiality.
– CCA 6.2: PCI HSM 3K DES - Support for triple length DES keys (standards
compliance)
EP11 Stage 4:
– New elliptic curve algorithms for PKCS#11 signature, key derivation operations:
• Ed448 elliptic curve
• EC25519 elliptic curve
– EP11 Concurrent Patch Apply: Allows service to be applied to the EP11 coprocessor
dynamically without taking the crypto adapter offline (already available for CCA
coprocessors).
– eIDAS compliance: eIDAS cross-border EU regulation for portable recognition of
electronic identification.
13
CCA 5.4 and 6.1 enhancements are also supported for z/OS V2R1 with ICSF HCR77C1 (WD17) with SPEs
(Small Program Enhancements (z/OS continuous delivery model).
The following software enhancements are available in ICSF Web Deliverable HCR77C1:
Crypto Usage Statistics: When enabled, ICSF aggregates statistics that are related to
crypto workloads and logs to an SMF record.
Panel-based CKDS Administration: ICSF added an ISPF, panel-driven interface that
allows interactive administration (View, Create, Modify, and Delete) of CKDS keys.
CICS End User Auditing: When enabled, ICSF retrieves the CICS user identity and
includes it as a log string in the SAF resource check. The user identity is not checked for
access to the resource. Instead, it is included in the resource check (SMF Type 80)
records that are logged for any of the ICSF SAF classes protecting crypto keys and
services (CSFKEYS, XCSFKEY, CRYPTOZ, and CSFSERV).
For more information about ICSF versions and FMID cross-references, see the z/OS: ICSF
Version and FMID Cross Reference, TD103782, abstract that is available at the IBM Techdoc
website.
For PTFs that allow previous levels of ICSF to coexist with the Cryptographic Support for
z/OS 2.1 - z/OS V2R3 (HCR77C1) web deliverable, check below FIXCAT, as shown in the
following example:
IBM.Coexistence.ICSF.z/OS_V2R1-V2R3-HCR77C1
Reporting can be done at an LPAR/domain level to provide more granular reports for capacity
planning and diagnosing problems. This feature requires fix for APAR OA54952.
The supported operating systems are listed in Table 7-10 on page 271.
Policy-driven z/OS Data Set Encryption enables users to perform the following tasks:
De-couple encryption from data classification; encrypt data automatically independent of
labor-intensive data classification work.
Encrypt data immediately and efficiently at the time it is written.
Reduce risks that are associated with mis-classified or undiscovered sensitive data.
Help protect digital assets automatically.
Achieve application transparent encryption.
IBM Db2® for z/OS and IBM Information Management System (IMS) intend to use z/OS Data
Set Encryption.
With z/OS, Data Set Encryption DFSMS enhances data security with support for data set
level encryption by using DFSMS access methods. This function is designed to give users the
ability to encrypt their data sets without changing their application programs.
z/OS Data Set Encryption requires CP Assist for Cryptographic Functions (CPACF).
Considering the significant enhancements that were introduced with z14, the encryption
mode of XTS is used by access method encryption to obtain the best performance possible. It
is not recommended to enable z/OS data set encryption until all sharing systems, fallback,
backup, and DR systems support encryption.
In addition to applying PTFs enabling the support, ICSF configuration is required. The
supported operating systems are listed in Table 7-10 on page 271.
Included in this support is the ability to dynamically control whether a running z/VM system is
encrypting this data. This support protects guest paging data from administrators or users
with access to volumes. Enabled with AES encryption, z/VM Encrypted Paging includes low
overhead by using CPACF.
The supported operating systems are listed in Table 7-10 on page 271.
Because of the potential costs and overheads, most of the organizations avoid the use of
host-based network encryption today. By using enhanced CPACF and SIMD on z15, TLS and
IPSec can use hardware performance gains while benefitting from transparent enablement.
Reduced cost of encryption enables broad use of network encryption.
Although z15 servers do not require any “functional” software, it is recommended to install all
z15 service before upgrading to the new server. The support matrix for z/OS releases and the
Z servers that support them are listed in Table 7-16.
V2R4 X X X - -
a. Server was withdrawn from marketing.
b. The IBM Software Support Services provides the ability for customers to purchase extended
defect support service for z/OS.
14
For example, the use of Crypto Express7S requires the Cryptographic Support for z/OS V2R1 - z/OS V2R3 web
deliverable.
The use of many functions covers fixes that are required to use the capabilities of the IBM
z15™ servers. They are identified by:
IBM.Device.Server.z15-8561.Exploitation
IBM.Device.Server.z15-8562.Exploitation
Support for z15 is provided by using a combination of web deliverables and PTFs, which are
documented in:
– For z15 T01: PSP Bucket Upgrade = 8561DEVICE, Subset = 8561/ZOS.
– For z15 T02: PSP Bucket Upgrade = 8562DEVICE, Subset = 8562/ZOS.
15
For more information, see the Tool to Compare IBM z15 Instruction Mnemonics with Macro Libraries IBM
Technote.
Use the SMP/E REPORT MISSINGFIX command to determine whether any FIXCAT APARs exist
that are applicable and are not yet installed, and whether any SYSMODs are available to
satisfy the missing FIXCAT APARs.
For more information about IBM Fix Category Values and Descriptions, see the IBM Fix
Category Values and Descriptions page of the IBM IT infrastructure website.
Configurations with a Coupling Facility on one of these servers can add a z15 Server to their
Sysplex for a z/OS or a Coupling Facility image. z15 does not support participating in a
Parallel Sysplex with System zEC12/zBC12 and earlier systems.
Each system can use, or not use, internal coupling links, InfiniBand coupling links, or ICA
coupling links independently of what other systems are using.
Coupling connectivity is available only when other systems also support the same type of
coupling. For more information about supported coupling link technologies on z15, see 4.6.4,
“Parallel Sysplex connectivity” on page 195, and the Coupling Facility Configuration Options
white paper.
The following new functions provide performance improvements for applications by using new
z15 instructions:
Vector Programming Enhancements
New z15 hardware instruction support
Packed Decimal support using vector registers
Auto-SIMD enhancements to make use of new data types
To enable the use of new functions, specify ARCH(13) and VECTOR for compilation. The binaries
that are produced by the compiler on z15 can be run only on z15 and above because it uses
the vector facility on z15 for new functions. The use of older versions of the compiler on z15
does not enable new functions.
For more information about the ARCHITECTURE, TUNE, and VECTOR compiler options, see z/OS
V2R2.0 XL C/C++ User’s Guide, SC09-4767.
16
IBM z15 does not support InfiniBand coupling links. More planning might be required to integrate the z15 in a Parallel Sysplex with z14
and z13/z13s servers.
For more information, see Migration from z/OS V2R1 to z/OS V2R2, GA32-0889.
Consider the following points before migrating z/OS 2.3 to IBM z15™:
IBM z/OS V2.3 with z15 requires a minimum of 8 GB of memory. When running as a z/VM
guest or on an IBM System z® Personal Development Tool, a minimum of 2 GB is required
for z/OS V2.3. If the minimum is not met, a warning WTOR is issued at IPL.
Continuing with less than the minimum memory might affect availability. A migration health
check will be introduced at z/OS V2.1 and z/OS V2.2 to warn if the system is configured
with less than 8 GB.
Dynamic splitting and merging of Coordinated Timing Network (CTN) is available with z15.
The z/OS V2.3 real storage manager (RSM) is planned to support a new asynchronous
memory clear operation to clear the data from 1M page frames by using I/O processors
(SAPs). The new asynchronous memory clear operation eliminates the CPU cost for this
operation and help improve performance of RSM first reference page fault processing and
system services, such as IARV64 and STORAGE OBTAIN.
RMF support is provided to collect SMC-D related performance measurements in SMF 73
Channel Path Activity and SMF 74 subtype 9 PCIE Activity records. It also provides these
measurements in the RMF Postprocessor and Monitor III PCIE and Channel Activity
reports. This support is also available on z/OS V2.2 with PTF UA80445 for APAR
OA49113.
HyperSwap support is enhanced to allow RESERVE processing. When a system runs a
request to swap to secondary devices that are managed by HyperSwap, z/OS detects
when RESERVEs are held and ensures that the devices that are swapped also hold the
RESERVE. This enhancement is provided with collaboration from z/OS, GDPS
HyperSwap, and CSM HyperSwap.
z/VM 7.1 includes SPEs shipped for z/VM 6.4, including Virtual Switch Enhanced Load
Balancing, DS8000 z-Thin Provisioning, and Encrypted Paging.
A z/VM Release Status Summary for supported z/VM versions is listed in Table 7-17.
The available PTF for APAR VM65976 provides infrastructure support for ESA/390
compatibility mode within z/VM V6.4. It must be installed on all members of an SSI cluster
before any z/VM V6.4 member of the cluster is run on an IBM z15™ server.
In addition to operating system support, all the stand-alone utilities a client uses must be at a
minimum level or need a PTF.
7.6.5 Capacity
For the capacity of any z/VM logical partition (LPAR) and any z/VM guest, in terms of the
number of Integrated Facility for Linux (IFL) processors and central processors (CPs), real or
virtual, you might want to adjust the number to accommodate the PU capacity of z15 servers.
Consider the following general guidelines when you are migrating z/VSE environment to z15
servers:
Collect reference information before migration
This information includes baseline data that reflects the status of, for example,
performance data, CPU utilization of reference workload, I/O activity, and elapsed times.
This information is required to size z15 and is the only way to compare workload
characteristics after migration.
For more information, see the z/VSE Release and Hardware Upgrade document.
Apply required maintenance for z15
Review the Preventive Service Planning (PSP) bucket 8561DEVICE for z15 and apply the
required PTFs for IBM and independent software vendor (ISV) products.
For the z15, the following metric groups for software licensing are available from IBM,
depending on the software product:
Monthly license charge (MLC)
MLC pricing metrics feature a recurring charge that applies monthly. In addition to the
permission to use the product, the charge includes access to IBM product support during
The subcapacity licensed products are charged monthly based on the highest observed
4-hour rolling average utilization of the logical partitions in which the product runs. The
exception is products that are licensed by using the Select Application License Charge
(SALC) pricing metric. This type of charge requires measuring the utilization and reporting it
to IBM.
The 4-hour rolling average utilization of the logical partition can be limited by a defined
capacity value on the image profile of the partition. This value activates the soft capping
function of the PR/SM, which limits the 4-hour rolling average partition utilization to the
defined capacity value. Soft capping controls the maximum 4-hour rolling average usage (the
last 4-hour average value at every 5-minute interval), but does not control the maximum
instantaneous partition use.
You can also use an LPAR group capacity limit, which sets soft capping by PR/SM for a group
of logical partitions that are running z/OS.
Even by using the soft capping option, the use of the partition can reach up to its maximum
share based on the number of logical processors and weights in the image profile. Only the
4-hour rolling average utilization is tracked, which allows utilization peaks above the defined
capacity value.
Some pricing metrics apply to stand-alone Z servers. Others apply to the aggregation of
multiple Z server workloads within the same Parallel Sysplex.
For more information about WLC and how to combine logical partition utilization, see z/OS
Planning for Sub-Capacity Pricing, SA23-2301.
One of the recent changes in software licensing for z/OS and z/VSE is Multi-Version
Measurement (MVM), which replaced Single Version Charging (SVC), Migration Pricing
Option (MPO), and the IPLA Migration Grace Period.
MVM for z/OS and z/VSE removes time limits for running multiple eligible versions of a
software program. Clients can run different versions of a program simultaneously for an
unlimited duration during a program version upgrade.
Clients can also choose to run multiple different versions of a program simultaneously for an
unlimited duration in a production environment. MVM allows clients to selectively deploy new
software versions, which provides more flexible control over their program upgrade cycles.
For more information, see Software Announcement 217-093, dated February 14, 2017.
Technology Update Pricing for the IBM z15™ extends the software price and performance
that is provided by AWLC and CMLC for z15 servers. The new and revised Transition
Charges for Sysplexes or Multiplexes offerings provide a transition to Technology Update
Pricing for the IBM z15™ for customers who did not yet fully migrate to z15 servers. This
When a z15 server is in an actively coupled Parallel Sysplex or a Loosely Coupled Complex,
you might choose aggregated Advanced Workload License Charges (AWLC) pricing or
aggregated Parallel Sysplex License Charges (PSLC) pricing (subject to all applicable terms
and conditions).
When a z15 server is part of a Multiplex under Country Multiplex Pricing (CMP) terms,
Country Multiplex License Charges (CMLC), Multiplex zNALC (MzNALC), and Flat Workload
License Charges (FWLC) are the only pricing metrics available (subject to all applicable
terms and conditions).
When a z15 server is running z/VSE, you can choose Mid-Range Workload License Charges
(MWLC), which are subject to all applicable terms and conditions.
For more information about AWLC, CMLC, MzNALC, PSLC, MWLC, or the Technology
Update Pricing and Transition Charges for Sysplexes or Multiplexes TTO offerings, see the
IBM z Software Pricing page of the IBM IT infrastructure website.
7.9 References
For more information about planning, see the home pages for the following operating
systems:
z/OS
z/VM
z/VSE
z/TPF
Linux on Z
KVM for IBM Z
IBM z15 servers support many dynamic provisioning features to give clients exceptional
flexibility and control over system capacity and costs.
A key resource for managing client IBM Z servers is the IBM Resource Link website. Once
registered, a client can view product information by clicking Resource Link → Client
Initiated Upgrade Information, and selecting Education. Select your particular product
from the list of available systems.
8.1.1 Overview
Upgrades can be categorized as described in this section.
Tip: An MES provides system upgrades that can result in more enabled processors, a
different central processor (CP) capacity level, more processor drawers, memory,
PCIe+ I/O drawers, and I/O features (physical upgrade). Extra planning tasks are
required for nondisruptive logical upgrades. An MES is ordered through your IBM
representative and installed by IBM service support representatives (IBM SSRs).
Temporary
All temporary upgrades are LICCC-based. The one billable capacity offering is On/Off
Capacity on Demand (On/Off CoD), which can be used for short-term capacity
requirements and are pre-paid or post-paid.
The two replacement capacity offerings available are Capacity Backup (CBU) and
Capacity for Planned Event (CPE).
System Recovery Boost zIIP capacity is a new pre-paid offering for z15 and is intended to
provide temporary zIIP capacity to be used to speed up IPL, shutdown, and stand-alone
dump events.
Activated capacity Capacity that is purchased and activated. Purchased capacity can be greater than the activated
capacity.
Billable capacity Capacity that helps handle workload peaks (expected or unexpected). The one billable offering
that is available is On/Off Capacity on Demand (OOCoD).
Capacity Hardware resources (processor and memory) that can process the workload can be added to
the system through various capacity offerings.
Capacity Backup Capacity Backup allows you to place model capacity or specialty engines in a backup system.
(CBU) CBU is used in an unforeseen loss of system capacity because of an emergency or for Disaster
Recovery testing.
Capacity for Planned Used when temporary replacement capacity is needed for a short-term event. CPE activates
Event (CPE) processor capacity temporarily to facilitate moving systems between data centers, upgrades,
and other routine management tasks. CPE is an offering of CoD.
Capacity levels Can be full capacity or subcapacity. For the z15 system, capacity levels for the CP engine are
7, 6, 5, and 4.
Capacity setting Derived from the capacity level and the number of processors.
For the z15 system, the capacity levels are 7nn, 6yy, 5yy, and 4xx, where xx, yy, or nn indicates
the number of active CPs.
Customer Initiated A web-based facility where you can request processor and memory upgrades by using the IBM
Upgrade (CIU) Resource Link and the system’s Remote Support Facility (RSF) connection.
Capacity on Demand The ability of a system to increase or decrease its performance capacity as needed to meet
(CoD) fluctuations in demand.
Capacity As a component of z/OS Capacity Provisioning, CPM monitors business-critical workloads that
Provisioning are running z/OS on z15 systems.
Manager (CPM)
Customer profile This information is on Resource Link and contains client and system information. A customer
profile can contain information about systems that are related to their IBM customer numbers.
Full capacity CP For z15 servers, feature (CP7) provides full capacity. Capacity settings 7nn are full capacity
feature settings with the ranges of 1 - 99 in decimal and A0 - J0, where A0 represents 100 and J0
represents 190, for capacity level 7nn.
Installed record The LICCC record is downloaded, staged to the Support Element (SE), and is installed on the
central processor complex (CPC). A maximum of eight different records can be concurrently
installed.
Model capacity Shows the current active capacity on the system, including all replacement and billable capacity.
identifier (MCI)
For z15 servers, the model capacity identifier is in the form of 4xx, 5yy, 6yy, or 7nn, where xx,
yy, or nn indicates the number of active CPs:
xx can have a range of 00 - 34. An all IFL or an all ICF system has a capacity level of 400.
yy can have a range of 01 - 34.
1 - 99 in decimal and A0 - J0, where A0 represents 100 and J0 represents 190, for capacity
level 7nn.
Model Permanent Keeps information about the capacity settings that are active before any temporary capacity is
Capacity Identifier activated.
(MPCI)
Model Temporary Reflects the permanent capacity with billable capacity only, without replacement capacity. If no
Capacity Identifier billable temporary capacity is active, MTCI equals the MPCI.
(MTCI)
On/Off Capacity on Represents a function that allows spare capacity in a CPC to be made available to increase the
Demand (CoD) total capacity of a CPC. For example, On/Off CoD can be used to acquire more capacity for
handling a workload peak.
Features on Demand FoD is a centralized way to flexibly entitle features and functions on the system.
(FoD)
Permanent capacity The capacity that a client purchases and activates. This amount might be less capacity than the
total capacity purchased.
Permanent upgrade LICC that is licensed by IBM to enable the activation of applicable computing resources, such
as processors or memory, for a specific CIU-eligible system on a permanent basis.
Purchased capacity Capacity that is delivered to and owned by the client. It can be higher than the permanent
capacity.
Permanent/ The internal representation of a temporary (TER) or permanent (PER) capacity upgrade that is
Temporary processed by the CIU facility. An entitlement record contains the encrypted representation of
entitlement record the upgrade configuration with the associated time limit conditions.
Replacement A temporary capacity that is used for situations in which processing capacity in other parts of
capacity the enterprise is lost. This loss can be a planned event or an unexpected disaster. The two
replacement offerings available are Capacity for Planned Events and Capacity Backup.
Resource Link The IBM Resource Link is a technical support website that provides a comprehensive set of
tools and resources (log in required).
Secondary approval An option that is selected by the client that requires second approver control for each CoD order.
When a secondary approval is required, the request is sent for approval or cancellation to the
Resource Link secondary user ID.
Staged record The point when a record that represents a temporary or permanent capacity upgrade is
retrieved and loaded on the SE disk.
Subcapacity For z15 servers, CP features (CP4, CP5, and CP6) provide reduced capacity relative to the full
capacity CP feature (CP7).
System Recovery Available on z15 servers, the optional System Recovery Boost Record is an orderable feature
Boost Record that provides more capacity for a limited time to enable speeding up shutdown, restart, and
catchup processing for a limited event duration.
Temporary capacity An optional capacity that is added to the current system capacity for a limited amount of time. It
can be capacity that is owned or not owned by the client.
Vital product data Information that uniquely defines system, hardware, software, and microcode elements of a
(VPD) processing system.
Tip: The use of the CIU facility for a system requires that the online CoD buying feature
(FC 9900) is installed on the system. The CIU facility is enabled through the permanent
upgrade authorization feature code (FC 9898).
Considerations: Most of the MESs can be concurrently applied without disrupting the
workload. For more information, see 8.2, “Concurrent upgrades” on page 340. However,
certain MES changes can be disruptive, such as adding PCIE IO drawers.
Memory upgrades that require dual inline memory module (DIMM) changes can be made
nondisruptively if multiple CPC drawers are available and the flexible memory option is
used.
Prepaid OOCoD tokensa: Beginning with IBM z15, new prepaid OOCoD tokens that
are purchased do not carry forward to future systems.
a. Statements by IBM regarding its plans, directions, and intent are subject to change or
withdrawal without notice at the sole discretion of IBM.
CBU
This offering allows you to replace model capacity or specialty engines in a backup system
that is used in an unforeseen loss of system capacity because of a disaster.
CPE
This offering allows you to replace model capacity or specialty engines because of a
relocation of workload during system migrations or a data center move.
System Recovery Boost Record
This offering allows you to add up to 20 zIIPs for use with the System Recovery Boost
facility. System Recovery Boost provides temporary extra capacity for CP workloads to
allow rapid shutdown, restart, and recovery of eligible systems. System Recovery Boost
records are prepaid, licensed for a one-year period, and can be renewed at any time.
CBU, CPE, and System Recovery Boost Records can be ordered by using the CIU
application through Resource Link or by contacting your IBM marketing representative.
Replacement capacity
When processing capacity is lost in part of an enterprise, replacement capacity can be
activated. It allows you to activate any PU type up to your authorized limit.
This capability is based on the flexibility of the design and structure, which allows concurrent
hardware installation and Licensed Internal Control Code (LICC) configuration changes.
The subcapacity models allow more configuration granularity within the family. The added
granularity is available for models that are configured with up to 34 CPs, and provides 102
extra capacity settings. Subcapacity models provide for CP capacity increase in two
dimensions that can be used together to deliver configuration granularity. The first dimension
is adding CPs to the configuration. The second is changing the capacity setting of the CPs
currently installed to a higher model capacity identifier.
z15 servers allow the concurrent and nondisruptive addition of processors to a running logical
partition (LPAR). As a result, you can have a flexible infrastructure to which you can add
capacity. This function is supported by z/OS, z/VM, and z/VSE. This addition is made by using
one of the following methods:
With planning ahead for the future need of extra processors. Reserved processors can be
specified in the LPAR’s profile. When the extra processors are installed, the number of
active processors for that LPAR can be increased without the need for a partition
reactivation and initial program load (IPL).
Another (easier) way is to enable the dynamic addition of processors through the z/OS
LOADxx member. Set the DYNCPADD parameter in member LOADxx to ENABLE.
Another function concerns the system assist processor (SAP). When more SAPs are
concurrently added to the configuration, the SAP-to-channel affinity is dynamically remapped
on all SAPs on the system to rebalance the I/O configuration.
The model capacity identifier can be concurrently changed. Concurrent upgrades can be
performed for permanent and temporary upgrades.
Tip: A drawer feature upgrade can be performed concurrently only for a max 34 or a max
71 machine if feature codes 2271 or 2272 were ordered with the base machine.
Note: Plan ahead memory is not offered on new z15 orders. It can be carried forward only
from z13 and z14 machines.
1 The z15 zero CP MCI is 400. This setting applies to an all-IFL or all-ICF system.
The concurrent I/O upgrade capability can be better used if a future target configuration is
considered during the initial configuration.
Important: The LICCC-based PU conversions require that at least one PU (CP, ICF, or
IFL), remains unchanged. Otherwise, the conversion is disruptive. The PU conversion
generates a LICCC that can be installed concurrently in two steps:
1. Remove the assigned PU from the configuration.
2. Activate the newly available PU as the new PU type.
LPARs also might have to free the PUs to be converted. The operating systems must include
support to configure processors offline or online so that the PU conversion can be done
nondisruptively.
Considerations: Client planning and operator action are required to use concurrent PU
conversion. Consider the following points about PU conversion:
It is disruptive if all current PUs are converted to different types.
It might require individual LPAR outages if dedicated PUs are converted.
The CIU facility is controlled through the permanent upgrade authorization FC 9898. A
prerequisite to FC 9898 is the online CoD buying feature code (FC 9900). Although FC 9898
can be installed on your z15 servers at any time, often it is added when ordering a z15 server.
After you place an order through the CIU facility, you receive a notice that the order is ready
for download. You can then download and apply the upgrade by using functions that are
available through the Hardware Management Console (HMC), along with the RSF. After all of
the prerequisites are met, the entire process (from ordering to activation of the upgrade) is
performed by the client and does not require any onsite presence of IBM SSRs.
As part of the setup, provide one resource link ID for configuring and placing CIU orders and,
if required, a second ID as an approver. The IDs are then set up for access to the CIU
support. The CIU facility allows upgrades to be ordered and delivered much faster than
through the regular MES process.
To order and activate the upgrade, log on to the IBM Resource Link website and start the CIU
application to upgrade a system for processors or memory. You can request a client order
approval to conform to your operational policies. You also can allow the definition of more IDs
to be authorized to access the CIU. More IDs can be authorized to enter or approve CIU
orders, or only view orders.
Permanent upgrades
Permanent upgrades can be ordered by using the CIU facility. Through the CIU facility, you
can generate online permanent upgrade orders to concurrently add processors (CPs, ICFs,
zIIPs, IFLs, and SAPs) and memory, or change the model capacity identifier. You can do so
up to the limits of the installed processor drawers on a system.
Temporary upgrades
The base model z15 server describes permanent and dormant capacity by using the capacity
marker and the number of PU features that are installed on the system. Up to eight temporary
offerings can be present. Each offering includes its own policies and controls, and each can
be activated or deactivated independently in any sequence and combination. Although
multiple offerings can be active at any time, only one On/Off CoD offering can be active at
any time if enough resources are available to fulfill the offering specifications.
Temporary upgrades are represented in the system by a record. All temporary upgrade
records are on the SE hard disk drive (HDD). The records can be downloaded from the RSF
or installed from portable media. At the time of activation, you can control everything locally.
API
HMC Application
Query Activation
R3 Up to 8 records installed
R1 R2 R4 R5 R6 R7 R8 and active
Dormant
capacity
Base Change permanent capacity
model through CIU or MES order
Purchased
capacity
The authorization layer enables administrative control over the temporary offerings. The
activation and deactivation can be driven manually or under the control of an application
through a documented application programming interface (API).
By using the API approach, you can customize at activation time the resources that are
necessary to respond to the current situation up to the maximum that is specified in the order
record. If the situation changes, you can add or remove resources without having to go back
to the base configuration. This process eliminates the need for temporary upgrade
specifications for all possible scenarios.
R1 R2 R3 R4
R3 R1
CBU CPE
R4 R4 R3 R3 R1
CBU CBU CBU CBU CPE
R2 R2 R2 R2 R2 R3
OOCoD OOCoD OOCoD OOCoD OOCoD CBU
As shown in Figure 8-2, if R2, R3, and R1 are active at the same time, only parts of R1 can be
activated because not enough resources are available to fulfill all of R1. When R2 is
deactivated, the remaining parts of R1 can be activated as shown.
Temporary capacity can be billable as On/Off CoD, or replacement capacity as CBU, CPE, or
System Recovery Boost. Consider the following points:
On/Off CoD is a function that enables concurrent and temporary capacity growth of the
system.
On/Off CoD can be used for client peak workload requirements, for any length of time, and
includes a daily hardware and maintenance charge. The software charges can vary
according to the license agreement for the individual products. For more information,
contact your IBM Software Group representative.
On/Off CoD can concurrently add processors (CPs, ICFs, zIIPs, IFLs, and SAPs),
increase the model capacity identifier, or both. It can do so up to the limit of the installed
processor drawers of a system. It is restricted to twice the installed capacity. On/Off CoD
requires a contractual agreement between you and IBM.
You decide whether to pre-pay or post-pay On/Off CoD. Capacity tokens that are inside the
records are used to control activation time and resources.
CBU is a concurrent and temporary activation of more CPs, ICFs, zIIPs, IFLs, and SAPs;
or an increase of the model capacity identifier; or both.
Note: CBU cannot be used for peak workload management in any form.
Note: CPE cannot be used for peak load management of client workload or for a
disaster situation.
The CPE feature requires unused capacity to be available on installed processor drawers
of the backup system. The capacity must be available as unused PUs, as a possibility to
increase the model capacity identifier on a subcapacity system, or as both.
A CPE contract must be in place before the special code that enables this capability can
be loaded on the system. The standard CPE contract provides for one 3-day planned
activation at a specific date. For more information, contact your IBM representative.
The System Recovery Boost Record allows a concurrent activation of extra zIIPs.
The System Recovery Boost Record offering can be used to provide extra zIIP capacity
that can be used by the System Recovery Boost facility. You might want to consider the
use of this offering if your server is a full capacity model (7nn) and can benefit from more
CPs during system shutdown and restart. The capacity is delivered as zIIPs that can
perform CP work during the boost periods for an LPAR.
A System Recovery Boost Record contract must be in place before the special code that
enables this capability can be loaded on the system. The standard contract provides for
one 6-hour activation for the specific purpose of System Recovery Boost only. For more
information, contact your IBM representative.
Activation of System Recovery Boost Record does not change the MCI of your system.
Permanent MES CPs, ICFs, zIIPs, IFLs, SAPs, processor Installed by IBM SSRs
drawer, memory, and I/Os
Online permanent CPs, ICFs, zIIPs, IFLs, SAPs, and Performed through the CIU facility
upgrade memory
Temporary On/Off CoD CPs, ICFs, zIIPs, IFLs, and SAPs Performed through the OOCoD facility
CBU CPs, ICFs, zIIPs, IFLs, and SAPs Activated through model conversion
CPE CPs, ICFs, zIIPs, IFLs, and SAPs Activated through model conversion
2
Other adapter types, such as zHyperlink, Coupling Express LR, and Remote Direct Memory Access (RDMA) over
Converged Ethernet (RoCE), also can be added to the PCIe+ I/O drawers through an MES.
To better use the MES upgrade function, carefully plan the initial configuration to allow a
concurrent upgrade to a target configuration. The availability of PCIe+ I/O drawers improves
the flexibility to perform unplanned I/O configuration changes concurrently.
The Store System Information (STSI) instruction gives more useful and detailed information
about the base configuration and temporary upgrades.
The model and model capacity identifiers that are returned by the STSI instruction are
updated to coincide with the upgrade. For more information, see “Store System Information
instruction” on page 378.
Upgrades: An MES provides the physical upgrade, which results in more enabled
processors, different capacity settings for the CPs, and more memory, I/O ports, I/O
adapters, and I/O drawers. Extra planning tasks are required for nondisruptive logical
upgrades. For more information, see “Guidelines to avoid disruptive upgrades” on
page 380.
Limits: The sum of CPs, inactive CPs, ICFs, zIIPs, IFLs, unassigned IFLs, and SAPs
cannot exceed the maximum limit of PUs available for client use. The number of zIIPs
cannot exceed twice the number of purchased CPs.
8561 T01 Max 34 MCI 708 (Drawer 0, 34 CPs) with Feat #2271
CP CP CP CP CP CP CP CP SP SP SP SP SP SP SP SP SP
SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP
MES + 31 CPs
(+ 1 Drawer)
CP CP CP CP CP SP SP SP SP SP SP SP SP SP SP SP SP SP SP
SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP
MES + 1 CP + 2 IFLs
CP CP CP CP CP CP SP SP SP SP SP SP SP SP SP SP SP SP SP
SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP SP IFL IFL
A model T01 max 34 (one processor drawer), model capacity identifier 708 (eight CPs), is
concurrently upgraded to a model T01 max 71 (two processor drawers), with MCI 739 (39
CPs). The model upgrade requires adding a processor drawer and assigning and activating
39 PUs as CPs. Then, model max 71, MCI 739, is concurrently upgraded to a capacity
identifier 740 (40 CPs) with two IFLs. This process is done by assigning and activating three
more unassigned PUs (one as CP and two as IFLs). If needed, more LPARs can be created
concurrently to use the newly added processors.
The example that is shown in Figure 8-3 was used to show how the addition of PUs as CPs
and IFLs and the addition of a processor drawer works. The addition of a processor drawer to
a z15 Max 34 upgrades the machine to Max 71.
After the second CPC drawer addition, CPC drawer 0 has 34 configurable PUs and CPC
drawer 1 has 37 configurable PUs, which allows 71 PUs to be characterized on the new Max
71 configuration.
Table 8-3 Number of processors that are supported by the operating system
Operating system Number of processors that are supported
z/OS V2R2 190 PUs per z/OS LPAR in non-SMT mode and 128 PUs per
z/OS LPAR in SMT mode. For both, the PU total is the sum of CPs
and zIIPs.
z/OS V2R3 190 PUs per z/OS LPAR in non-SMT mode and 128 PUs per
z/OS LPAR in SMT mode. For both, the PU total is the sum of CPs
and zIIPs.
z/OS V2R4 190 PUs per z/OS LPAR in non-SMT mode and 128 PUs per
z/OS LPAR in SMT mode. For both, the PU total is the sum of CPs
and zIIPs.
z/TPF 86 CPs.
Linux on IBM Z -190 CPs Linuxa supports 256 cores without SMT and 128 cores with SMT
(256 threads).
a. Supported Linux on Z distributions (for more information, see Chapter 7, “Operating system
support” on page 253).
Software charges, which are based on the total capacity of the system on which the software
is installed, are adjusted to the new capacity after the MES upgrade.
Software products that use Workload License Charges (WLC) or Taylor Fit Pricing (TFP)
might not be affected by the system upgrade. Their charges are based on partition usage, not
on the system total capacity. For more information about WLC, see 7.8, “Software licensing”
on page 328.
The Flexible Memory Feature is available to allow better control over future memory
upgrades. For more information about flexible memory features, see 2.5.7, “Flexible Memory
Option” on page 65.
If the z15 server is a multiple processor drawer configuration, you can use the EDA feature to
remove a processor drawer and add DIMM memory cards. It can also be used to upgrade the
installed memory cards to a larger capacity size. You can then use LICCC to enable the extra
memory.
Concurrency: Upgrades that require DIMM changes can be concurrent by using the EDA
feature. Planning is required to see whether this option is a viable for your configuration.
The use of the flexible memory option ensures that EDA can work with the least disruption.
You can also add memory by concurrently adding a second processor drawer with sufficient
memory into the configuration and then using LICCC to enable that memory. Changing
DIMMs in a single CPC drawer system is disruptive.
An LPAR can dynamically take advantage of a memory upgrade if reserved storage is defined
to that LPAR. The reserved storage is defined to the LPAR as part of the image profile.
Reserved memory can be configured online to the LPAR by using the LPAR dynamic storage
reconfiguration (DSR) function. DSR allows a z/OS operating system image and z/VM
partitions to add reserved storage to their configuration if any unused storage exists.
The nondisruptive addition of storage to a z/OS and z/VM partition requires the correct
operating system parameters to be set. If reserved storage is not defined to the LPAR, the
LPAR must be deactivated, the image profile changed, and the LPAR reactivated. This
process allows the extra storage resources to be available to the operating system image.
For more information about PCIe+ I/O drawers, see 4.2, “I/O system overview” on page 156.
The number of PCIe+ I/O drawers that can be present in a z15 server depends on how many
CPC drawers are present and on whether the configuration includes the Bulk Power
Assembly (BPA) offering. It also depends on whether the CPC drawer reserve features are
present.
The number of drawers for specific configuration options are listed in Table 8-4 on page 352.
It is based on no CPC drawer reserve options being configured.
Note: The maximum number of IO drawers in the table is reduced by 1 for each CPC
drawer reserve feature present.
Depending on the number of I/O features, the configurator determines the number of PCIe+
I/O drawers required.
To better use the MES for I/O capability, carefully plan the initial configuration to allow
concurrent upgrades up to the target configuration.
If a PCIe+ I/O drawer is added to a z15 server and original features must be physically moved
to another PCIe+ I/O drawer, original card moves are disruptive.
z/VSE, z/TPF, and Linux on Z do not provide dynamic I/O configuration support. Although
installing the new hardware is done concurrently, defining the new hardware to these
operating systems requires an IPL.
Tip: z15 servers feature a hardware system area (HSA) of 256 GB. z14 servers have a 192
GB HSA. HSA is not part of the client-purchased memory.
A staged record can be removed without installing it. An FoD record can be installed only
completely; no selective feature or partial record installation is available. The features that are
installed are merged with the CPC LICCC after activation.
An FoD record can be installed only once. If it is removed, a new FoD record is needed to
reinstall. A remove action cannot be undone.
Tip: Accurate planning and the definition of the target configuration allows you to maximize
the value of these plan-ahead features.
Adding permanent upgrades to a system through the CIU facility requires that the permanent
upgrade enablement feature (FC 9898) is installed on the system. A permanent upgrade
might change the system model capacity identifier (4xx, 5yy, 6yy, or 7nn) if more CPs are
requested, or if the capacity identifier is changed as part of the permanent upgrade. If
necessary, more LPARs can be created concurrently to use the newly added processors.
Software charges that are based on the total capacity of the system on which the software is
installed are adjusted to the new capacity after the permanent upgrade is installed. Software
products that use WLC or customers with TFP might not be affected by the system upgrade
because their charges are based on LPAR usage rather than system total capacity. For more
information about WLC, see 7.8, “Software licensing” on page 328.
The CIU facility process on IBM Resource Link is shown in Figure 8-5.
Customer ibm.com/servers/resourcelink
Internet
Online
permanent
Optional customer order
secondary order
approval
Remote Support
Facility
Figure 8-5 Permanent upgrade order example
The following sample sequence shows how to start an order on the IBM Resource Link:
1. Sign on to Resource Link.
2. Select Customer Initiated Upgrade from the main Resource Link page. Client and
system information that is associated with the user ID are displayed.
3. Select the system to receive the upgrade. The current configuration (PU allocation and
memory) is shown for the selected system.
4. Select Order Permanent Upgrade. The Resource Link limits the options to those options
that are valid or possible for the selected configuration (system).
5. After the target configuration is verified by the system, accept or cancel the order. An order
is created and verified against the pre-established agreement.
6. Accept or reject the price that is quoted. A secondary order approval is optional. Upon
confirmation, the order is processed. The LICCC for the upgrade is available within hours.
z15
Figure 8-6 CIU-eligible order activation example
8.4.1 Ordering
IBM Resource Link provides the interface that enables you to order a concurrent upgrade for
a system. You can create, cancel, or view the order, and view the history of orders that were
placed through this interface.
Configuration rules enforce that only valid configurations are generated within the limits of the
individual system. Warning messages are issued if you select invalid upgrade options. The
process allows only one permanent CIU-eligible order for each system to be placed at a time.
The initial view of the Machine profile on Resource Link is shown in Figure 8-7.
The number of CPs, ICFs, zIIPs, IFLs, SAPs, memory size, and unassigned IFLs on the
current configuration are displayed on the left side of the page.
Resource Link retrieves and stores relevant data that is associated with the processor
configuration, such as the number of CPs and installed memory cards. It allows you to select
only those upgrade options that are deemed valid by the order process. It also allows
upgrades only within the bounds of the currently installed hardware.
When the order is available for download, you receive an email that contains an activation
number. You can then retrieve the order by using the Perform Model Conversion task from the
SE, or through the Single Object Operation to the SE from an HMC.
The window provides several possible options. If you select the Retrieve and apply data
option, you are prompted to enter the order activation number to start the permanent
upgrade, as shown in Figure 8-9.
8.5.1 Overview
The capacity for CPs is expressed in millions of service units (MSUs). Capacity for speciality
engines is expressed in number of speciality engines. Capacity tokens are used to limit the
resource consumption for all types of processor capacity.
Each speciality engine type features its own tokens, and each On/Off CoD record includes
separate token pools for each capacity type. During the ordering sessions on Resource Link,
select how many tokens of each type to create for an offering record. Each engine type must
include tokens for that engine type to be activated. Capacity that has no tokens cannot be
activated.
When resources from an On/Off CoD offering record that contains capacity tokens are
activated, a billing window is started. A billing window is always 24 hours. Billing occurs at
the end of each billing window.
The resources that are billed are the highest resource usage inside each billing window for
each capacity type. An activation period is one or more complete billing windows. The
activation period is the time from the first activation of resources in a record until the end of
the billing window in which the last resource in a record is deactivated.
At the end of each billing window, the tokens are decremented by the highest usage of each
resource during the billing window. If any resource in a record does not have enough tokens
to cover usage for the next billing window, the entire record is deactivated.
Note: On/Off CoD requires that the Online CoD Buying feature (FC 9900) is installed on
the system that you want to upgrade.
The On/Off CoD to Permanent Upgrade Option gives customers a window of opportunity to
assess capacity additions to your permanent configurations by using On/Off CoD. If a
purchase is made, the hardware On/Off CoD charges during this window (three days or less)
are waived. If no purchase is made, you are charged for the temporary use.
The resources eligible for temporary use are CPs, ICFs, zIIPs, IFLs, and SAPs. The
temporary addition of memory and I/O ports or adapters is not supported.
Unassigned PUs that are on the installed processor drawers can be temporarily and
concurrently activated as CPs, ICFs, zIIPs, IFLs, and SAPs through LICCC. You can assign
PUs up to twice the currently installed CP capacity, and up to twice the number of ICFs, zIIPs,
or IFLs. An On/Off CoD upgrade cannot change the system capacity feature. The addition of
new processor drawers is not supported. However, the activation of an On/Off CoD upgrade
can increase the model capacity identifier (4xx, 5yy, 6yy, or 7nn).
The Capacity Provisioning Manager Console is provided as part of z/OS MF, which provides a
browser interface for managing z/OS systems.
8.5.3 Ordering
Concurrently installing temporary capacity by ordering On/Off CoD is possible in the following
manner:
CP features equal to the MSU capacity of installed CPs
IFL features up to the number of installed IFLs
ICF features up to the number of installed ICFs
zIIP features up to the number of installed zIIPs
SAPs - up to 8
On/Off CoD can be ordered as prepaid or postpaid. A prepaid On/Off CoD offering record
contains resource descriptions, MSUs, speciality engines, and tokens that describe the total
capacity that can be used. For CP capacity, the token contains MSU-days. For speciality
engines, the token contains speciality engine-days.
A postpaid On/Off CoD offering record contains resource descriptions, MSUs, speciality
engines, and can contain capacity tokens that denote MSU-days and speciality engine-days.
When resources in a postpaid offering record without capacity tokens are activated, those
resources remain active until they are deactivated, or until the offering record expires. The
record often expires 180 days after its installation.
When resources in a postpaid offering record with capacity tokens are activated, those
resources must include enough capacity tokens to allow the activation for an entire billing
window (24 hours). The resources remain active until they are deactivated, until all of the
resource tokens are used, or until the record expires. The record usually expires 180 days
after its installation. If one capacity token type is used, resources from the entire record are
deactivated.
For example, for a z15 server with capacity identifier 502 (two CPs), a capacity upgrade
through On/Off CoD can be delivered in the following ways:
Add CPs of the same capacity setting. With this option, the model capacity identifier can
be changed to a 503, which adds another CP to make it a three-way CP. It can also be
changed to a 504, which adds two CPs, making it a four-way CP.
Change to a different capacity level of the current CPs and change the model capacity
identifier to a 602 or 702. The capacity level of the CPs is increased, but no other CPs are
added. The 502 also can be temporarily upgraded to a 603, which increases the capacity
level and adds another processor. The capacity setting 434 does not have an upgrade
path through On/Off CoD because you cannot reduce the number of CPs and a 534 is
more than twice the capacity.
Use the Large System Performance Reference (LSPR) information to evaluate the capacity
requirements according to your workload type. For more information about LSPR data for
current IBM processors, see the Large Systems Performance Reference for IBM Z page of
the IBM Systems website.
The On/Off CoD hardware capacity is charged on a 24-hour basis. A grace period is granted
at the end of the On/Off CoD day. This grace period allows up to an hour after the 24-hour
billing period to change the On/Off CoD configuration for the next 24-hour billing period or
deactivate the current On/Off CoD configuration. The times when the capacity is activated
and deactivated are maintained in the z15 server and sent back to the support systems.
If On/Off capacity is active, On/Off capacity can be added without having to return the system
to its original capacity. If the capacity is increased multiple times within a 24-hour period, the
charges apply to the highest amount of capacity active in that period.
If more capacity is added from an active record that contains capacity tokens, the systems
checks whether the resource has enough capacity to be active for an entire billing window
(24 hours). If that criteria is not met, no extra resources are activated from the record.
If necessary, more LPARs can be activated concurrently to use the newly added processor
resources.
To participate in this offering, you must accept contractual terms for purchasing capacity
through the Resource Link, establish a profile, and install an On/Off CoD enablement feature
on the system. Later, you can concurrently install temporary capacity up to the limits in On/Off
CoD and use it for up to 180 days.
Monitoring occurs through the system call-home facility. An invoice is generated if the
capacity is enabled during the calendar month. You are billed for the use of temporary
capacity until the system is returned to the original configuration. Remove the enablement
code if the On/Off CoD support is no longer needed.
On/Off CoD orders can be pre-staged in Resource Link to allow multiple optional
configurations. The pricing of the orders is done at the time that you order them, and the
pricing can vary from quarter to quarter. Staged orders can have different pricing.
When the order is downloaded and activated, the daily costs are based on the pricing at the
time of the order. The staged orders do not have to be installed in the order sequence. If a
staged order is installed out of sequence and later a higher-priced order is staged, the daily
cost is based on the lower price.
Another possibility is to store multiple On/Off CoD LICCC records on the SE with the same or
different capacities, which gives you greater flexibility to enable quickly needed temporary
capacity. Each record is easily identified with descriptive names, and you can select from a
list of records that can be activated.
Resource Link provides the interface to order a dynamic upgrade for a specific system. You
can create, cancel, and view the order. Configuration rules are enforced, and only valid
configurations are generated based on the configuration of the individual system. After you
complete the prerequisites, orders for the On/Off CoD can be placed. The order process uses
the CIU facility on Resource Link.
This test can have a maximum duration of 24 hours, which commences upon the activation of
any capacity resource that is contained in the On/Off CoD record. Activation levels of capacity
can change during the 24-hour test period. The On/Off CoD test automatically stops at the
end of the 24-hour period.
You also can perform administrative testing. No capacity is added to the system, but you can
test all the procedures and automation for the management of the On/Off CoD facility.
The example order that is shown in Figure 8-11 is an On/Off CoD order for 0% more CP
capacity (system is at capacity level 7), and for two more ICFs and two more zIIPs. The
maximum number of CPs, ICFs, zIIPs, and IFLs is limited by the current number of available
unused PUs of the installed processor drawers. The maximum number of SAPs is determined
by the model number and the number of available PUs on the already installed processor
drawers.
To finalize the order, you must accept Terms and Conditions for the order, as shown in
Figure 8-12.
If the On/Off CoD offering record does not contain resource tokens, you must deactivate the
temporary capacity manually. Deactivation is done from the SE and is nondisruptive.
Depending on how the capacity was added to the LPARs, you might be required to perform
tasks at the LPAR level to remove it. For example, you might have to configure offline any CPs
that were added to the partition, deactivate LPARs that were created to use the temporary
capacity, or both.
On/Off CoD orders can be staged in Resource Link so that multiple orders are available. An
order can be downloaded and activated only once. If a different On/Off CoD order is required
or a permanent upgrade is needed, it can be downloaded and activated without having to
restore the system to its original purchased capacity.
In support of automation, an API is if allows the activation of the On/Off CoD records. The
activation is performed from the HMC, and requires specifying the order number. With this
API, automation code can be used to send an activation command along with the order
number to the HMC to enable the order.
8.5.6 Termination
A client is contractually obligated to end the On/Off CoD right-to-use feature when a transfer
in asset ownership occurs. A client also can choose to end the On/Off CoD right-to-use
feature without transferring ownership.
Removing FC 9898 ends the right to use the On/Off CoD. This feature cannot be ordered if a
temporary session is active. Similarly, the CIU enablement feature cannot be removed if a
temporary session is active. When the CIU enablement feature is removed, the On/Off CoD
right-to-use feature is simultaneously removed. Reactivating the right-to-use feature subjects
the client to the terms and fees that apply then.
Monitoring
When you activate an On/Off CoD upgrade, an indicator is set in vital product data. This
indicator is part of the call-home data transmission, which is sent on a scheduled basis. A
time stamp is placed into the call-home data when the facility is deactivated. At the end of
each calendar month, the data is used to generate an invoice for the On/Off CoD that was
used during that month.
Software
Software Parallel Sysplex license charge (PSLC) clients are billed at the MSU level that is
represented by the combined permanent and temporary capacity. All PSLC products are
billed at the peak MSUs that are enabled during the month, regardless of usage. Clients with
WLC licenses are billed by product at the highest four-hour rolling average for the month. In
this instance, temporary capacity does not increase the software bill until that capacity is
allocated to LPARs and used.
Results from the STSI instruction reflect the current permanent and temporary CPs. For more
information, see “Store System Information instruction” on page 378.
z/OS Capacity Provisioning is delivered as part of the z/OS MVS™ Base Control Program
(BCP).
The Provisioning Manager monitors the workload on a set of z/OS systems and organizes the
provisioning of extra capacity to these systems when required. You define the systems to be
observed in a domain configuration file.
The details of extra capacity and the rules for its provisioning are stored in a policy file. These
two files are created and maintained through the Capacity Provisioning Management Console
(CPMC).The operational flow of Capacity Provisioning is shown in Figure 8-13 on page 366.
PR/SM
RMF
Ethernet WLM RMF RMF WLM
Switch
DDS
CIM
Capacity server
Provisioning CIM
(SNMP)
Manager (CPM) server
Provisioning
Policy z/OS MF
Capacity Provisioning
Management Console
z/OS console(s)
The z/OS WLM manages the workload by goals and business importance on each z/OS
system. WLM metrics are available through existing interfaces, and are reported through IBM
Resource Measurement Facility (RMF) Monitor III, with one RMF gatherer for each z/OS
system.
Sysplex-wide data aggregation and propagation occur in the RMF Distributed Data Server
(DDS). The RMF Common Information Model (CIM) providers and associated CIM models
publish the RMF Monitor III data.
CPM retrieves critical metrics from one or more z/OS systems’ CIM structures and protocols.
CPM communicates to local and remote SEs and HMCs by using the Simple Network
Management Protocol (SNMP).
CPM can see the resources in the individual offering records and the capacity tokens. When
CPM activates resources, a check is run to determine whether enough capacity tokens
remain for the specified resource to be activated for at least 24 hours. If insufficient tokens
remain, no resource from the On/Off CoD record is activated.
If a capacity token is used during an activation that is driven by the CPM, the corresponding
On/Off CoD record is deactivated prematurely by the system. This process occurs even if the
CPM activates this record, or parts of it. However, you do receive warning messages if
capacity tokens are close to being fully used.
You receive the messages five days before a capacity token is fully used. The five days are
based on the assumption that the consumption is constant for the five days. You must put
operational procedures in place to handle these situations. You can deactivate the record
manually, allow it occur automatically, or replenish the specified capacity token by using the
Resource Link application.
z15 HMC
z/OS MF Browser Interface
SE
Capacity Provisioning SE
Manager Console Linux on
CF System z CF
Sysplex A z/VM Sysplex B
CF CF
Sysplex B z/OS image Sysplex A
z/OS image
Sysplex A
Sysplex A
z/OS image
Capacity
Provisioning z/OS
z/OSimage
image z/OS
z/OSimage
image
Manager Sysplex A Sysplex A
z/OS image
z/OS image
z/OS
z/OSimage
image z/OS image Sysplex B
Sysplex B
The CPD configuration defines the CPCs and z/OS systems that are controlled by an
instance of the CPM. One or more CPCs, sysplexes, and z/OS systems can be defined into a
domain. Although sysplexes and CPCs do not have to be contained in a domain, they must
not belong to more than one domain.
CPM operates in the following modes, which allows four different levels of automation:
Manual mode
Use this command-driven mode when no CPM policy is active.
Analysis mode
In analysis mode, CPM processes capacity-provisioning policies and informs the operator
when a provisioning or deprovisioning action is required according to policy criteria.
Also, the operator determines whether to ignore the information or to manually upgrade or
downgrade the system by using the HMC, SE, or available CPM commands.
Several reports are available in all modes that contain information about the workload,
provisioning status, and the rationale for provisioning guidelines. User interfaces are provided
through the z/OS console and the CPMC application.
The provisioning policy defines the circumstances under which more capacity can be
provisioned (when, which, and how). The criteria features the following elements:
A time condition is when provisioning is allowed:
– Start time indicates when provisioning can begin.
– Deadline indicates that provisioning of more capacity is no longer allowed.
– End time indicates that deactivation of capacity must begin.
A workload condition is which work qualifies for provisioning. It can have the following
parameters:
– The z/OS systems that can run eligible work.
– The importance filter indicates eligible service class periods, which are identified by
WLM importance.
– Performance Index (PI) criteria:
• Activation threshold: PI of service class periods must exceed the activation
threshold for a specified duration before the work is considered to be suffering.
• Deactivation threshold: PI of service class periods must fall below the deactivation
threshold for a specified duration before the work is considered to no longer be
suffering.
– Included service classes are eligible service class periods.
– Excluded service classes are service class periods that must not be considered.
Tip: If no workload condition is specified, the full capacity that is described in the policy
is activated and deactivated at the start and end times that are specified in the policy.
Provisioning scope is how much more capacity can be activated and is expressed in
MSUs.
The number of zIIPs must be one specification per CPC that is part of the CPD and are
specified in MSUs.
The maximum provisioning scope is the maximum extra capacity that can be activated for
all the rules in the CPD.
In the specified time interval, the provisioning rule is that up to the defined extra capacity can
be activated if the specified workload is behind its objective.
The rules and conditions are named and stored in the Capacity Provisioning Policy.
For more information about z/OS Capacity Provisioning functions, see z/OS MVS Capacity
Provisioning User’s Guide, SC34-2661.
The provisioning management routines can interrogate the installed offerings, their content,
and the status of the content of the offering. To avoid the decrease in capacity, create only
one On/Off CoD offering on the system by specifying the maximum allowable capacity. The
CPM can then, when an activation is needed, activate a subset of the contents of the offering
sufficient to satisfy the demand. If more capacity is needed later, the Provisioning Manager
can activate more capacity up to the maximum allowed increase.
Multiple offering records can be pre-staged on the SE hard disk. Changing the content of the
offerings (if necessary) is also possible.
Remember: CPM controls capacity tokens for the On/Off CoD records. In a situation
where a capacity token is used, the system deactivates the corresponding offering record.
Therefore, you must prepare routines for catching the warning messages about capacity
tokens being used, and have administrative procedures in place for such a situation.
The messages from the system begin five days before a capacity token is fully used. To
avoid capacity records being deactivated in this situation, replenish the necessary capacity
tokens before they are used.
Important: The base System Recovery Boost capability is BUILT INTO z15 firmware and
does not require ordering more features. System Recovery Boost Upgrade (consisting of
FC 9930 and FC 6802) is an optional, orderable feature that provides more temporary zIIP
capacity for use during boost periods. Consider the following points:
FC 9930 is not required to use the base System Recovery Boost capability.
FC 9930 is only needed if more zIIP temporary capacity is required.
The System Recovery Boost Upgrade optional feature is offered with z15 servers to provide
more capacity for System Recovery Boost processing. For example, if a z/OS operating
system change requires a sequence of system shutdowns and restarts, and these
procedures can benefit from extra CPU capacity, the System Recovery Boost record can be
used to activate more zIIPs on the server at the commencement of the change window.
These zIIPs are used by the z/OS systems to run general CP work during the boost periods.
Note: At least one permanent zIIP record must be present for ordering System
Recovery Boost Upgrade (FC 6802).
Important: The System Recovery Boost Upgrade record is for System Recovery Boost
capacity only, and cannot be used for peak workload management. The customer must
deactivate the boost capacity at the end of the system restart procedure.
The zIIP processors that can be activated by System Recovery Boost record come from the
“dark core” capacity on the server. They can be added to a z15 server nondisruptively.
The base system configuration must have sufficient memory and channels to accommodate
the potential requirements of the larger capacity system.
Note: The System Recovery Boost configuration is activated temporarily and provides up
to a maximum of 20 extra zIIPs to the system’s original, permanent configuration and can
violate the 2:1 zIIP rule. The number of zIIPs that can be activated is limited by the unused
capacity that is available on the system.
When activating the System Recovery Boost record, the extra zIIPs are added to the zIIP pool
when they are activated. Review the LPAR zIIP assignments and weights in the image
profiles to ensure that the LPAR can use the extra capacity when it becomes available.
Configure a quantity for the initial and reserved zIIPs in the image profile so that extra zIIPs
can be brought online dynamically when the boost record is activated. Also consider adjusting
the LPAR zIIP weight.
When the system recovery event ends, the system must be returned to its original
configuration. The System Recovery Boost Upgrade record can be used only once and must
be replenished before it can be used again.
A System Recovery Boost Upgrade contract (through FC 9930) must be in place before the
special code that enables this capability can be installed on the system.
The feature codes are calculated automatically when the CPE offering is configured. Whether
the eConfig tool or the Resource Link is used, a target configuration must be ordered. The
configuration consists of a model identifier, several speciality engines, or both. Based on the
target configuration, several feature codes from the list are calculated automatically, and a
CPE offering record is constructed.
CPE is intended to replace capacity that is lost within the enterprise because of a planned
event, such as a facility upgrade or system relocation.
Note: CPE is intended for short duration events that last a maximum of three days.
After each CPE record is activated, you can access dormant PUs on the system for which you
have a contract, as described by the feature codes. Processor units can be configured in any
combination of CP or specialty engine types (zIIP, SAP, IFL, and ICF). At the time of CPE
activation, the contracted configuration is activated. The general rule of two zIIPs for each
configured CP is enforced for the contracted configuration.
The processors that can be activated by CPE come from the available unassigned PUs on
any installed processor drawer. CPE features can be added to a z15 server nondisruptively. A
one-time fee is applied for each CPE event. This fee depends on the contracted configuration
and its resulting feature codes. Only one CPE record can be ordered at a time.
The base system configuration must include sufficient memory and channels to
accommodate the potential requirements of the large CPE-configured system. Ensure that all
required functions and resources are available on the system where CPE is activated. These
functions and resources include CF LEVELs for coupling facility partitions, memory, and
cryptographic functions, and include connectivity capabilities.
The CPE configuration is activated temporarily and provides more PUs in addition to the
system’s original, permanent configuration. The number of extra PUs is predetermined by the
number and type of feature codes that are configured, as described by the feature codes. The
number of PUs that can be activated is limited by the unused capacity that is available on the
system; for example:
A z15 Max 71 with 26 CPs, and no IFLs or ICFs has 45 unassigned PUs available.
A z15 Max 145 with 38 CPs, 1 IFL, and 1 ICF has 105 unassigned PUs available.
A CPE contract must be in place before the special code that enables this capability can be
installed on the system. CPE features can be added to a z15 server nondisruptively.
Important: CBU is for disaster and recovery purposes only. It cannot be used for peak
workload management or for a planned event.
8.9.1 Ordering
The CBU process allows for CBU to activate CPs, ICFs, zIIPs, IFLs, and SAPs. To use the
CBU process, a CBU enablement feature (FC 9910) must be ordered and installed. You must
order the quantity and type of PU that you require by using the following feature codes:
FC 6805: More CBU test activations
FC 6817: Total CBU years ordered
FC 6818: CBU records that are ordered
FC 6820: Single CBU CP-year
FC 6821: 25 CBU CP-year
FC 6822: Single CBU IFL-year
FC 6823: 25 CBU IFL-year
FC 6824: Single CBU ICF-year
FC 6825: 25 CBU ICF-year
FC 6828: Single CBU zIIP-year
FC 6829: 25 CBU zIIP-year
FC 6830: Single CBU SAP-year
FC 6831: 25 CBU SAP-year
FC 6832: CBU replenishment
The CBU entitlement record (FC 6818) contains an expiration date that is established at the
time of the order. This date depends on the quantity of CBU years (FC 6817). You can extend
your CBU entitlements through the purchase of more CBU years.
The number of FC 6817 per instance of FC 6818 remains limited to five. Fractional years are
rounded up to the nearest whole integer when calculating this limit.
3
All new CBU contract documents contain new CBU test terms to allow execution of production workload during
CBU test. CBU clients must sign the IBM client Agreement Amendment for IBM Z Capacity Backup Upgrade Tests
(US form #Z125-8145).
FC 6805 allows for ordering more tests in increments of one. The maximum number of tests
that is allowed is 15 for each FC 6818.
The processors that can be activated by CBU come from the available unassigned PUs on
any installed processor drawer. The maximum number of CBU features that can be ordered is
190. The number of features that can be activated is limited by the number of unused PUs on
the system; for example:
A z15 Max 34 with Capacity Model Identifier 401 can activate up to 34 CBU features.
These CBU features can be used to change the capacity setting of the CPs, and to
activate unused PUs.
A z15 Max 71 with 15 CPs, 4 IFLs, and 1 ICF has 51 unused PUs available. It can activate
up to 51 CBU features.
The ordering system allows for over-configuration in the order. You can order up to 190 CBU
features, regardless of the current configuration. However, at activation, only the capacity that
is installed can be activated. At activation, you can decide to activate only a subset of the
CBU features that are ordered for the system.
Subcapacity makes a difference in the way that the CBU features are completed. On the
full-capacity models, the CBU features indicate the amount of extra capacity that is needed. If
the amount of necessary CBU capacity is equal to four CPs, the CBU configuration is four
CBU CPs.
The subcapacity models feature multiple capacity settings of 4xx, 5yy, or 6yy. The standard
models use the capacity setting 7nn. To change the capacity setting, the number of CBU CPs
must be equal to or greater than the number of CPs in the base configuration.
For example, if the base configuration is a two-way 402, providing a CBU configuration of a
four-way of the same capacity setting requires two CBU feature codes. If the required CBU
capacity changes the capacity setting of the CPs, going from model capacity identifier 402 to
a CBU configuration of a four-way 504 requires four CBU feature codes with a capacity setting
of 5yy.
If the capacity setting of the CPs is changed, more CBU features are required, not more
physical PUs. Therefore, your CBU contract requires more CBU features when the capacity
setting of the CPs is changed.
CBU can add CPs through LICCC only, and the z15 server must have the correct number of
installed processor drawers to allow the required upgrade. CBU can change the model
capacity identifier to a higher value than the base setting (4xx, 5yy, or 6yy), but the CBU
feature cannot decrease the capacity setting.
A CBU contract must be in place before the special code that enables this capability can be
installed on the system. CBU features can be added to a z15 server nondisruptively. For each
system enabled for CBU, the authorization to use CBU is available for 1 - 5 years.
The alternative configuration is activated temporarily, and provides more capacity than the
system’s original, permanent configuration. At activation time, determine the capacity that you
require for that situation. You can decide to activate only a subset of the capacity that is
specified in the CBU contract.
When the emergency is over (or the CBU test is complete), the system must be returned to its
original configuration. The CBU features can be deactivated at any time before the expiration
date. Failure to deactivate the CBU feature before the expiration date can cause the system to
downgrade resources gracefully to the original configuration. The system does not deactivate
dedicated engines, or the last of in-use shared engines.
Planning: CBU for processors provides a concurrent upgrade. This upgrade can result in
more enabled processors, changed capacity settings that are available to a system
configuration, or both. You can activate a subset of the CBU features that are ordered for
the system. Therefore, more planning and tasks are required for nondisruptive logical
upgrades. For more information, see “Guidelines to avoid disruptive upgrades” on
page 380.
For more information, see the IBM Z Capacity on Demand User’s Guide, SC28-6846.
CBU activation
CBU is activated from the SE by using the HMC and SSO to the SE, by using the Perform
Model Conversion task, or through automation by using the API on the SE or the HMC.
During a real disaster, use the Activate CBU option to activate the 90-day period.
Image upgrades
After CBU activation, the z15 server can have more capacity, more active PUs, or both. The
extra resources go into the resource pools and are available to the LPARs. If the LPARs must
increase their share of the resources, the LPAR weight can be changed or the number of
logical processors can be concurrently increased by configuring reserved processors online.
The operating system must concurrently configure more processors online. If necessary,
more LPARs can be created to use the newly added capacity.
CBU deactivation
To deactivate the CBU, the extra resources must be released from the LPARs by the
operating systems. In some cases, this process involves varying the resources offline. In
other cases, it can mean shutting down operating systems or deactivating LPARs. After the
resources are released, the same facility on the HMC/SE is used to turn off CBU. To
deactivate CBU, select the Undo temporary upgrade option from the Perform Model
Conversion task on the SE.
Tip: The CBU test activation is done the same way as the real activation; that is, by using
the same SE Perform a Model Conversion window and selecting the Temporary
upgrades option. The HMC windows were changed to avoid accidental real CBU
activations by setting the test activation as the default option.
The test CBU must be deactivated in the same way as the regular CBU. Failure to deactivate
the CBU feature before the expiration date can cause the system to degrade gracefully back
to its original configuration. The system does not deactivate dedicated engines or the last
in-use shared engine.
CBU example
An example of a CBU operation is shown in Figure 8-15. The permanent configuration is a
504, and a record contains seven CP CBU features. During an activation, multiple target
configurations are available. With 7 CP CBU features, you can add up to 7CPs within the
same MCI, which allows the activation of a 506, 507, through to a 511(the blue path).
Alternatively, 4 CP CBU features can be used to change the MCI (in the example from a 504
to a 704) and then add the remaining 3 CP CBU features to upgrade to a 707 (the red path).
3
7xx 707
4
6xx
7
7
5xx 504 511
4xx
N-Way 1 2 3 4 5 6 7 8 9 10 11 12 13
Note: System Recovery Boost record does not affect model capacity identifier.
The GDPS service is for z/OS only, or for z/OS in combination with Linux on Z.
z15 servers allow concurrent upgrades, which means that dynamically adding capacity to the
system is possible. If the operating system images that run on the upgraded system do not
require disruptive tasks to use the new capacity, the upgrade is also nondisruptive. This
process avoids power-on resets (POR), LPAR deactivation, and IPLs.
If the concurrent upgrade is intended to satisfy an image upgrade to an LPAR, the operating
system that is running in this partition must concurrently configure more capacity online. z/OS
operating systems include this capability. z/VM can concurrently configure new processors
and I/O devices online, and memory can be dynamically added to z/VM partitions.
If the concurrent upgrade is intended to satisfy the need for more operating system images,
more LPARs can be created concurrently on the z15 system. These LPARs include all
resources that are needed. These extra LPARs can be activated concurrently.
These enhanced configuration options are available through the HSA, which is an IBM
reserved area in system memory.
Linux operating systems, in general, cannot add more resources concurrently. However,
Linux, and other types of virtual machines that run under z/VM, can benefit from the z/VM
capability to nondisruptively configure more resources online (processors and I/O).
With z/VM, Linux guests can manipulate their logical processors by using the Linux CPU
hotplug daemon. The daemon can start and stop logical processors that are based on the
Linux load average value. The daemon is available in Linux SLES 10 SP2 and later, and in
Red Hat Enterprise Linux (RHEL) V5R4 and up.
8.10.1 Components
The following components can be added, depending on the considerations as described in
this section:
PUs
Memory
I/O
Cryptographic adapters
Special features
PUs
CPs, ICFs, zIIPs, IFLs, and SAPs can be added concurrently to a z15 server if unassigned
PUs are available on any installed processor drawer. The number of zIIPs cannot exceed
twice the number of CPs. The z15 allows the concurrent addition of a second and third
processor drawer if the CPC reserve features are installed.
If necessary, more LPARs can be created concurrently to use the newly added processors.
The Coupling Facility Control Code (CFCC) can also configure more processors online to
coupling facility LPARs by using the CFCC image operations window.
By using the previously defined reserved memory, z/OS operating system images, and z/VM
partitions, you can dynamically configure more memory online. This process allows
nondisruptive memory upgrades. Linux on Z supports Dynamic Storage Reconfiguration.
I/O
I/O features can be added concurrently if all the required infrastructure (I/O slots and PCIe
Fanouts) is present in the configuration. PCIe+ I/O drawers can be added concurrently
without planning if free space is available in one of the frames and the configuration permits.
Dynamic I/O configurations are supported by certain operating systems (z/OS and z/VM),
which allows nondisruptive I/O upgrades. Dynamic I/O reconfiguration on a stand-alone
coupling facility system is also possible using the Dynamic I/O activation for stand-alone CF
CPCs features
Cryptographic adapters
Crypto Express7S features can be added concurrently if all the required infrastructure is in
the configuration.
Special features
Special features such as zHyperlink, Coupling Express LR, and RoCE features can be added
concurrently if all infrastructure is available in the configuration.
Enabling and using the extra processor capacity is not apparent to most applications.
However, certain programs depend on processor model-related information, such as ISV
products. Consider the effect on the software that is running on a z15 server when you
perform any of these configuration upgrades.
Processor identification
The following instructions are used to obtain processor information:
Store System Information (STSI) instruction
The STSI instruction can be used to obtain information about the current execution
environment and any processing level below the current environment. It can be used to
obtain processor model and model capacity identifier information from the basic machine
configuration form of the system information block (SYSIB). It supports concurrent
upgrades and is the recommended way to request processor information.
Store CPU ID (STIDP) instruction
STIDP returns information that identifies the execution environment, system serial
number, and machine type.
Note: To ensure unique identification of the configuration of the issuing CPU, use the STSI
instruction specifying basic machine configuration (SYSIB 1.1.1).
The model capacity identifier contains the base capacity, On/Off CoD, and CBU. The Model
Permanent Capacity Identifier and the Model Permanent Capacity Rating contain the base
capacity of the system. The Model Temporary Capacity Identifier and Model Temporary
Capacity Rating contain the base capacity and On/Off CoD.
For more information about the STSI instruction, see z/Architecture Principles of Operation,
SA22-7832.
For more information about the STIDP instruction, see z/Architecture Principles of Operation,
SA22-7832.
You can minimize the need for these outages by carefully planning and reviewing “Guidelines
to avoid disruptive upgrades” on page 380.
One major client requirement was to eliminate the need for a client authorization connection
to the IBM Resource Link system when activating an offering. This requirement is met by the
z13, z14, and z15 servers.
After the offerings are installed on the z15 SE, they can be activated at any time at the client’s
discretion. No intervention by IBM or IBM personnel is necessary. In addition, the activation of
CBU does not require a password.
Resource usage (and therefore, financial exposure) can be controlled by using capacity
tokens in the On/Off CoD offering records.
The CPM is an example of an application that uses the CoD APIs to provision On/Off CoD
capacity that is based on the requirements of the workload. The CPM cannot control other
offerings.
For more information about any of the topics in this chapter, see IBM Z Capacity on Demand
User’s Guide, SC28-6943.
Note: Throughout this chapter, z15 refers to IBM z15 Model T01 (Machine Type 8561),
unless otherwise specified.
The key objectives, in order of priority, are to ensure data integrity, computational integrity,
reduce or eliminate unscheduled outages, reduce scheduled outages, reduce planned
outages, and reduce the number of Repair Actions.
RAS can be accomplished with improved concurrent replace, repair, and upgrade functions
for processors, memory, drawers, and I/O. RAS also extends to the nondisruptive capability
for installing Licensed Internal Code (LIC) updates. In most cases, a capacity upgrade can be
concurrent without a system outage. As an extension to the RAS capabilities, environmental
controls are implemented in the system to help reduce power consumption and meet cooling
requirements.
The following overriding RAS requirements are principles as shown in Figure 9-1:
Inclusion of existing (or equivalent) RAS characteristics from previous generations.
Learn from current field issues and addressing the deficiencies.
Understand the trend in technology reliability (hard and soft) and ensure that the RAS
design points are sufficiently robust.
Invest in RAS design enhancements (hardware and firmware) that provide IBM Z and
Customer valued differentiation.
9.2 Technology
This section introduces some of the RAS features that are incorporated in the z15 design.
z15 processor memory and cache structure are shown in Figure 9-2.
1 Key in storage error uncorrected: Indicates that the hardware cannot repair a storage key that was in error.
9.3 Structure
The z15 server is built in a new form factor of 1- 4 19-inch frames. The z15 server can be
delivered as an air-cooled and water-cooled system and fulfills the requirements for an
ASHRAE A3 environment.
The z15 server can have up to 11 PCIe+ I/O drawers when delivered with Bulk Power
Assembly (BPA) and 12 PCIe+ I/O drawers when delivered with Power Distribution Unit
(PDU). The structure changes to the z15 server are done with the following goals:
Enhanced system modularity
Standardization to enable rapid integration
Platform simplification
Cables are keyed to ensure that correct lengths are plugged. Plug detection ensures correct
location, and custom latches ensure retention. Further improvements to the fabric bus include
symmetric multiprocessing (SMP) cables that connect the drawers.
To improve field-replaceable unit (FRU) isolation, TDR techniques are applied to the SMP
cables, between chips (PU-PU, and PU-SC), and between the PU chips and dual inline
memory modules (DIMMs).
Firmware was updated to improve filtering and resolution of errors that do not require action.
Enhanced integrated sparing in processor cores, cache relocates, N+1 SEEPROM and POL
N+2 redundancies, and DRAM marking also are incorporated to reduce touches. The
following RAS enhancements reduce service touches:
Improved error resolution to enable filtering
Enhanced integrated sparing in processor cores
Cache relocates
N+1 SEEPROM
N+2 POL
DRAM marking
(Dynamic) Spare lanes for PU-SC, PU-PU, PU-mem, and SC-SMP fabric
N+1 radiator pumps, controllers, blowers, and sensors
N+1 Ethernet switches
N+1 Support Element (SE) (with N+1 SE power supplies)
Redundant SEEPROM on memory DIMM
Redundant temperature sensor (one SEEPROM and one temperature sensor per I2C
bus)
FICON forward error correction
The new IBM Z Channel Subsystem Function performs periodic polling from the channel
to the end points for the logical paths that are established and reduces the number of
useless Repair Actions (RAs).
The RDP data history is used to validate Predictive Failure Algorithms and identify Fibre
Channel Links with degrading signal strength before errors start to occur. The new Fibre
Channel Extended Link Service (ELS) retrieves signal strength.
FICON Dynamic Routing
FICON Dynamic Routing (FIDR) enables the use of storage area network (SAN) dynamic
routing policies in the fabric. With the z15 server, FICON channels are no longer restricted
to the use of static routing policies for inter-switch links (ISLs) for cascaded FICON
directors.
FICON Dynamic Routing dynamically changes the routing between the channel and
control unit based on the Fibre Channel Exchange ID. Each I/O operation has a unique
exchange ID. FIDR is designed to support static SAN routing policies and dynamic routing
policies.
FICON Dynamic Routing can help clients reduce costs by providing the following features:
– Share SANs between their FICON and FCP traffic.
– Improve performance because of SAN dynamic routing policies that better use all the
available ISL bandwidth through higher use of the ISLs,
– Simplify management of their SAN fabrics by using static routing policies that assign
different ISL routes with each power-on-reset (POR), which makes the SAN fabric
performance difficult to predict.
The difference between scheduled outages and planned outages might not be obvious. The
general consensus is that scheduled outages occur sometime soon. The time frame is
approximately two weeks.
Planned outages are outages that are planned well in advance and go beyond this
approximate two-week time frame. This chapter does not distinguish between scheduled and
planned outages.
Preventing unscheduled, scheduled, and planned outages was addressed by the IBM Z
system design for many years.
z15 servers have a fixed size HSA of 256 GB (up from 192 GB on z14). This size helps
eliminate pre-planning requirements for HSA and provides the flexibility to update dynamically
the configuration. You can perform the following tasks dynamically:2
Add a logical partition (LPAR)
Add a logical channel subsystem (LCSS)
Add a subchannel set
Add a logical PU to an LPAR
Add a cryptographic coprocessor
Remove a cryptographic coprocessor
Enable I/O connections
Swap processor types
Add memory
Add a physical processor
By addressing the elimination of planned outages, the following tasks also are possible:
Concurrent driver upgrades
Concurrent and flexible customer-initiated upgrades
2
Some pre-planning considerations might exist. For more information, see Chapter 8, “System upgrades” on
page 333.
The EDA procedure and careful planning help ensure that all the resources are still available
to run critical applications in an (n-1) drawer configuration. This process allows you to avoid
planned outages. Consider the flexible memory option to provide more memory resources
when you are replacing a drawer. For more information about flexible memory, see 2.5.7,
“Flexible Memory Option” on page 65.
To minimize the effect on current workloads, ensure that sufficient inactive physical resources
exist on the remaining drawers to complete a drawer removal. Also, consider deactivating
non-critical system images, such as test or development LPARs. After you stop these
non-critical LPARs and free their resources, you might find sufficient inactive resources to
contain critical workloads while completing a drawer replacement.
The following configurations especially enable the use of the EDA function. These z15
features need enough spare capacity so that they can cover the resources of a fenced or
isolated drawer. This configuration imposes limits on the following number of the client-owned
PUs that can be activated when one drawer within a model is fenced:
A maximum of 34 client PUs are configured on the Max34.
A maximum of 71 client PUs are configured on the Max71.
A maximum of 108 client PUs are configured on the Max108.
A maximum of 145 client PUs are configured on the Max145.
A maximum of 190 client PUs are configured on the Max190.
No special feature codes are required for PU and model configuration.
Feature Max34 to Max145 each have 4 SAPs in each drawer. Max190 has in total 22
standard SAPs.
The flexible memory option delivers physical memory so that 100% of the purchased
memory increment can be activated even when one drawer is fenced.
The I/O connectivity must also support drawer removal. Most of the paths to the I/O feature
redundant I/O interconnect support in the I/O infrastructure (drawers) that enable connections
through multiple fanout cards.
If sufficient resources are not present on the remaining drawers, certain non-critical LPARs
might need to be deactivated. One or more PUs or storage might need to be configured
offline to reach the required level of available resources. Plan to address these possibilities to
help reduce operational errors.
Include the planning as part of the initial installation and any follow-on upgrade that modifies
the operating environment. A client can use the Resource Link machine information report to
determine the number of drawers, active PUs, memory configuration, and channel layout.
If the z15 server is installed, click Prepare for Enhanced Drawer Availability in the Perform
Model Conversion window of the EDA process on the Hardware Management Console
(HMC). This task helps you determine the resources that are required to support the removal
of a drawer with acceptable degradation to the operating system images.
The EDA process determines which resources, including memory, PUs, and I/O paths, are
free to allow for the removal of a drawer. You can run this preparation on each drawer to
determine which resource changes are necessary. Use the results as input in the planning
stage to help identify critical resources.
With this planning information, you can examine the LPAR configuration and workload
priorities to determine how resources might be reduced and still allow the drawer to be
concurrently removed.
When you perform the review, document the resources that can be made available if the EDA
is used. The resources on the drawers are allocated during a POR of the system and can
change after that process. Perform a review when changes are made to z15 servers, such as
adding drawers, PUs, memory, or channels. Also, perform a review when workloads are
added or removed, or if the HiperDispatch feature was enabled and disabled since the last
time you performed a POR.
Processor availability
Processor resource availability for reallocation or deactivation is affected by the type and
quantity of the resources in use, such as:
Total number of PUs that are enabled through LICCC
PU definitions in the profiles that can be dedicated and dedicated reserved or shared
Active LPARs with dedicated resources at the time of the drawer repair or replacement
To maximize the PU availability option, ensure that sufficient inactive physical resources are
on the remaining drawers to complete a drawer removal.
Memory availability
Memory resource availability for reallocation or deactivation depends on the following factors:
Physically installed memory
Image profile memory allocations
Amount of memory that is enabled through LICCC
Flexible memory option
Virtual Flash Memory if enabled and configured
For more information, see 2.7.2, “Enhanced drawer availability (EDA)” on page 73.
Preparation: The preparation step does not reallocate any resources. It is used only to
record client choices and produce a configuration file on the SE that is used to run the
concurrent drawer replacement operation.
The preparation step can be done in advance. However, if any changes to the configuration
occur between the preparation and the physical removal of the drawer, you must rerun the
preparation phase.
The process can be run multiple times because it does not move any resources. To view the
results of the last preparation operation, click Display Previous Prepare Enhanced Drawer
Availability Results from the Perform Model Conversion window in the SE.
The preparation step can be run without performing a drawer replacement. You can use it to
dynamically adjust the operational configuration for drawer repair or replacement before IBM
SSR activity. The Perform Model Conversion window in you click Prepare for Enhanced
Drawer Availability is shown in Figure 9-4 on page 399.
After you click Prepare for Enhanced Drawer Availability, the Enhanced Drawer Availability
window opens. Select the drawer that is to be repaired or upgraded; then, select OK, as
shown in Figure 9-5. Only one target drawer can be selected at a time.
The system verifies the resources that are required for the removal, determines the required
actions, and presents the results for review. Depending on the configuration, the task can take
from a few seconds to several minutes.
The preparation step determines the readiness of the system for the removal of the targeted
drawer. The configured processors and the memory in the selected drawer are evaluated
against unused resources that are available across the remaining drawers. The system also
analyzes I/O connections that are associated with the removal of the targeted drawer for any
single path I/O connectivity.
If insufficient resources are available, the system identifies the conflicts so that you can free
other resources.
Preparation tabs
The results of the preparation are presented for review in a tabbed format. Each tab indicates
conditions that prevent the EDA option from being run. The following tab selections are
available:
Processors
Memory
Single I/O
Single Domain I/O
Single Alternate Path I/O
Only the tabs that feature conditions that prevent the drawer from being removed are
displayed. Each tab indicates the specific conditions and possible options to correct them.
For example, the preparation identifies single I/O paths that are associated with the removal
of the selected drawer. These paths must be varied offline to perform the drawer removal.
After you address the condition, rerun the preparation step to ensure that all the required
conditions are met.
After you review the reassignment results and make any necessary adjustments, click OK.
The final results of the reassignment, which include the changes that are made as a result of
the review, are displayed (see Figure 9-7). These results are the assignments when the
drawer removal phase of the EDA is completed.
By understanding the system configuration and the LPAR allocation for memory, PUs, and
I/O, you can make the best decision about how to free the necessary resources to allow for
drawer removal.
The preparation process can be run multiple times to ensure that all conditions are met. It
does not reallocate any resources; instead, it produces only a report. The resources are not
reallocated until the Perform Drawer Removal process is started.
Review the results. The result of the preparation task is a list of resources that must be made
available before the drawer replacement can occur.
Reserved storage: If you plan to use the EDA function with z/OS LPARs, set up
reserved storage and an RSU value. Use the RSU value to specify the number of
storage units that are to be kept free of long-term fixed storage allocations. This
configuration allows for storage elements to be varied offline.
When correctly configured, z15 servers support concurrently activating a selected new LIC
Driver level. Concurrent activation of the selected new LIC Driver level is supported only at
specific released sync points. Concurrently activating a selected new LIC Driver level
anywhere in the maintenance stream is not possible. Certain LIC updates do not allow a
concurrent update or upgrade.
The EDM function does not eliminate the need for planned outages for driver-level upgrades.
Upgrades might require a system level or a functional element scheduled outage to activate
the new LIC. The following circumstances require a scheduled outage:
Specific complex code changes might dictate a disruptive driver upgrade. You are alerted
in advance so that you can plan for the following changes:
– Design data or hardware initialization data fixes
– CFCC release level change
OSA CHPID code changes might require PCHID Vary OFF/ON to activate new code.
Crypto code changes might require PCHID Vary OFF/ON to activate new code.
Note: zUDX clients should contact their User Defined Extensions (UDX) provider
before installing Microcode Change Levels (MCLs). Any changes to Segments 2 and 3
from a previous MCL level might require a change to the client’s UDX. Attempting to
install an incompatible UDX at this level results in a Crypto checkstop.
Consider the following points for managing native PCIe adapters microcode levels:
Updates to the Resource Group require all native PCIe adapters that are installed in that
RG to be offline.
Updates to the native PCIe adapter require the adapter to be offline. If the adapter is not
defined, the MCL session automatically installs the maintenance that is related to the
adapter.
The PCIe native adapters are configured with Function IDs (FIDs) and might need to be
configured offline when changes to code are needed. To help alleviate the number of
adapters (and FIDs) that are affected by the Resource Group code update, z15 have four
Resource Groups per system (CPC).
Note: Other adapter types, such as FICON Express, OSA Express, and Crypto Express
that are installed in the PCIe+ I/O drawer are not effected because they are not managed
by the Resource Groups.
The adapter locations and PCHIDs for the four Resource Groups are listed in Table 9-2.
1 100-101,108-109,110-111,118-119
2 140-141,148-149,150-151,158-159
3 180-181,188-189,190-191,198-199
4 1C0-1C1,1C8-1C9,1D0-1D1,1D8-1D9
5 200-201,208-209,210-211,218-219
8 2C0-2C1,2C8-2C9,2D0-2D1,2D8-2D9
9 300-301,308-309,310-311,318-319
10 340-341,348-349,350-351,358-359
11 380-381,388-389,390-391,398-399
12 3C0-3C1,3C8-3C9,3D0-3D1,3D8-3D9
1 120-121,128-129,130-131,138-139
2 160-161,168-169,170-171,178-179
3 1A0-1A1,1A8-1A9,1B0-1B1,1B8-1B9
4 1E0-1E1,1E8-1E9,1F0-1F1,1F8-1F9
5 220-221,228-229,230-231,238-239
8 2E0-2E1,2E8-2E9,2F0-2F1,2F8-2F9
9 320-321,328-329,330-331,338-339
10 360-361,368-369,370-371,378-379
11 3A0-3A1,3A8-3A9,3B0-3B1,3B8-3B9
12 3E0-3E1,3E8-3E9,3F0-3F1,3F8-3F9
1 104-105,10C-10D,114-115,11C-11D
2 144-145,14C-14D,154-155,15C-15D
3 184-185,18C-18D,194-195,19C-19D
4 1C4-1C5,1CC-1CD,1D4-1D5,1DC-1DD
5 204-205,20C-20D,214-215,21C-21D
8 2C4-2C5,2CC-2CD,2D4-2D5,2DC-2DD
9 304-305,30C-30D,314-315,31C-31D
10 344-345,34C-34D,354-355,35C-35D
11 384-385,38C-38D,394-395,39C-39D
12 3C4-3C5,3CC-3CD,3D4-3D5,3DC-3DD
1 124-125,12C-12D,134-135,13C-13D
2 164-165,16C-16D,174-175,17C-17D
3 1A4-1A5,1AC-1AD,1B4-1B5,1BC-1BD
4 1E4-1E5,1EC-1ED,1F4-1F5,1FC-1FD
5 224-225,22C-22D,234-235,23C-23D
8 2E4-2E5,2EC-2ED,2F4-2F5,2FC-2FD
9 324-325,32C-32D,334-335,33C-33D
10 364-365,36C-36D,374-375,37C-37D
11 3A4-3A5,3AC-3AD,3B4-3B5,3BC-3BD
12 3E4-3E5,3EC-3ED,3F4-3F5,3FC-3FD
This chapter describes the newest elements for the HMC and SE.
Note: The Help function is a good starting point to get more information about all of the
functions that can be used by the HMC and SE. The Help feature is available by clicking
Help from a drop-down menu that appears when you click your user ID.
Note: Throughout this chapter, z15 refers to IBM z15 Model T01 (Machine Type 8651),
unless otherwise specified.
The HMC runs a set of management applications. On z15, the HMC can be a stand-alone
computer (mini-tower or rack mounted), or (new with z15) can run (as a Hardware
Management Appliance) on the SEs hardware (1U rack-mounted servers that are integrated
in z15 A frame).
The SEs are two 1U servers integral to the z15 frame. One SE is the primary SE (active) and
the other is the alternative SE (backup). As with the HMCs, the SEs are closed systems, and
no other applications can be installed on them.
The HMC is used to set up, manage, monitor, and operate one or more CPCs. It manages
IBM Z hardware, its logical partitions (LPARs), and provides support applications. At least one
HMC is required to operate an IBM Z. An HMC can manage multiple Z CPCs, and can be at a
local or a remote site.
When tasks are performed at the HMC, the commands are routed to the active SE of the z15.
The SE then issues those commands to their CPC. One HMC can control up to 100 SEs and
1 SE can be controlled by up to 32 HMCs.
New with z15, a number of “traditional” SE-only functions moved to HMC tasks. On z15, these
functions appear as native HMC tasks, but run on the SE. These HMC functions run in
parallel with Single Object Operations (SOOs), which simplifies and streamlines system
management. For more information about SOOs, see “Single Object Operations” on
page 431.
HMC 2.14.1 includes DPM 3.2, with enhanced storage management capabilities. HMC driver
41 (Version 2.15.0) with MCLs added support for FCP and FICON support to the DPM
Release 4.0. DPM is a mode of operation that enables customers with little or no knowledge
of IBM Z technology to set up the system efficiently and with ease.
For more information, see IBM Knowledge Center, click the search engine window, and enter
DPM.
Updating the driver or applying HMC MCLs requires planning to ensure that these
operations are not disruptive to the CPC by ensuring availability of the primary or
alternative SE appliance.
The HMC FC0063 (2461-SE3) is a 1U server that can be ordered with an optional IBM 1U
rack-mounted tray that features a monitor and a keyboard/pointing device (KMM FC 0154).
The HMC system unit and the KMM tray must be rack-mounted in two adjacent 1U locations
in the “ergonomic zone” between 21U and 26U in a standard 19-inch rack.
The rack-mounted HMC can be installed in a customer-provided rack (it cannot be mounted in
the z15 CPC frames). Three C13 power receptacles are required: two for the system unit and
one for the display and keyboard, as shown in Figure 10-5.
For more information about the mini-KMM and how to attach it to the A frame, see IBM Z 8561
Installation Manual for Physical Planning, GC28-7002.
10.2.4 New service and functional operations for HMCs and SEs
Because a DVD drive is not available on the HMC or the SE, this section describes some
service and functional operations for HMC Version 2.15.0.
Firmware load
Firmware can be loaded by using the following modes:
USB
If the HMC and SE firmware is shipped on a USB drive when a new system is ordered, the
load procedure is similar that was used with a DVD load.
Electronic
If USB load is not allowed, or when FC 0846 is ordered, an ISO image is used for a
firmware load over a local area network (LAN). New build HMCs include HMC/SE ISO
images and the HMC provides the server function for loading the code. ISO images can
also be downloaded through zRSF or FTP from IBM.
Important: The ISO image server (HMC) must be on the same subnet with the target
system; that is, the system to be loaded with HMC or SE code from ISO images.
An HMC with Version 2.15.0 can support N-2 IBM Z server generations. Some functions that
are available on Version 2.15.0 and later are supported only when the HMC is connected to
an IBM Z with Driver 41 (z15).
The SE drivers and versions that are supported by the z15 HMC Version 2.15.0 (Driver 41)
and earlier versions are listed in Table 10-1.
The following HMC feature codes are available for a new order:
FC 0062: M/T 2461-TW3
This feature is the new tower HMC that supports z15, z14 ZR1, z14, z13, and z13s
systems.
FC 0063: M/T 2461-SE3
This feature is the new rack-mounted HMC that supports z15, z14 ZR1, z14, z13, and
z13s systems.
The following older HMCs can be carried forward (the carry forward HMCs do not provide all
enhancements that are available with FC 0062 and FC 0063):
Tower FC 0082
Tower FC 0095
1U Rack FC 0083
1U Rack FC 0096
In z14, the status bar is in Home tab, and the HMC tasks are started in tabs. While working in
a task tab, the console status is not visible. As such, the user cannot be notified of hardware
or operating system messages, and other unacceptable status messages until the current
task is closed and user returns to the Home tab.
In z15, the status bar was moved to the masthead of the HMC interface, which is always
visible (even when working in a task tab), as shown in Figure 10-8.
The task to authorize IBM Product Engineering access to the console is shown in
Figure 10-9. When access is authorized, an IBM product engineer can use an exclusive user
ID and reserved password to log on to the console for problem determination actions.
As shown in Figure 10-9, the task is available only to users with ACSADMIN authority.
Consider the following points:
Customers must ensure that redundant administrator users (ACSADMIN role) are
available for each console.
Customers must document contact information and procedures.
The “Welcome Text” task can be used to identify contact information so that IBM Service
personnel are informed about how to engage customer administrators if HMC or SE
access is needed.
The Product Engineering access options are disabled by default.
The for CPC management, the SEs on z15 are connected to the Ethernet switches that are
installed at the top of the z15 rack, under the SEs. In previous IBM Z systems, the customer
network was connected directly to the bulk power hub (BPH). Now, the SEs are directly
connected to the customer network by using distinct (separate) Ethernet ports than the SE
Ethernet ports that are used for internal CPC management.
Note: The HMC must be connected to the SEs by using a switch. Direct connection
between the HMC and the SEs is not supported.
With FC 0100, HMCs and SEs are packaged redundantly inside the Z CPC frame, which
eliminates the need for managing separate HMC hardware appliances outside of the z15
CPC.
SE redundant physical appliances use increased capacity devices (1U Rack Mounted). The
SE physical appliances run virtual instances of HMC and SE on each physical appliance. This
configuration provides redundancy for the HMC and SE.
With FC 0100, the SE hardware (2461-SE3) provides the required processing resources and
networking connectivity. With Hardware Management Appliance, client access to the HMC is
performed by using a browser (remote access because no HMC KMM console is required).
Hardware Management Appliance connectivity is shown in Figure 10-11.
Various methods are available for setting up the network. Designing and planning the HMC
and SE connectivity is the clients’ responsibility, based on the environment’s connectivity and
security requirements.
For more information about the HMC settings that are related to access and security, see
the HMC and SE console help function or IBM Knowledge Center.
FTPS is based on Secure Sockets Layer cryptographic protocol (SSL) and requires
certificates to authenticate the servers. SFTP is based on Secure Shell protocol (SSH) and
requires SSH keys to authenticate the servers. Certificates and key pairs are hosted on the
z15 HMC Console.
The recommended network topology for HMC, SE, and FTP server is shown in Figure 10-13.
Figure 10-13 Recommended Network Topology for HMC, SE, and FTP server
1 CIM support was removed from the HMC with Version 2.14.0.
With z15 HMC, all FTP connections that originate from the SEs are taken to HMC consoles.
Secure FTP server credentials must be imported to one or more managing HMC consoles.
After the HMC console completes all FTP operations, the HMC console performs the FTP
operation on SE’s behalf and returns the results. The IBM Z platform must be managed by at
least one HMC to allow FTP operations.
Similar to z14, the z15 HMC consoles abandon anonymous cipher suite and implement an
industry standard-based, password-driven cryptography system. The Domain Security
Settings are used to provide authentication and high-quality encryption. Because of these
changes, we now recommend that clients use unique Domain Security settings to provide
maximum security. The new system provides greater security than anonymous cipher suites,
even if the default settings are used.
To allow greater flexibility in password selection, the password limit was increased to 64
characters and special characters are allowed for z15 and z14 installations. If communication
with older systems before z14 is needed, the previous password limits must be followed (6 - 8
characters, only uppercase and number characters allowed).
For more information about HMC networks, see the following resources:
The HMC and SE (Version 2.15.0) console help system, or IBM Knowledge Center
IBM Z 8561 Installation Manual for Physical Planning, GC28-7002
DVD/CD Drive
Starting with z15, a DVD/CD drive is not available with the HMC (nor with the SE).
Ethernet switches
Ethernet switches for HMC and SE connectivity are provided by the client. Existing supported
switches can still be used.
Note: The recommendation is to use a switch with 1000 Mbps/Full duplex support.
RSF is broadband-only
RSF through a modem is not supported the z15 HMC. Broadband is needed for hardware
problem reporting and service. For more information, see 10.4, “Remote Support Facility” on
page 429.
The HMC uses IPv4 and IPv6 multicasting2 to automatically discover the SEs. The HMC
Network Diagnostic Information task can be used to identify the IP addresses (IPv4 and IPv6)
which are used by the HMC to communicate to the SEs (of a CPC).
Because many IPv6 addresses are not fully qualified, shorthand notation can be used. In
shorthand notation, the leading zeros can be omitted, and a series of consecutive zeros can
be replaced with a double colon. The address in the previous example also can be written in
the following manner:
2001:db8::202:b3ff:fe1e:8329
If an IPv6 address is assigned to the HMC for remote operations that use a web browser,
browse to it by specifying that address. The address must be surrounded with square
brackets in the browser’s address field, as shown in the following example:
https://[fdab:1b89:fc07:1:201:6cff:fe72:ba7c]
Multi-factor authentication first factor is login and password; the second factor is TOTP
(Time-based One-Time Password) that is sent to your smartphone, desktop, or app (for
example, Google Authenticator). This TOTP is defined in RFC 6238 standard and uses a
cryptographic hash function that combines a secret key with the current time to generate a
one-time password.
The secret key is generated by HMC/SE/TKE while the user is performing first factor logon.
The secret key is known only to HMC/SE/TKE and to the user’s smartphone. For that reason,
it must be protected as much as your first factor password.
Multi-factor authentication code (MFA code) that was generated as a second factor is
time-sensitive. Therefore, it is important to remember that it should be used soon after it is
generated.
The algorithm within the HMC that is responsible for MFA code generation changes the code
every 30 seconds. However, to make things easier, the HMC and SE console accepts current,
previous, and next MFA codes. It is also important to have HMC, SE, and smartphone clocks
synchronized. If the clocks are not synchronized, the MFA logon attempt fails. Time zone
differences are irrelevant because the MFA code algorithm uses UTC.
On z15, HMC Version 2.15.0 provides integration of HMC authentication and z/OS MFA
support, which means RSA SecurID authentication is achieved by way of centralized support
from IBM MFA for z/OS, with the MFA policy defined in RACF and the HMC IDs assigned to
RACF user IDs. The RSA SecurID passcode (from an RSA SecurID Token) is verified by the
RSA authentication server. This authentication is supported on HMC only, not on the SE.
User Management task is changed on User Definition and User Template Definition to define
and select the MFA Server, and for mapping the HMC user ID to the RACF user ID and
selecting the RACF policy.
Consideration: RSF through a modem is not supported on the z15 HMC. Broadband
connectivity is needed for hardware problem reporting and service.
More information: For more information about the benefits of Broadband RSF and the
SSL/TLS-secured protocol, and a sample configuration for the Broadband RSF
connection, see Integrating the HMC Broadband Remote Support Facility into Your
Enterprise, SC28-6986.
10.4.2 RSF connections to IBM and Enhanced IBM Service Support System
If the HMC and SE are at Driver 22 or later, the driver uses a new remote infrastructure at IBM
when the HMC connects through RSF for certain tasks. Check your network infrastructure
settings to ensure that this new infrastructure works.
At the time of this writing, RSF still uses the “traditional” RETAIN connection. You must add
access to the new Enhanced IBM Service Support System to your current RSF infrastructure
(proxy, firewall, and so on).
To have the best availability and redundancy and to be prepared for the future, the HMC must
access IBM by using the internet through RSF in the following manner: Transmission to the
enhanced IBM Support System requires a domain name server (DNS). The DNS must be
configured on the HMC if a proxy for RSF is not used. If a proxy for RSF is used, the proxy
must provide the DNS.
The following host names and IP addresses are used and your network infrastructure must
allow the HMC to access the following host names:
www-945.ibm.com on port 443
esupport.ibm.com on port 443
Note: Remote browser access is the default for the Hardware Management Appliance.
Access by using the UFD requires physical access to the HMC. Logon security for a web
browser is provided by the local HMC user logon procedures. Certificates for secure
communications are provided, and can be changed by the user. A remote browser session
to the primary HMC that is managing an ensemble allows a user to perform
ensemble-related actions, such as limiting remote web browser access.
Microsoft Internet Explorer, Mozilla Firefox, and Goggle Chrome were tested as remote
browsers. For more information about web browser requirements, see the HMC and SE
console help system or IBM Knowledge Center.
Note: With HMC 2.15.0, certain tasks that required in the past to access to the SE in SOO
mode were implemented as HMC tasks. With this enhancement, the HMC runs the tasks
on the SE directly, without the need to lock the SE in SOO mode.
A full set of granular security controls are provided from the HMC console, to the user,
monitor only, and mobile app password, including multi-factor authentication. This mobile
interface is optional and is disabled by default.
With the introduction of the DPM mode for Linux on Z, only CPCs the user interface and user
interaction with the HMC changed dramatically; the capabilities underneath are still the same.
The figures and command examples that are shown in this section were taken in PR/SM
mode.
The HMC is used to start the power-on reset (POR) of the system. During the POR,
processor units (PUs) are characterized and placed into their respective pools, memory is put
into a single storage pool, and the IOCDS is loaded and started into the hardware system
area (HSA).
The hardware messages task displays hardware-related messages at the CPC, LPAR, or SE
level. It also displays hardware messages that relate to the HMC.
You can use the Load task on the HMC to perform an IPL of an operating system. This task
causes a program to be read from a designated device, and starts that program. You can
perform the IPL of the operating system from storage, the USB flash memory drive (UFD), or
an FTP server.
When an LPAR is active and an operating system is running in it, you can use the HMC to
dynamically change certain LPAR parameters. The HMC provides an interface to change
partition weights, add logical processors to partitions, and add memory.
LPAR weights can also be changed through a scheduled operation. Use the Customize
Scheduled Operations task to define the weights that are set to LPARs at the scheduled time.
Channel paths can be dynamically configured on and off (as needed for each partition) from
an HMC.
The Change LPAR Controls task for z15 can export the Change LPAR Controls table data to a
comma-separated value (.csv)-formatted file. This support is available to a user when they
are connected to the HMC remotely by a web browser.
One example of managing the LPAR settings is the absolute physical hardware LPAR
capacity setting. Driver 15 (zEC12/zBC12) introduced the capability to define (in the image
profile for shared processors) the absolute processor capacity that the image is allowed to
use (independent of the image weight or other cappings).
To indicate that the LPAR can use the non-dedicated processors absolute capping, select
Absolute capping on the Image Profile Processor settings to specify an absolute number of
processors at which to cap the LPAR’s activity. The absolute capping value can be “None” or
a value for the number of processors (0.01 - 255.0).
The LPAR group absolute capping was the next step in partition capping options that are
available on z15, z14 M0x, z14 ZR1, z13s, and z13 CPCs at Driver level 27 and greater.
Following on to LPAR absolute capping, LPAR group absolute capping uses a similar
methodology to enforce the following components:
Customer licensing
Non-z/OS partitions where group soft capping is not an option
z/OS partitions where ISV does not support software capping
A group name, processor capping value, and partition membership are specified at the
hardware console, along with the following properties:
Set an absolute capacity cap by CPU type on a group of LPARs.
Allow each of the partitions to use capacity up to their individual limits if the group's
aggregate consumption does not exceed the group absolute capacity limit.
Include updated SysEvent QVS support (used by vendors who implement software
pricing).
Only shared partitions are managed in these groups.
Specify caps for one or more processor types in the group.
Specify in absolute processor capacity (for example, 2.5 processors).
The value is not tied to the Licensed Internal Code (LIC) configuration code (LICCC). Any
value 0.01 - 255.00 can be specified. This configuration makes the profiles more portable,
which means that you do not have issues in the future when profiles are migrated to new
machines.
Although the absolute cap can be specified to hundredths of a processor, the exact amount
might not be that precise. The same factors that influence the “machine capacity” also
influence the precision with which the absolute capping works.
LPAR absolute capping can be changed through scheduled operations start with HMC
version 2.15.0.
The HMC also provides integrated 3270 and ASCII consoles. These consoles allow an
operating system to be accessed without requiring other network or network devices, such as
TCP/IP or control units.
For example, if the certificate that is associated with the 3270 server on the IBM host is
signed and issued by a corporate certificate, it must be imported, as shown in Figure 10-18.
The import from the remote server option can be used if the connection between the console
and the IBM host can be trusted when the certificate is imported, as shown in Figure 10-19.
Otherwise, import the certificate by using removable media.
A secure Telnet connection is established by adding the prefix L: to the IP address:port of the
IBM host, as shown in Figure 10-20.
When you perform a driver upgrade, always check the Driver (41) Customer Exception Letter
option in the Fixes section at the IBM Resource Link.
For more information, see 9.9, “z15 Enhanced Driver Maintenance” on page 404.
Tip: The IBM Resource Link provides access to the system information for your IBM Z
system according to the system availability data that is sent on a scheduled basis. It
provides more information about the MCL status of your z15 systems.
For more information about accessing the Resource Link, see the IBM Resource Link
website (login required).
At the Resource Link website, click Tools → Machine Information, choose your IBM Z
system, and then, click EC/MCL.
Microcode terms
The microcode features the following characteristics:
The driver contains engineering change (EC) streams.
Each EC stream covers the code for a specific component of z15. It includes a specific
name and an ascending number.
The EC stream name and a specific number are one MCL.
MCLs from the same EC stream must be installed in sequence.
MCLs can include installation dependencies on other MCLs.
Combined MCLs from one or more EC streams are in one bundle.
An MCL contains one or more Microcode Fixes (MCFs).
MCL Application
By design, all MCLs can be applied concurrently, including:
During MCL Bundle application.
During Enhanced Driver Maintenance (Concurrent Driver Upgrade). For example, for z14
at Driver 32, the original GA level, upgrade from Driver 32 to Driver 36 (also applicable to
z14 ZR1).
MCL Activation
By design and with planning, MCLs can be activated concurrently. Consider the following
points:
Most MCLs activate concurrently when applied.
A few MCLs are “Pended” for scheduled activation because activation is disruptive in
some way (Recent History: Most commonly seen for traditional OSA-Express features or
Crypto-Express features).
Activate traditional I/O Feature Pended MCL – LIC on the hardware feature:
– Display Pending MCLs using HMC function or Resource Link Machine Information
Reports
Note: For hardware that does not need CHPID or a FUNCTION definition (for example,
Crypto Express), a different method that is specific to the feature is used.
Alternative: Apply and activate all Pended MCLs disruptively with a scheduled Power On
Reset (POR)
To discover this “Pended” situation, the following actions are done whenever an MCL is
applied:
Logon HMC and select CPC under “System Management”
Change Management
System Information
Query Additional Actions
Or:
Logon HMC and select CPC under “System Management”
Change Management
Query Channel/Crypto Configure Off/On Pending
10.5.5 Monitoring
This section describes monitoring considerations.
You can display more information for the following components (see Figure 10-25 on
page 441):
Power consumption
Environmental
Aggregated processors
Processors (with SMT information)
System Assist Processors
Logical Partitions
Channels
Adapters: Crypto use percentage is displayed according to the physical channel ID
(PCHID number)
The data is presented in table format and graphical “histogram” format. The data also can be
exported to a .csv-formatted file so that the data can be imported into a spreadsheet. For this
task, you must use a web browser to connect to an HMC.
The HMC for IBM z15 features the following CoD capabilities:
SNMP API support:
– API interfaces for granular activation and deactivation
– API interfaces for enhanced CoD query information
– API event notification for any CoD change activity on the system
– CoD API interfaces, such as On/Off CoD and Capacity Back Up (CBU)
SE window features (accessed through HMC Single Object Operations):
– Window controls for granular activation and deactivation
– History window for all CoD actions
– Description editing of CoD records
HMC/SE provides the following CoD information:
– Millions of service units (MSU) and processor tokens
– Last activation time
– Pending resources that are shown by processor type instead of only a total count
– Option to show more information about installed and staged permanent records
– More information for the Attention state by providing seven more flags
For more information about using and setting up CPM, see IBM Knowledge Center or the
following publications:
z/OS MVS Capacity Provisioning User’s Guide, SC33-8299
IBM Z System Capacity on-Demand User’s Guide, SC28-6985
Important: The Sysplex Time task on the SE was discontinued for z15. Therefore, an
HMC at Version 2.15.0 is required to manage system time for z15 CPCs.
With the Server Time Protocol (STP) functions, the role of the HMC is extended to provide the
user interface for managing the Coordinated Timing Network (CTN). Consider the following
points:
IBM Z CPCs rely on STP for time synchronization, and continue to provide support of a
pulse per second (PPS) port. Consider the following points:
– New for z15 - STP can be configured to use Precision Time Protocol (PTP, IEEE 1588)
as external time source. Current implementation requires that the support element is
connected to a PTP capable network infrastucture and has access to a PTP server.
– An STP that uses Network Time Protocol (NTP) or PTP as External Time Source
(ETS) server with PPS maintains an accuracy of 10 ms
– An STP that uses ETS without PPS maintains accuracy of 100 ms
The z15 cannot be in the same CTN with zEC12, zBC12, or earlier systems and cannot
become member of a mixed CTN.
An STP-only CTN can be managed by using different HMCs. However, the HMC must be at
the same driver level (or later) than any SE that is to be managed. Also, all SEs to be
managed must be known (defined) to that HMC.In a STP-only CTN, the HMC can be used to
perform the following tasks:
Start or modify the CTN ID.
Start the time (manually or by contacting an NTP server).
Start the time zone offset, Daylight Saving Time offset, and leap second offset.
Assign the roles of preferred, backup, and current time servers, and arbiter.
Adjust time by up to plus or minus 60 seconds.
Schedule changes to the offsets listed. STP can automatically schedule Daylight Saving
Time, based on the selected time zone.
Monitor the status of the CTN.
Monitor the status of the coupling links that are started for STP message exchanges.
For diagnostic purposes, the PPS port state on a z15 can be displayed and fenced ports
can be reset individually.
Important: The Sysplex Time task on the SE was discontinued for z15. Therefore, an
HMC at Version 2.15.0 is required to manage system time for z15 CPCs.
Detailed instructions and guidelines are provided within task workflow. z15 HMC provides a
visual representation of the CTN topology. A preview of any configuration action is also shown
in topological display. An example of the topology view is shown in Figure 10-26.
Figure 10-26 CTN topology visible on HMC Manage System Time window
Click Current time details for more information and available options (for example, adjusting
time zone and leap second offset), as shown in Figure 10-27.
ECAR support is faster than the original CAR support because the console path changes
from a 2-way path to a 1-way path. Also, almost no lag time is incurred between the system
checkstop and the start of CAR processing. Because the request is generated from the PTS
before system logging, it avoids the potential of recovery being held up.
Requirements
ECAR is available on z15, z14 M0x, z14 ZR1, and z13/z13s systems on Driver 27 and later.
Attention: z15 and z14 ZR1 do not support InfiniBand connectivity; therefore, these
servers cannot be connected by using IFB coupling/timing links to a z14 (3906), z13, or
z13s. As such, in a CTN with servers that still use InfiniBand coupling, CTN roles (PTS,
CTS, or Arbiter) must be assigned carefully in such a way that a failure of a CTN role
playing server does not affect CTN functionality (loss of synchronization or transition to
lower STP Stratum levels).
For more information about planning and setup, see the following publications:
Server Time Protocol Planning Guide, SG24-7280
Server Time Protocol Implementation Guide, SG24-7281
Server Time Protocol Recovery Guide, SG24-7380
CTN Split
The HMC menus for Server Time Protocol (STP) provide support when one or more systems
must be split in to a separate CTN without interruption in the clock source.
The task is available under the Advanced Actions menu in the Manage System Time task.
Several checks are performed to avoid potential disruptive actions. If targeted CTN includes
only members with the roles, task start fails with error message. If targeted CTN includes at
least one system without any roles, the task starts. An informational warning is presented to
the user to acknowledge that sysplex workloads are divided appropriately.
Note: After joining the selected CTN, all systems within the current CTN are synchronized
with the Current Time Server of the selected CTN. A coupling link must be in place that
connects the CTS of the selected CTN and the CTS of the current CTN.
During the transition state, most of the STP actions for the two affected CTNs are disabled.
After the merge is completed, STP actions are enabled again.
For more information about planning and understanding STP server roles, see the following
publications:
Server Time Protocol Planning Guide, SG24-7280
Server Time Protocol Implementation Guide, SG24-7281
Server Time Protocol Recovery Guide, SG24-7380
The NTP server becomes the single time source (the ETS) for STP and other servers that are
not IBM Z systems (such as AIX®, and Microsoft Windows) that include NTP clients.
The HMC can act as an NTP server. With this support, the z15 can receive the time from the
HMC without accessing a LAN other than the HMC and SE network. When the HMC is used
as an NTP server, it can be configured to receive the NTP source from the internet. For this
type of configuration, a LAN that is separate from the HMC/SE LAN can be used.
The HMC offers the following symmetric key and autokey authentication and NTP commands:
Symmetric key (NTP V3-V4) authentication
Symmetric key authentication is described in RFC 1305, which was made available in
NTP Version 3. Symmetric key encryption uses the same key for encryption and
decryption. Users that are exchanging data keep this key to themselves. Messages
encrypted with a secret key can be decrypted only with the same secret key. Symmetric
key authentication supports network address translation (NAT).
Symmetric key autokey (NTP V4) authentication
This autokey uses public key cryptography, as described in RFC 5906, which was made
available in NTP Version 4. You can generate keys for the HMC NTP by clicking Generate
Local Host Key in the Autokey Configuration window. This option issues the ntp-keygen
command to generate the specific key and certificate for this system. Autokey
authentication is not available with the NAT firewall.
Issue NTP commands
NTP command support is added to display the status of remote NTP servers and the
current NTP server (HMC).
For more information about planning and setup for STP and NTP, see the following
publications:
Server Time Protocol Planning Guide, SG24-7280
Server Time Protocol Implementation Guide, SG24-7281
Server Time Protocol Recovery Guide, SG24-7380
With z15, you can offload the following HMC and SE log files for customer audit:
Console event log
Console service history
Tasks performed log
Security logs
System log
Full log offload and delta log offload (since the last offload request) are provided. Offloading
to removable media and to remote locations by FTP is available. The offloading can be
manually started by the new Audit and Log Management task or scheduled by the Customize
Scheduled Operations task. The data can be offloaded in the HTML and XML formats.
Each HMC user ID template defines the specific authorization levels for the tasks and objects
for the user who is mapped to that template. The HMC user is mapped to a specific user ID
template by user ID pattern matching. The system then obtains the name of the user ID
template from content in the LDAP server schema data.
If you want to change the roles for a default user ID, create your own version by copying a
default user ID.
The Secure FTP infrastructure allows HMC and SE applications to query whether a public key
is associated with a host address and to use the Secure FTP interface with the appropriate
public key for a host. Tasks that use FTP now provide a selection for the secure host
connection.
When selected, the task verifies that a public key is associated with the specified host name.
If a public key is not provided, a message window opens that points to the Manage SSH Keys
task to enter a public key. The following tasks provide this support:
Import/Export IOCDS
Advanced Facilities FTP IBM Content Collector Load
Audit and Log Management (Scheduled Operations only)
FCP Configuration Import/Export
OSA view Port Parameter Export
OSA-Integrated Console Configuration Import/Export
The information that is needed to manage a system’s I/O configuration must be obtained from
many separate sources. The System Input/Output Configuration Analyzer task enables the
system hardware administrator to access, from one location, the information from those
sources. Managing I/O configurations then becomes easier, particularly across multiple
systems.
The System Input/Output Configuration Analyzer task runs the following functions:
Analyzes the current active IOCDS on the SE.
Extracts information about the defined channel, partitions, link addresses, and control
units.
Requests the channels’ node ID information. The Fibre Channel connection (FICON)
channels support remote node ID information, which is also collected.
The System Input/Output Configuration Analyzer is a view-only tool. It does not offer any
options other than viewing. By using the tool, data is formatted and displayed in five different
views. The tool provides various sort options, and data can be exported to a UFD for later
viewing.
The older system, such as z13’s HMC, supports the CIM as an extra systems management
API. Starting with z14, the CIM support is removed.
For more information about APIs, see IBM Z Application Programming Interfaces,
SB10-7164.
Cryptographic hardware
z15 systems include standard cryptographic hardware and optional cryptographic features for
flexibility and growth capability.
The Crypto Express7S, which is a new Peripheral Component Interconnect Express (PCIe)
cryptographic coprocessor, is an optional z15 exclusive feature. Crypto Express7S provides a
secure programming and hardware environment on which crypto processes are run. Each
Crypto Express7S adapter can be configured by the installation as a Secure IBM CCA
coprocessor, a Secure IBM Enterprise Public Key Cryptography Standards (PKCS) #11
(EP11) coprocessor, or an accelerator.
When EP11 mode is selected, a unique Enterprise PKCS #11 firmware is loaded into the
cryptographic coprocessor. It is separate from the Common Cryptographic Architecture
(CCA) firmware that is loaded when a CCA coprocessor is selected. CCA firmware and
PKCS #11 firmware cannot coexist in a card.
The Trusted Key Entry (TKE) Workstation with smart card reader feature is required to
support the administration of the Crypto Express7S when configured as an Enterprise
PKCS #11 coprocessor.
To support the new Crypto Express7S card, the TKE9.2 is needed. An example of the
Cryptographic Configuration window is shown in Figure 10-28 on page 450.
The Usage Domain Zeroize task is provided to clear the appropriate partition crypto keys for a
usage domain when you remove a crypto card from a partition. Crypto Express7/6/5S in
EP11 mode is configured to the standby state after the zeroize process.
For more information, see IBM z15 (8561) Configuration Setup, SG24-8860.
This operation ensures that any changes that are made to the data are detected during the
upgrade process by verifying the digital signature. It helps ensure that no malware can be
installed on IBM Z products during firmware updates. It also enables the z15 Central
Processor Assist for Cryptographic Function (CPACF) functions to comply with Federal
Information Processing Standard (FIPS) 140-2 Level 1 for Cryptographic LIC changes. The
enhancement follows the IBM Z focus of security for the HMC and the SE.
The Crypto Express7S (CEX7S) is compliant with CCA PCI HSM. TKE workstation is optional
when used to manage a Crypto Express7S feature that is defined as a CCA coprocessor in
normal mode. However, it is mandatory when it is used to manage a Crypto Express7S
feature that is defined as a CCA coprocessor in PCI-HSM mode or is defined as an EP11
coprocessor (CCA in PCI-HSM mode and EP11 also require a smart card reader plus smart
cards with FIPS certification).
Setting up is a disruptive action. The selection of the DPM mode of operation is done by using
the Enable Dynamic Partition Manager function, which is available in the SE CPC
Configuration menu. Enabling DPM is performed on the SE and requires a system POR.
Attention: The Enabling Dynamic Partition Manager task is run on the SE and is
performed by your IBM system service representative (SSR).
After the CPC is restarted and you log on to the HMC in which this CPC is defined, the HMC
shows the Welcome window that is shown in Figure 10-29.
New partitions can be added by selecting Get Started. For more information, see IBM
Knowledge Center.
Naming: Throughout this chapter, z15 refers to IBM z15 Model T01 (Machine Type 8561),
unless otherwise specified.
The following options are available for physically installing the server:
Air or water cooling
Power Distribution Unit (PDU) or Bulk Power Assembly (BPA) for power
Installation on a raised floor or non-raised floor
I/O and power cables can exit under the raised floor or off the top of the server frames
For more information about physical planning, see IBM Z 8561 Installation Manual for
Physical Planning, GC28-7002.
The z15 servers are available in the following power and cooling options:
Intelligent Power Distribution Unit-based power (iPDU) - or PDU
Radiator-based cooling (air cooled system)
Bulk Power Assembly-based power (BPA):
– Radiator-based cooling (air cooled system)
– Water Cooling Unit that uses customer-provided chilled water supply
A PDU-based system can have 2 - 8 power cords, depending on the configuration. The use of
iPDU on z15 might enable fewer frames, which allows for more I/O slots to be available and
improves power efficiency to lower overall energy costs. It also offers some standardization
and ease of data center installation planning, which allows the z15 to easily coexist with other
platforms within the data center.
Power requirements
The z15 is designed with a fully redundant power system. To make full use of the redundancy
that is built into the server, the PDUs within one pair must be powered from different power
distribution panels. In that case, if one PDU in a pair fails, the second PDU ensures continued
operation of the server without interruption.
The second, third, and fourth PDU pairs are installed dependent on other CPC or PCIe+ I/O
drawers installed. The locations of the PDU pairs and frames are listed in Table 11-1.
A rear view of a maximum configured PDU-powered system with five CPC drawers and 12
PCIe+ I/O drawers is shown in Figure 11-1.
:
The number of PDUs and power cords that are required based on the number of CPC
drawers and PCIe+ I/O drawers are in Table 11-3.
1 2 2 2 2 6 6 6 6
2 4 4 4 4 4 4 4 4 6 6 6 6 6
3 4 4 4 4 4 4 4 6 6 6 6 6 N/A
4 6 6 6 6 6 6 6 6 6 6 8 8 8
5 6 6 6 6 6 6 6 6 6 8 8 8 8
Power consumption
The utility power consumption for the z15 for PDU option is listed in Table 11-4.
0 1 2 3 4 5 6 7 8 9 10 11 12
max34 3.8 4.8 5.9 6.9 8.0 9.0 10.0 11.1 12.1 13.2 14.2 15.3 16.1
FC0655 kw kw kw kw kw kw kw kw kw kw kw kw kw
max71 6.9 7.9 9.0 10.0 11.1 12.1 13.2 14.2 15.3 16.3 17.4 18.4 19.2
FC0656 kw kw kw kw kw kw kw kw kw kw kw kw kw
max108 10.1 11.1 12.2 13.2 14.3 15.3 16.4 17.4 18.5 19.5 20.6 21.6 N/A
FC0657 kw kw kw kw kw kw kw kw kw kw kw kw
max145 13.5 14.5 15.6 16.6 17.7 18.7 19.8 20.8 21.9 22.9 24.0 25.0 25.8
FC0658 kw kw kw kw kw kw kw kw kw kw kw kw kw
max190 16.6 17.6 18.7 19.7 20.8 21.8 22.9 23.9 25.0 26.0 27.1 28.1 28.9
FC0659 kw kw kw kw kw kw kw kw kw kw kw kw kw
Power estimation for any configuration, power source, and room condition can be obtained
by using the power estimation tool that is available at the IBM Resource Link website (login
required).
On the Resource Link page, click Tools → Power and weight estimation.
The BPA option is required for clients who order an Internal Battery Feature (IBF), Water
Cooling Unit (WCU), or Balanced Power. The BPA requires two or four power cords. All BPAs
are concurrently repairable in the field.
Power requirements
The 8561 operates from 2 or 4 fully redundant power supplies. These redundant power
supplies each have their own power cords, or pair of power cords, which allows the system to
survive the loss of customer power to either power cord or power cord pair.
If power is interrupted to one of the power supplies, the other power supply assumes the
entire load and the system continues to operate without interruption. Therefore, the power
cords for each power supply must be wired to support the entire power load of the system.
For the most reliable availability, the power cords in the rear of the frame should be powered
from different PDUs. All power cords exit through the rear of the frame. The utility current
distribution across the phase conductors (phase current balance) depends on the system
configuration. Each front end power supply is provided with phase switching redundancy.
The loss of an input phase is detected and the total input current is switched to the remaining
phase pair without any power interruption. Depending on the configuration input power draw,
the system can run from several minutes to indefinitely in this condition. Because most single
phase losses are transients that recover in seconds, this redundancy provides protection
against virtually all single phase outages.
Power cords for the BPAs are attached to the options that are listed in Table 11-5.
The source power cords ratings for BPA systems are listed in Table 11-6.
A rear view of a maximum configured BPA powered system with two BPA pairs, IBF, five CPC
drawers, and 11 PCIe+ I/O drawers is shown in Figure 11-2.
The number of power cords that are required based on the number of CPC drawers and
PCIe+ I/O drawers are listed in Table 11-7.
1 1 1 2 2 2 2 2 3 3 3 4 4
2 1 2 2 2 2 2 3 3 3 4 4 4
3 1 2 2 2 2 2 3 3 3 4 4 4
4 2 2 3 3 3 3 3 4 4 4 4 4
5 2 2 3 3 3 3 3 4 4 4 4 4
The number of power cords and number of Bulk Power Regulators required based on the
number of CPC processor drawers and number of PCIe+ I/O drawers in the configuration
(radiator or WCU) is shown in Table 11-8 on page 459.
0 1 2 3 4 5 6
1 CPC 1 1 2 2 2 2 2
2 CPC 2 2 2 2 2 3 3
3 CPC 2 2 3 3 3 3 -
Note: Balanced power feature includes all BPRs that are plugged in all BPAs in the system.
Power consumption
The utility power consumption for the z15 for the BPA radiator cooled system option is listed in
Table 11-9.
max34 4.2 5.3 6.4 7.5 8.6 9.7 10.9 12.0 13.1 14.2 15.3 16.4
FC0655 kw kw kw kw kw kw kw kw kw kw kw kw
(1)
max71 7.5 8.6 9.7 10.8 11.9 13.0 14.2 15.3 16.4 17.5 18.6 18.6
FC0656 kw kw kw kw kw kw kw kw kw kw kw kw
(2)
max108 10.9 12.0 13.1 14.2 15.3 16.4 17.6 18.7 19.8 20.9 22.0 22.0
FC0657 kw kw kw kw kw kw kw kw kw kw kw kw
(3)
max145 14.8 15.9 17.0 18.1 19.2 20.3 21.4 22.5 23.6 24.8 25.9 25.9
FC0658 kw kw kw kw kw kw kw kw kw kw kw kw
(4)
max190 18.1 19.2 20.3 21.4 22.5 23.6 24.7 25.8 26.9 28.0 29.2 29.2
FC0659 kw kw kw kw kw kw kw kw kw kw kw kw
(5)
max34 4.0 5.1 6.2 7.3 8.4 9.5 10.6 11.8k 12.9 14.0 15.1 16.2
FC0655 kw kw kw kw kw kw kw w kw kw kw kw
(1)
max71 7.2 8.3 9.4 10.5 11.6 12.7 13.8 14.9 16.1 17.2 18.3 19.4
FC0656 kw kw kw kw kw kw kw kw kw kw kw kw
(2)
max108 10.4 11.5 12.6 13.7 14.8 15.9 17.0 18.1 19.2 20.4 21.5 22.6
FC0657 kw kw kw kw kw kw kw kw kw kw kw kw
(3)
max145 14.0 15.1 16.2 17.3 18.5 19.6 20.7 21.8 22.9 24.0 25.1 26.2
FC0658 kw kw kw kw kw kw kw kw kw kw kw kw
(4)
max190 17.2 18.3 19.4 20.5 21.6 22.8 23.9 25.0 26.1 27.2 28.3 29.4
FC0659 kw kw kw kw kw kw kw kw kw kw kw kw
(5)
If the z15 server is configured with the Internal Battery Feature (IBF), Balanced Power Plan
Ahead automatically supplies the maximum number of batteries (IBFs) with the system.
Power estimation for any configuration, power source, and room condition can be obtained
by using the power estimation tool at IBM Resource Link website (authentication required).
On the Resource Link page, click Tools → Power and weight estimation.
The z15 servers include a recommended (long-term) ambient temperature range of 18°C
(64.4°F) - 27°C (80.6°F). The minimum allowed ambient temperature is 5°C (41°F) and the
maximum allowed temperature is 40°C (104°F).
For more information about cooling requirements, see IBM Z 8561 Installation Manual for
Physical Planning, GC28-7002.
The radiator cooling system requires chilled air to fulfill the air-cooling requirements. z15
system airflow is from the front (intake, chilled air) to the rear (exhausts, warm air) of the
frames. The chilled air is provided through perforated floor panels in front of the system
The hot and cold airflow and the arrangement of server aisles are shown in Figure 11-3.
As shown in Figure 11-3, rows of servers must be placed front-to-front. Chilled air is provided
through perforated floor panels that are placed in rows between the fronts of servers (the cold
aisles). Perforated tiles generally are not placed in the hot aisles. If your computer room
causes the temperature in the hot aisles to exceed a comfortable temperature, add as many
perforated tiles as necessary to create a satisfactory comfort level. Heated exhaust air exits
the computer room above the computing equipment.
The water cooled 8561 system requires four or eight connections to the facility water,
depending on the system configuration. Consider the following points:
Systems with 1 - 3 CPC drawers require two feeds and two returns in the rear of the A
frame.
Systems with more than three CPC drawers require two feeds and two returns in the rear
for both the A and the B frames.
These connections are made by using hoses that are fixed to the facility plumbing. They are
routed up through the rear tailgate of the machine and terminated by using quick connect
couplings.
Hoses and a means to fasten them to the facility plumbing are provided with the server.
The water cooled 8561 systems with one to three CPC drawers feature one water cooling
assembly (WCA). The water cooled 8561 systems with four or five CPC drawers include two
WCAs. Each WCA contains two fully redundant water control units (WCUs). These water
control units each have their own facility feed and return water connections. If water is
interrupted to one of the WCUs, the other water control unit assumes the entire load and the
server continues to operate without interruption.
Therefore, each water connection to the facility plumbing must support the entire flow
requirement for the WCA. If water is lost to both WCUs, the system attempts to reject heat by
using the inner door heat exchangers in each frame and increasing system blower speeds.
The server can also run in a degraded mode during this event.
Raised floor: The minimum raised floor height for a water-cooled system is 22.86 cm
(8.6 in.).
These connections are made by using hoses that are fixed to the facility plumbing and are
routed up through the rear tailgate of the system. They end with quick connect couplings.
Before you install z15 servers with water-cooled option, your facility must meet following
requirements:
Total water hardness cannot exceed 200 mg/L of calcium carbonate.
The facility water pH is 7 - 9.
Turbidity is less than 10 Nephelometric Turbidity Units (NTUs).
Bacteria is less than 1000 colony-forming units (CFUs)/ml.
The water is as free of particulate matter as feasible.
The allowable system inlet water temperature range is 6°C - 20°C (43°F - 68°F) by using
standard building chilled water. A special water system is typically not required.
The client’s ends of the hoses are left open, which allows you to cut the hose to a custom
length. An insulation clamp is provided to secure the insulation and protective sleeving after
you cut the hose to the correct length and install it onto your plumbing.
Important: Use shut-off valves in front of the hoses. This configuration allows for the
removal of the hoses for a service procedure or relocation. Valves are not included in the
order. A stainless steel fitting is available for ordering. The fitting is barbed on one side and
has a 2.54 cm (1 in.) male national pipe thread (NPT).
For more information about the tools that are needed for the water supply connections, see
IBM Z 8561 Installation Manual for Physical Planning, GC28-7002.
Removal of IBF supporta: IBM z15 is planned to be the last IBM Z server to offer an
Internal Battery Feature (IBF). As client data centers continue to improve power stability
and uninterruptible power supply (UPS) coordination, IBM Z continues to innovate to help
clients take advantage of common power efficiency and monitoring across their
ecosystems. Additional support for data center power planning can be requested through
your IBM Sales contact.
a. Statements by IBM regarding its plans, directions, and intent are subject to change or
withdrawal without notice at the sole discretion of IBM.
If a power shutdown occurs, the optional Internal Battery Feature (IBF) provides sustained
system operations for a relatively short time, which allows a proper z15 server’s shutdown. In
addition, an external uninterrupted power supply system can be connected, which allows for
longer periods of sustained operation.
Attention: The optional Internal Battery Feature (FC 3217) contains lithium ion batteries,
which are packaged separately from the frame. A pair of batteries weighs approximately
43.5 kg (96 lb), as each individual battery weighs approximately 21.8 kg (48 lb).
Because of the hazardous nature of these Lithium ion batteries, only a certified Dangerous
Goods Transportation employee is authorized to package and ship the IBFs. In the event of
a system relocation, this restriction must be considered and is the responsibility of the
client.
For more information, see IBM Z 8561 Installation Manual for Physical Planning,
GC28-7002.
The IBF (when installed) can provide emergency power for the estimated time that is listed in
Table 11-11. The number of IBFs depends on the number of BPRs. For the number of BPRs
that are installed in relation to I/O units and the number of CPC drawers, see Table 11-8 on
page 459. They are installed in pairs. You can have two, four, or six batteries (odd numbers
are not allowed) per BPA.
Table 11-11 Battery hold-up times (minutes) for IBM z15 Model T01
CPC number of PCIe+ I/O drawers
Feature
0 1 2 3 4 5 6 7 8 9 10 11
Max34 10 7 12 10 9 8 7 7 7 7 7 7
FC 0655
Max71 11 9 8 7 6 9 9 9 9 9 9 9
FC 0656
Max108 7 6 9 8 8 7 7 7 7 7 7 7
FC 0657
Max145 7 7 6 8 8 7 7 7 7 7 7 7
FC 0658
Max190 7 7 6 9 8 8 7 7 7 7 7 7
FC 0659
Consideration: The system holdup times that are listed in Table 11-11 on page 464
assume that both sides are functional and have fresh batteries under normal room ambient
conditions.
Holdup times are greater for configurations that do not have every I/O slot plugged, the
maximum installed memory, and are not using the maximum processors.
These holdup times are estimates. Your particular battery holdup time for any specific
circumstance might be different.
Holdup times vary depending on the number of BPRs that are installed. As the number of
BPRs increases, the holdup time also increases until the maximum number of BPRs is
reached. After six BPRs (three per side) are installed, no other batteries are added;
therefore, the time decreases from that point.
Holdup times for actual configurations are provided in the power estimation tool at IBM
Resource Link website.
On the Resource Link page, click Tools → Machine information. Then, select your IBM
Z system and click Power Estimation Tool.
The BPR plugging order in the Bulk Power Enclosure is shown in Figure 11-5. Only the A
frame is shown, but the same order is used in the B or C racks if present. Consider the
following points:
BPRs (and IBF units when the feature is installed) always are plugged in pairs.
The number of IBFs (when installed) always equals the number of BPRs.
The z15 can be installed on a raised or non-raised floor. For more information about weight
distribution and floor loading tables, see IBM 8561 Installation Manual for Physical Planning,
GC28-7002. This data is used with the maximum frame weight, frame width, and frame depth
to calculate the floor loading.
Weight estimates for the maximum system configurations on the 8561 are listed in
Table 11-12.
A - 795 kg - - 795 kg
(1753 lbs) (1753 lbs)
The power and weight estimation tool for Z servers on Resource Link covers the estimated
weight for your designated configuration. It is available on IBM Resource Link website.
On the Resource Link page, click Tools → Power and weight estimation.
Note: On the z15, all I/O cabling and power cords come from the rear of the machine;
therefore, all related features for Bottom and Top Exit cabling are in the rear of the frame.
Raised floor
If the z15 server is installed in a raised floor environment, air-cooled and water-cooled models
are supported. You can select top exit features to manage I/O cables from the top frame of the
z15 server.
The following top exit options are available for z15 servers:
Top Exit I/O cabling feature code (FC 7917)
Bottom Exit cabling feature code (FC 7919)
Top Exit Cabling without Tophat feature code (FC 7928)
Note: The feature provides the same extra hardware for every frame in the configuration.
The Top Exit cabling feature adds 117.5 mm (4.63 in.) to the height of the frame and
approximately 5.4 kg (12 lbs) to the weight.
If the Top Exit cabling feature is not ordered, two sliding plates are available on the top of the
frame (one on each side of the rear of the frame) that can be partially opened. By opening
these plates, I/O cabling and power cords can exit the frame. The plates should be removed
to install the The Top Exit cabling feature as shown in Figure 11-7 on page 468.
All external cabling enters the rear of the rack from under floor or from above the rack.
Different from previous Z Systems, no cabling access or cable plugging is available at the
front of the rack. The top view of the rack with and without FC 7917 is shown in Figure 11-8.
The Top Exit Cabling feature provides new hardware. The new hardware resembles a
rectangular box with an open side that faces the front or rear of the rack. It includes other
hardware to organize and fasten cables.
The Top Exit Cabling option can be used for routing power cables and IO cables out the top of
the machine.
Without the Top Exit Cabling feature, power and cables still can be run out the top of the rack
through two adjustable openings at the top rear of the rack, as shown on the left side of
Figure 11-8.
The Bottom Exit Cabling feature provides tailgate hardware for routing power cables or IO
cables out the bottom of the machine.
For more information, see IBM 8561 Installation Manual for Physical Planning, GC28-7002,
and 11.3, “Physical planning” on page 467
The frame tie-down kit can be used on a non-raised floor where the frame is secured directly
to a concrete floor, or on a raised floor where the frame is secured to the concrete floor
underneath the raised floor. Raised floors 241.3 mm (9.5 inches) - 1270 mm (50 inches) are
supported.
The kits help secure the frames and their contents from damage when they are exposed to
shocks and vibrations, such as in a seismic event. The frame tie-downs are intended for
securing a frame that weighs up to 1308 kg (2885 lbs).
For more information, see IBM 8561 Installation Manual for Physical Planning, GC28-7002.
For more information, see IBM 8561 Installation Manual for Physical Planning, GC28-7002.
The following tools are available to plan and monitor the energy consumption of z15 servers:
Power estimation tool on Resource Link
Energy Optimization Advisor task for maximum potential power on HMC and SE
Monitors Dashboard and Environmental Efficiency Statistics tasks on HMC and SE
Select the advice hyperlinks to provide specific recommendations for your system, as shown
in Figure 11-12.
The data is presented in table format and graphical “histogram” format. The data can also be
exported to a .csv-formatted file so that the data can be imported into a spreadsheet. For this
task, you must use a web browser to connect to an HMC.
Note: Throughout this chapter, z15 refers to IBM z15 Model T01 (Machine Type 8561)
unless otherwise specified.
Uniprocessor performance also increased. On average, a z15 Model 701 offers performance
improvements of more than 12% over the z14 Model 701. Figure 12-1 shows a system
performance comparison of successive IBM Z servers.
Operating system support varies for the number of “engines” that are supported.
The zEDC Express feature was adopted by enterprises because it helps with software costs
for compression/decompression operations (by offloading these operations), and increases
data encryption (compression before encryption) efficiency.
With z15, the zEDC Express functionality was moved off from the PCIe infrastructure into the
processor nest. By moving the compression and decompression into the processor nest
(on-chip), IBM z15 processor provides a new level of performance for these tasks and
eliminates the need for the zEDC Express feature virtualization. It also brings new use cases
to the platform.
For more information, see Appendix C, “IBM Integrated Accelerator for zEnterprise Data
Compression” on page 509.
For z/OS V2R3 studies, the capacity scaling factor that is commonly associated with the
reference processor is set to a 2094-701 with a Processor Capacity Index (PCI) value of 593.
This value is unchanged since z/OS V1R11 LSPR. The use of the same scaling factor across
LSPR releases minimizes the changes in capacity results for an older study and provides
more accurate capacity view for a new study.
Performance data for z15 servers were obtained with z/OS V2R3 (running Db2 for z/OS V12,
CICS TS V5R3, IMS V14, Enterprise COBOL V6R2, and WebSphere Application Server for
z/OS V9.0.0.8). All IBM Z server generations are measured in the same environment with the
same workloads at high usage.
Note: If your software configuration is different from what is described here, the
performance results might vary.
Consult the LSPR when you consider performance on the z15. The range of performance
ratings across the individual LSPR workloads is likely to include a large spread. Performance
of the individual logical partitions (LPARs) varies depending on the fluctuating resource
requirements of other partitions and the availability of processor units (PUs). Therefore, it is
important to know which LSPR workload type suite your production environment. For more
information, see 12.8, “Workload performance variation” on page 485.
For more information about performance, see the Large Systems Performance Reference for
IBM Z page of the Resource Link website.
For more information about millions of service units (MSU) ratings, see the IBM z Systems
Software Contracts page of the IBM IT infrastructure website.
The CPU Measurement Facility (CPU MF) data that was introduced on the z10 provides
insight into the interaction of workload and hardware design in production workloads. CPU
MF data helps LSPR to adjust workload capacity curves that are based on the underlying
hardware sensitivities; in particular, the processor access to caches and memory. This
processor access to caches and memory is called nest. By using this data, LSPR introduces
three workload capacity categories that replace all older primitives and mixes.
LSPR contains the internal throughput rate ratios (ITRRs) for the z15 and the previous
generation processor families. These ratios are based on measurements and projections that
use standard IBM benchmarks in a controlled environment.
The throughput that any user experiences can vary depending on the amount of
multiprogramming in the user’s job stream, the I/O configuration, and the workload
processed. Therefore, no assurance can be given that an individual user can achieve
throughput improvements that are equivalent to the performance ratios that are stated.
The path length varies for each transaction or job, and depends on the complexity of the tasks
that must be run. For a particular transaction or job, the application path length tends to stay
the same, assuming that the transaction or job is asked to run the same task each time.
However, the path length that is associated with the operating system or subsystem can vary
based on the following factors:
Competition with other tasks in the system for shared resources. As the total number of
tasks grows, more instructions are needed to manage the resources.
The number of logical processors (n-way) of the image or LPAR. As the number of logical
processors grows, more instructions are needed to manage resources that are serialized
by latches and locks.
As workloads are moved between microprocessors with various designs, performance varies.
However, when on a processor, this component tends to be similar across all models of that
processor.
Workload performance is sensitive to how deep into the memory hierarchy the processor
must go to retrieve the workload instructions and data for running. The best performance
occurs when the instructions and data are in the caches nearest the processor because little
time is spent waiting before running. If the instructions and data must be retrieved from farther
out in the hierarchy, the processor spends more time waiting for their arrival.
As workloads are moved between processors with various memory hierarchy designs,
performance varies because the average time to retrieve instructions and data from within the
memory hierarchy varies. Also, when on a processor, this component continues to vary
because the location of a workload’s instructions and data within the memory hierarchy is
affected by several factors that include, but are not limited to, the following factors:
Locality of reference
I/O rate
Competition from other applications and LPARs
The term Relative Nest Intensity (RNI) indicates the level of activity to this part of the memory
hierarchy. By using data from CPU MF, the RNI of the workload that is running in an LPAR can
be calculated. The higher the RNI, the deeper into the memory hierarchy the processor must
go to retrieve the instructions and data for that workload.
The “Nest”
L2LP L2RP
L1 MEMP
L3P L4LP L4RP
Many factors influence the performance of a workload. However, these factors often are
influencing the RNI of the workload. The interaction of all these factors results in a net RNI for
the workload, which in turn directly relates to the performance of the workload.
These factors are tendencies, not absolutes. For example, a workload might have a low I/O
rate, intensive processor use, and a high locality of reference, which all suggest a low RNI.
However, it might be competing with many other applications within the same LPAR and many
other LPARs on the processor, which tends to create a higher RNI. It is the net effect of the
interaction of all these factors that determines the RNI.
The traditional factors that were used to categorize workloads in the past are shown with their
RNI tendency in Figure 12-4.
Little can be done to affect most of these factors. An application type is whatever is necessary
to do the job. The data reference pattern and processor usage tend to be inherent to the
nature of the application. The LPAR configuration and application mix are mostly a function of
what must be supported on a system. The I/O rate can be influenced somewhat through
buffer pool tuning.
However, one factor, software configuration tuning, is often overlooked but can have a direct
effect on RNI. This term refers to the number of address spaces (such as CICS
application-owning regions (AORs) or batch initiators) that are needed to support a workload.
Tuning to reduce the number of simultaneously active address spaces to the optimum number
that is needed to support a workload can reduce RNI and improve performance. In the LSPR,
the number of address spaces for each processor type and n-way configuration is tuned to be
consistent with what is needed to support the workload. Therefore, the LSPR workload
capacity ratios reflect a presumed level of software configuration tuning. Retuning the
software configuration of a production workload as it moves to a larger or faster processor
might be needed to achieve the published LSPR ratios.
These categories are based on the RNI. The RNI is influenced by many variables, such as
application type, I/O rate, application mix, processor usage, data reference patterns, LPAR
configuration, and the software configuration that is running. CPU MF data can be collected
by z/OS System Measurement Facility on SMF 113 records or z/VM Monitor starting with
z/VM V5R4.
The IBM Processor Capacity Reference for IBM Z (zPCR) tool supports the following
workload categories:
Low
Low-Average
For more information about the no-charge IBM zPCR tool (which reflects the latest IBM LSPR
measurements), see the Getting Started with zPCR (IBM's Processor Capacity Reference)
page of the IBM Techdoc Library website.
As described in 12.5, “LSPR workload categories based on RNI” on page 483, the underlying
performance sensitive factor is how a workload interacts with the processor hardware.
For more information about RNI, see 12.5, “LSPR workload categories based on RNI” on
page 483.
The AVERAGE RNI LSPR workload is intended to match most client workloads. When no
other data is available, use the AVERAGE RNI LSPR workload for capacity analysis.
The CPU MF data can be used determine workload type. When available, this data allows the
RNI for a production workload to be calculated.
By using the RNI and another factor from CPU MF, the L1MP (percentage of data and
instruction references that miss the L1 cache), a workload can be classified as LOW,
AVERAGE, or HIGH RNI. This classification and resulting hit are automated in the zPCR tool.
It is preferable to use zPCR for capacity sizing.
Starting with z/OS V2R1 with APAR OA43366, zFS file is not required any more for CPU MF
and Hardware Instrumentation Services (HIS). HIS is a z/OS function that collects hardware
event data for processors in SMF records type 113, and a z/OS UNIX System Services output
files.
Only SMF 113 record is required to know proper workload type by using CPU MF counter
data. CPU overhead of CPUMF is minimal. Also, the amount of SMF 113 record is 1% of
typical SMF 70 and 72 which RMF writes.
CPU MF and HIS can use not only for deciding workload type but also use another purpose.
For example, starting with z/OS V2R1, you can record Instruction Counts in SMF type 30
record when you activate CPU MF. Therefore, we strongly recommend that you always
activate CPU MF.
For more information about getting CPUMF counter data, see the CPU MF - 2017 Update
and WSC Experiences of the IBM Techdoc Library website.
A holistic performance approach is required when the performance gains are reduced
because of frequency. Therefore, hardware and software synergy becomes an absolute
requirement.
Starting with z13, Instructions Per Cycle (IPC) improvements in core and cache became the
driving factor for performance gains. As these microarchitectural features increase (which
contributes to instruction parallelism), overall workload performance variability also increases
because not all workloads react the same way to these enhancements.
Because of the nature of the z15 multi-CPC drawer system and resource management
across those drawers, performance variability from application to application is expected.
Also, the memory and cache designs affect various workloads in many ways. All workloads
are improved, with cache-intensive loads benefiting the most. For example, having more PUs
per CPC drawer, each with higher capacity than z14, more workload can fit on a z15 CPC
drawer. This configuration can result in better performance. For example, z14 two drawer
system model M02 can populate maximum 69 PUs.
In contrast, z15 two drawer system Max71 can populate maximum 71 PUs. Therefore, two
more PUs can share caches and memories within the drawer, so the performance
improvements is expected.
The workload variability for moving from z13 and z14 to z15 expected to be stable. Workloads
that are migrating from z10 EC, z196, and zEC12 to z15 can expect to see similar results with
slightly less variability than the migration from z13 and z14.
Experience demonstrates that IBM Z servers can be run at up to 100% utilization levels,
sustained. However, most clients prefer to leave some room and run at 90% or slightly under.
Do not use MIPs or MSUs for capacity planning: Do not use “one number” capacity
comparisons, such as MIPs or MSUs. IBM does not officially announce the processor
performance as “MIPs”. MSU is only a number for software license charge and it does not
represent for performance for the processor.
CP3KEXTR is offered as a “no-charge” application. It can also create the EDF file for
ZCP3000. ZCP3000 is an IBM internal tool, but you can create the EDF file for it on your
system. For more information about CP3KEXTR, see the IBM Techdoc z/OS Data Extraction
Program (CP3KEXTR) for zPCR and zBNA.
For all optical links, the connector type is LC Duplex (except for the zHyperLink) and the
ICA SR connections, which are established with multifiber push-on (MPO) connectors.
The MPO connector of the zHyperLink, and the ICA connection features two rows of 12 fibers
and are interchangeable.
The electrical Ethernet cable for the Open Systems Adapter (OSA) connectivity is connected
through an RJ45 jack.
The attributes of the channel options that are supported on z15 are listed in Table A-1.
Open Systems Adapter (OSA) and Remote Direct Memory over Converged Ethernet (RoCE)
Parallel Sysplex
a. The link data rate does not represent the performance of the link. The performance depends
on many factors, including latency through the adapters, cable lengths, and the type of
workload.
b. Where applicable, the minimum fiber bandwidth distance in MHz-km for multi-mode fiber optic
links is included in parentheses.
c. A 600 meters maximum when sharing the switch across two RoCE Express features.
The unrepeated distances for different multimode fiber optic types for zHyperlink Express are
listed in Table A-2.
The maximum unrepeated distances for FICON SX features are listed in Table A-3.
OM3 860 meters 500 meters 380 meters 150 meters 100 meters
(50 µm at 2000 MHz·km)
2822 feet 1640 feet 1247 feet 492 feet 328 feet
OM4a N/A 500 meters 400 meters 190 meters 125 meters
(50 µm at 4700 MHz-km)
N/A 1640 feet 1312 feet 693 feet 410 feet
a. Fibre Channel Standard (not certified for Ethernet)
Note: System Recovery Boost is a new firmware feature and is available on z15 CPC. It is
not available on older systems.
Naming: The IBM z15 server generation is available as the following machine types and
models:
Machine Type 8561 (M/T 8561), Model T01, Features Max34, Max71, Max108,
Max145, and Max190, which is further identified as IBM z15 Model T01, or z15 T01,
unless otherwise specified.
Machine Type 8562 (M/T 8562), Model T02, Features Max4, Max13, Max21, Max32,
and Max65, which is further identified as IBM z15 Model T02, or z15 T02, unless
otherwise specified.
In the remainder of this appendix, IBM z15 (z15) refers to both machine types.
B.2 Functions
IBM Z Recovery boost is a new feature that was introduced with IBM z15 firmware (Driver 41),
which is designed to provide higher temporary processing capacity to LPARs for speeding up
shutdown, IPL, and stand-alone memory dump operations, as well as for short duration
The boost capacity does not contribute to other IBM software license charges.
Figure B-1 shows a typical System Recovery Boost timeline for z/OS.
D
A. Boost period: Faster
planned shutdown
A B. Faster GDPS-driven HW
C Reconfiguration Activities
(especially around DR
and site switch)
B C. Boost period: Faster IPL
D. Boost period: Faster
Middleware Restart and
Recovery
E. Boost period: Boosted
Capacity to do work
following IPL
Figure B-1 z/OS typical System Recovery Boost timeline
Boost timeline
For z/OS, current implementation allows 60 minutes boost period for IPL and 30 minutes for
shutdown events2. The 60 minutes boost period for IPL-ing the z/OS system also allow
catching up with the backlog work.
1
Supported operating images are enabled for boost that is running in an LPAR.
2 Boost periods are operating system-specific. Different operating systems can feature different boost periods.
For process recovery boosts, boost is provided up to 30 minutes (sum of boost times for
multiple processes) period over a period of 24 hours per system image (z/OS LPAR).
At the time of this writing, IBM z/OS (2.3 and 2.4 with PTFs), IBM z/VM 7.1 with PTFs, z/VSE
V6.2, and z/TPF 1.1 with PTFs can use subcapacity boost.
Note: Speed boost applies to general-purpose processors (CPs) only. All other engines
run at full capacity (IFLs, zIIPs, and ICFs).
In this example, three LPARs are defined in the z15 model 403. In the normal operation, all
work is dispatched on subcapacity CPs.
When LPARs enter a boost period, work that is dispatched from LPARz runs at CP7 (full
capacity). Other LPARs continue to be dispatched at CP4 (subcapacity). One boost period
In this period, the system can use following processors to run CP workload:
Entitled purchased CPs
Entitled purchased zIIPs
Extra zIIPs, which can be added by using the temporary boost capacity records (FC 9930
and FC 6802 on z15 T013).
If more logical zIIPs are available and configured in the LPAR profile (whether added by way
of the temporary boost capacity record or other temporary capacity activation) while in the
boost period, images bring more logical zIIP processors online to use the extra physical zIIP
capacity.
After the boost period ends, z/OS dispatching of work on CPs versus zIIPs returns to normal.
Important: These extra logical processors should be taken offline at the end of the boost
period (either manually or by using automation), or they are taken offline automatically
when the temporary boost record expires.
The start and end of the boost period is signaled by way of a console message, ENF signal
(84), and cutting an SMF record. Also, the start and end of boost period starts new SMF
interval. A system command or PROC (IEABE) is provided to allow for early opt-out of the
boost, if wanted.
3
System Recovery Boost upgrade (FC 9930 and 6802) is not available for z15 T02. Only base System Recovery
Boost is available on z15 T02 (built into the system).
In this example, three LPARs are defined on the z15 model 703 (M/T 8561) with two zIIPs.
Two zIIPs are shared between LPARy and LPARz.
During normal operation, only zIIP eligible work is dispatched to the zIIPs. When LPARz
enters a boost period, general-purpose work and zIIP eligible work might be dispatched to the
zIIPs.
When the boost period ends, only zIIPs-eligible work is dispatched to the zIIPs.
B.3.3 Optional System Recovery Boost Upgrade capacity record (z15 T01)
You can temporarily increase the number of physical zIIPs to use for System Recovery Boost.
This feature is the new System Recovery Boost Upgrade record that you can activate from
the HMC/SE Perform Model Conversion menu, or by using automation that calls the hardware
API services.
After it is activated, zIIPs are added to the zIIP pool and LPAR shares of the extra physical
zIIP capacity follows normal LPAR weight rules. You can set up the LPAR image profiles in
advance with extra logical zIIP processors to enable the effective use of the extra physical
zIIP processor capacity.
Deactivate the temporary boost capacity record at the end of the IPL or shutdown actions or
change windows for which they intend to provide a boost (the record auto-deactivates after 6
hours).
For systems that ordered the new System Recovery Boost Upgrade record, the zIIP:CP ratio
of 2:1 can be exceeded during the boost periods for the boost opt-in images (LPARs). The
boost record activates the zIIPs for up to 6 hours. The boost record has an expiration date of
one year.
Consider the following points regarding the optional System Recovery Boost Upgrade record:
In this example, three LPARs are defined in the z15 model 703 with two zIIPs. Two zIIPs are
assigned to both LPARy and LPARz. LPARy is in normal operation.
When LPARz is in a boost period; therefore, both general-purpose work and zIIP eligible work
can be dispatched to the zIIPs. LPARz includes a reserved zIIP specified in its image profile.
By activating temporary boost capacity record, one zIIP is added to the zIIP pool and
automatically allocated to the LPARz and brought online. In this boost period, general CP
work for LPARz is dispatched to zIIPs and CPs.
In a sysplex, WLM sysplex routing starts to route work away from a system after the shutdown
PROC is started to further accelerate shutdown.
In this example, three LPARs are defined in the z15 model 403 (M/T 8561) with two zIIPs. Two
zIIPs are assigned to LPARy and LPARz. During normal operation, all CP work is dispatched
at subcapacity (CP4). Only zIIP eligible work is dispatched to zIIPs.
Note: Stand-alone memory dump does not use zIIP for boost purposes.
Before the planned shutdown of LPARz, the IEASDBS proc is started by an operator or
automation. This process starts the shutdown boost. CP work that is dispatched by LPARz is
run at full-capacity (CP7) and also general-purpose workload is dispatched to zIIPs. LPARx
and LPARy continue in the normal operation at subcapacity (CP4) and only zIIP-eligible
workload is dispatched to zIIP.
Firmware changes on the z15 HMC and SEs support greater parallelism and performance
improvements in the HW API services.
4 z/OS use requires z/OS 2.4 with rollback to 2.3 with PTFs (for subcapacity and zIIP boost, on z15 CPC).
Important: The capabilities described in this section are available for both IBM z15 T01
and IBM z15 T02 running z/OS images member of a parallel Sysplex. These require:
Specific (concurrently installable) LPAR MCL for z15 Driver 41 or higher (check IBM
ResourceLink).
z/OS 2.4 with PTFs for APAR OA59813
z/OS 2.3 with PTFs for APAR OA59813
Note that z/OS APARs will be associated with new FIXCAT for System Recovery Boost
(Keyword SRB/K and Fixcat Name: IBM.Function.SystemRecoveryBoost).
With the enhanced System Recovery Boost support, IBM is extending the boost technologies
to provide short-term recovery process boost acceleration for other specific sysplex recovery
events in z/OS.
Currently, these sysplex recovery events often cause short-duration workload impacts or
workload spikes while the sysplex is busy recovering for a recovery event. Recovery affects
the normal execution of the client workload in the sysplex, until the recovery processing
completes.
The process recovery boost is designed to provide boosted processor capacity to mitigate
short-term recovery impacts, and restore normal sysplex steady-state sysplex operation as
quickly as possible following specific recovery events, as well as to provide boosted processor
capacity for a short period of time following restoration of steady-state operation, to help with
workload “catch-up” from the recovery event.
z/OS manages the recovery process boosts internally, with the operating system initiating the
boosts as these recovery events take place, and only on the images that are affected by these
events. If recovery process boosts happen to “overlap” – a second recovery process boost
occurs before a first one has used its entire boost period – then the overlapping boosts are
merged and the boost period may be extended to allow the full boost period time for the
second recovery process
Operational considerations
During a Recovery Process boost period, WLM neither routes work away from the system
(as it does during shutdown boost) nor towards the system (as it does during startup
boost). WLM ignores short-duration recovery boosts for workload routing purposes.
When bringing reserved logical zIIP processors online and offline at the start and end of a
recovery process boost period, z/OS limits the number of “transient” zIIPs brought
online/offline automatically to at most two (more transient zIIPs during IPL and shutdown
boost periods may be configured).
System Recovery Boost Upgrade record activation (used to activate additional temporary
zIIP capacity on the z15 T01) is NOT supported for use with recovery process boost
periods.
z/OS starts and ends a new SMF interval during a recovery process boost period, but
when two or more recovery process boosts “overlap” they are merged into a single boost
period and a single SMF interval
z/OS issues ENF signals and console messages as appropriate when starting, extending,
or stopping a recovery process boost
z/OS does not permit overlap between the use of recovery process boosts, and the longer
image-level startup/shutdown boosts:
– Recovery process boosts are not initiated while an image-level startup boost is still in
progress – the system is already boosted
– If a recovery process boost is in progress when a system image-level shutdown is
initiated, then z/OS “cancels” the in-progress recovery process boost, and initiates the
shutdown boost period for system shutdown
– If additional transient zIIPs were already online during the recovery process boost,
z/OS would potentially have to bring additional ones online for the shutdown boost, up
to the full quota of reserved logical zIIPs.
Important: The base System Recovery Boost capability is built into z15 firmware and does
not require ordering extra features. System Recovery Boost Upgrade (consisting of FC
9930 and FC 6802) is an optional, orderable feature that provides more temporary zIIP
capacity for use during boost periods, and is available on z15 T01 (M/T 8561) only.
Consider the following points:
FC 9930 is not required to use the base System Recovery Boost capability.
FC 9930 is only needed if more zIIP temporary capacity is required.
By default, System Recovery Boost is enabled for z/OS (opt-in). z/OS must run on z15
CPC.
For extra zIIP temporary boost capacity, FC 9930 and FC 6802 (System Recovery
Boost Upgrade) must be ordered with the z15 T01 CPC.
You can configure a z/OS system-level parameter (IEASYSxx) to control whether a specific
z/OS image opts in for the zIIP processor Boost, as shown in the following example:
BOOST=SYSTEM | ZIIP | SPEED | NONE
If you want to use offline zIIPs or extra zIIPs that are provided by the System Recovery Boost
record, you must define reserved zIIPs in the image profile, as shown in Figure B-6 on
page 503.
You also should review LPAR weights and storage allocation to ensure that they meet your
system requirements.
A new “boost class” is available for Recovery Process boosts, which appears in various
system messages, ENF signals, SMF fields, and other z/OS APIs.
Example 2:
IEE257I Boost State
Boost class: Recovery Process
Requestor: Hyperswap
zIIP boost: active with 2 transient zIIP cores
Speed boost: active
In addition, the DISPLAY M=CPU has also been enhanced (see Example B-2):
– “I” indicates zIIPs
– “B” indicates (transient) boost zIIPs. This CPU was configured online at the start of the
boost period, and will be configured offline when the boost ends
B.6 Automation
Your automation product can be used in the following ways to automate and control System
Recovery Boost activities:
To activate and deactivate the eBod temporary capacity record to provide extra physical
zIIPs for an IPL or Shutdown Boost.
To dynamically modify LPAR weights, as might be needed to modify or “skew” the sharing
of physical zIIP capacity during a Boost period.
To drive the invocation of the PROC that indicates the beginning of a shutdown process
(and the start of the shutdown Boost).
To take advantage of new composite hardware API reconfiguration actions.
To control the level of parallelism present in the workload at start (for example, starting
middleware regions) and shutdown (for example, perform an orderly shutdown of
middleware).
Automation can pace or throttle these activities to varying degrees; with Boost, less pacing
or more parallelism might be wanted.
To automate on the new z/OS messages that are issued at start or end of boost periods to
take whatever actions are needed.
System Recovery Boost records on z15 T01 require Boost Enablement contract (FC 9930)
and temporary pre-paid zIIP boost records (FC 6802).
An optional feature that is available for z14, z13, and z13s servers, zEDC Express addressed
customer requirements by providing hardware-based acceleration for data compression and
decompression. zEDC provided data compression with lower CPU consumption than
compression technology that was available on the IBM Z server.
Existing clients deployed zEDC compression to deliver the following types of compression:
Storage
Data transfer
Database
In-application
Many clients are increasing their zEDC footprint to 8GBps with up to 16 features per z14
system at 1 GBps throughput per feature (redundancy reduces total throughput to 8 GBps).
The z15 further addresses the growth of data compression requirements with the integrated
on-chip compression unit (implemented in processor Nest, one per PU chip) that significantly
increases compression throughput and speed compared to zEDC deployments.
The z15 Integrated Accelerator for zEDC delivers industry-leading throughput and replaces
the zEDC Express PCIe adapter that is available on the IBM z14 and earlier servers.
One Nest Accelerator Unit is available per processor chip, which is shared by all cores on the
chip and features the following benefits:
New concept of sharing and operating an accelerator function in the nest
Supports DEFLATE compliant compression/decompression and GZIP CRC/ZLIB Adler
Low latency
High bandwidth
Problem state execution
Hardware and firmware interlocks to ensure system responsiveness
Designed instruction
Run in millicode
Moving the compression function from the (PCIe) I/O drawer to the processor chip means that
compression can operate directly in L3 cache and data does not need to be passed by using
I/O operations.
Compression modes
Compression is run in one of the following modes:
Synchronous
Execution occurs in problem state where the user application starts the instruction in its
virtual address space.
Asynchronous
Execution is optimized for Large Operations under z/OS for authorized applications (for
example, BSAM) and issues I/O by using EADMF for asynchronous execution.
This type of execution maintains the current user experience and provides a transparent
implementation for authorized users of zEDC.
Note: The zEDC Express feature does not carry forward to z15.
The IFAPRDxx feature is still required for authorized services. For problem state services,
such as zlib usage of Java, it is not required.
The current zlib and the new zlib function on z14 and earlier servers and z15 hardware. It
functions with or without the z15 z/OS PTFs on z14 and earlier servers.
Software support
Support of the On-Chip Compression function is compatible with zEDC support and is
available in z/OS V2R1 or later for data compression and decompression. Support for data
recovery (decompression) in the case that zEDC or On-Chip Compression not available;
however, it is provided through software in z/OS V2R2, and V2R1 with the appropriate
program temporary fixes (PTFs).
A specific fix category that is named IBM.Function.zEDC identifies the fixes that enable or use
the zEDC and On-Chip Compression function.
z/OS guests that run under z/VM V6.4 with PTFs and later can use the zEDC Express feature
and z15 On-Chip Compression.
For more information, see the Enhancements to z/VM 6.4 page of the IBM Systems website.
IBM 31-bit and 64-bit SDK for z/OS Java Technology Edition, Version 7 Release 1 (5655-W43
and 5655-W44) (IBM SDK 7 for z/OS Java) now provides use of the zEDC Express feature
and Shared Memory Communications-Remote Direct Memory Access (SMC-R), which is
used by the 10GbE RoCE Express feature.
For more information about how to implement and use the IBM Z compression features, see
Reduce Storage Occupancy and Increase Operations Efficiency with IBM zEnterprise Data
Compression, SG24-8259.
zBNA is based on Microsoft Windows, and provides graphical and text reports, including
Gantt charts, and support for alternative processors.
zBNA can be used to analyze client-provided System Management Facilities (SMF) records
to identify jobs and data sets that are candidates for zEDC and z15 On-Chip Compression
across a specified time window (often a batch window).
Therefore, zBNA can help you estimate the use of On-Chip Compression features and help
identify savings. The following resources are available:
IBM Employees can obtain zBNA and other CPS tools at the IBM Z Batch Network
Analyzer (zBNA) Tool page of the IBM Techdoc website.
IBM Business Partners can obtain zBNA and other CPS tools at the IBM PartnerWorld
website (log in required).
IBM clients can obtain zBNA and other CPS tools at the IBM Z Batch Network Analyzer
(zBNA) Tool page of the IBM Techdoc Library website.
The z15 On-Chip Compression accelerator solves these virtualization limitations because the
function is no longer an I/O device and is available as a problem state instruction to all Linux
on Z guests without constraints.
The common building blocks are displayed and range from 1 - 4 frames, with various numbers
of CPC drawers, PCIe+ I/O drawers.
The publications that are listed in this section are considered particularly suitable for a more
detailed discussion of the topics that are covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide more information about the topic in this
document. Note that some publications that are referenced in this list might be available in
softcopy only:
IBM z15 Technical Introduction, SG24-8850
IBM Z Connectivity Handbook, SG24-5444
IBM z14 (M/T 3907) Technical Guide, SG24-8451
IBM z14 ZR1 Technical Guide, SG24-8651
IBM z14 (M/T 3906) Technical Introduction, SG24-8450
IBM z14 ZR1 Technical Introduction, SG24-8550
You can search for, view, download, or order these documents and other Redbooks,
Redpapers, Web Docs, draft and additional materials, at the following website:
ibm.com/redbooks
Other publications
The publication IBM Z 8561 Installation Manual for Physical Planning, GC28-7002 also is
relevant as another information source.
Online resources
The IBM Resource Link for documentation and tools website is also relevant as another
information source:
http://www.ibm.com/servers/resourcelink
SG24-8851-00
ISBN 0738458120
Printed in U.S.A.
®
ibm.com/redbooks