IBM EC12 sg248049
IBM EC12 sg248049
IBM EC12 sg248049
Ivan Dobos Wolfgang Fries Parwez Hamid Octavian Lascu Gerard Laumay Swee Seng Ng Fernando Nogal Frank Packheiser Vicente Ranieri Jr Karan Singh Andr Spahni Esra Ufacik Hans Wijngaard Zhaoxu Zhang
ibm.com/redbooks
8049edno.fm
International Technical Support Organization IBM zEnterprise EC12 Technical Guide August 2012
SG24-8049-00
8049edno.fm
Note: Before using this information and the product it supports, read the information in Notices on page xv.
First Edition (August 2012) This edition applies to the IBM zEnterprise EC12 and the IBM zEnterprise BladeCenter Extension Model 003. This document was created or updated on August 29, 2012. Note: This book is based on a pre-GA version of a product and may not apply when the product becomes generally available. We recommend that you consult the product documentation or follow-on versions of this redbook for more current information.
Copyright International Business Machines Corporation 2010. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp.
8049edno.fm
iii
8049edno.fm
iv
8049TOC.fm
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix The team who wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxii Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxii Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxii Chapter 1. Introducing the IBM zEnterprise EC12 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.1 IBM zEnterprise EC12 elements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.2 zEC12 highlights . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.2.1 Flash Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.3 zEC12 technical overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.3.1 Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.3.2 Model upgrade paths . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.3.3 Frames . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 1.3.4 Processor cage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 1.3.5 I/O connectivity, PCIe, and InfiniBand . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 1.3.6 I/O subsystems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 1.3.7 Cryptography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 1.3.8 Capacity on Demand (CoD) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 1.3.9 Parallel Sysplex support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 1.4 IBM zEnterprise BladeCenter Extension (zBX) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 1.4.1 Blades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 1.4.2 IBM WebSphere DataPower Integration Appliance XI50 for zEnterprise . . . . . . . 18 1.5 Unified Resource Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 1.6 Hardware Management Consoles and Support Elements . . . . . . . . . . . . . . . . . . . . . . 19 1.7 Operating systems and software. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 1.8 Reliability, availability, and serviceability. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 1.9 Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 1.9.1 LSPR workload suite. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 1.9.2 Fundamental components of workload capacity performance . . . . . . . . . . . . . . . 22 1.9.3 Relative nest intensity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 1.9.4 LSPR workload categories based on relative nest intensity . . . . . . . . . . . . . . . . . 25 1.9.5 Relating production workloads to LSPR workloads . . . . . . . . . . . . . . . . . . . . . . . 26 1.9.6 Workload performance variation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 1.9.7 Main performance improvement drivers with zEC12 . . . . . . . . . . . . . . . . . . . . . . 28 Chapter 2. CPC Hardware Components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1 Frames and cage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1 Frame A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2 Frame Z . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.3 I/O cages, I/O drawers, and PCIe I/O drawers . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.4 Top exit I/O cabling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Book concept . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.1 Book interconnect topology. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2 Oscillator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.3 System control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Copyright IBM Corp. 2010. All rights reserved.
31 32 33 34 34 35 36 37 39 40 v
8049TOC.fm
2.2.4 Book power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Multi-chip module . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4 Processor units and storage control chips . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.1 PU chip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.2 Processor unit (core) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.3 PU characterization. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.4 Storage control (SC) chip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.5 Cache level structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5 Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.1 Memory subsystem topology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.2 Redundant array of independent memory (RAIM) . . . . . . . . . . . . . . . . . . . . . . . . 2.5.3 Memory configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.4 Memory upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.5 Book replacement and memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.6 Flexible Memory Option . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.7 Preplanned Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.6 Reliability, availability, serviceability (RAS) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.6.1 RAS in the CPC memory subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.6.2 General zEC12 RAS features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7 Connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7.1 Redundant I/O interconnect . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7.2 Enhanced book availability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.7.3 Book upgrade . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8 Model configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.1 Upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.2 Concurrent PU conversions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.3 Model capacity identifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.4 Model capacity identifier and MSU value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.5 Capacity Backup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.8.6 On/Off Capacity on Demand and CPs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9 Power and cooling. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.1 Power consumption . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.2 High voltage DC power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.3 Internal Battery Feature (IBF) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.4 Power capping and power saving . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.5 Power estimation tool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.6 Cooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.7 Radiator Unit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.8 Water Cooling Unit (WCU) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.9.9 Backup air cooling system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.10 Summary of zEC12 structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 3. CPC System design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1 Design highlights . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Book design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.1 Cache levels and memory structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.2 Book interconnect topology. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3 Processor unit design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.1 Out-of-order execution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.2 Superscalar processor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.3 Compression and cryptography accelerators on a chip . . . . . . . . . . . . . . . . . . . . 3.3.4 Decimal floating point accelerator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.5 IEEE floating point . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
40 40 41 42 43 45 45 47 48 49 49 50 53 53 53 54 55 55 56 57 59 60 61 61 62 63 64 65 65 67 68 68 69 69 70 70 70 71 72 75 76 77 78 79 79 82 82 83 86 86 87 88
vi
8049TOC.fm
3.3.6 Processor error detection and recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 3.3.7 Branch prediction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 3.3.8 Wild branch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 3.3.9 Translation look-aside buffer. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 3.3.10 Instruction fetching, decode, and grouping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 3.3.11 Extended translation facility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 3.3.12 Instruction set extensions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 3.3.13 Transactional execution (TX) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 3.3.14 Run-time instrumentation (RI) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 3.4 Processor unit functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 3.4.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 3.4.2 Central processors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 3.4.3 Integrated Facility for Linux . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94 3.4.4 Internal Coupling Facility. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94 3.4.5 System z Application Assist Processors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95 3.4.6 System z Integrated Information Processor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 3.4.7 zAAP on zIIP capability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100 3.4.8 System Assist Processors. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 3.4.9 Reserved processors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 3.4.10 Processor unit assignment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 3.4.11 Sparing rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 3.4.12 Increased flexibility with z/VM-mode partitions . . . . . . . . . . . . . . . . . . . . . . . . . 103 3.5 Memory design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 3.5.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 3.5.2 Central storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 3.5.3 Expanded storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 3.5.4 Hardware system area . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 3.6 Logical partitioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 3.6.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 3.6.2 Storage operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112 3.6.3 Reserved storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115 3.6.4 Logical partition storage granularity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 3.6.5 LPAR dynamic storage reconfiguration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 3.7 Intelligent resource director . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117 3.8 Clustering technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118 3.8.1 Coupling facility control code . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120 3.8.2 Dynamic CF dispatching . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121 Chapter 4. CPC I/O System Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1 Introduction to InfiniBand and PCIe . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.1 Infrastructure types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.2 InfiniBand specification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.3 Data, signalling, and link rates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.4 PCIe . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2 I/O system overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.1 Characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.2 Summary of supported I/O features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3 I/O cages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4 I/O drawers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.5 PCIe I/O drawers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.6 I/O cage, I/O drawer and PCIe I/O drawer offerings . . . . . . . . . . . . . . . . . . . . . . . . . . 4.6.1 I//O cage and I/O drawer carry forward . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7 Fanouts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 124 124 125 125 125 126 126 126 127 130 132 134 135 136
Contents
vii
8049TOC.fm
4.7.1 HCA2-C fanout (FC0162) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.2 PCIe copper fanout (FC0169) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.3 HCA2-O (12xIFB) fanout (FC0163). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.4 HCA2-O LR (1xIFB) fanout (FC0168) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.5 HCA3-O (12xIFB) fanout (FC0171). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.6 HCA3-O LR (1xIFB) fanout (FC0170) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.7 Fanout considerations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.7.8 Fanout summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8 I/O feature cards . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.1 I/O feature card types ordering information. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.8.2 PCHID report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9 Connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.1 I/O feature support and configuration rules. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.2 FICON channels . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.3 OSA-Express4S features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.4 OSA-Express3 features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.5 OSA-Express for ensemble connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.9.6 HiperSockets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10 Parallel Sysplex connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10.1 Coupling links . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.10.2 External clock facility. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11 Cryptographic functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.1 CPACF functions (FC 3863) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.2 Crypto Express4S feature (FC 0865) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.11.3 Crypto Express3 feature (FC 0864) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.12 Flash Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 5. CPC Channel Subsystem. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1 Channel subsystem. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.1 Multiple CSSs concept . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.2 CSS elements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.3 Multiple subchannel sets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.4 Parallel access volumes and extended address volumes. . . . . . . . . . . . . . . . . . 5.1.5 Logical partition name and identification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.6 Physical channel ID . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.7 Channel spanning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.8 Multiple CSS construct . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.9 Adapter ID . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1.10 Channel subsystem enhancement for I/O resilience. . . . . . . . . . . . . . . . . . . . . 5.2 I/O configuration management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.3 Channel subsystem summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4 System-initiated CHPID reconfiguration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5 Multipath initial program load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 6. Cryptography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.1 Cryptographic synchronous functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2 Cryptographic asynchronous functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.1 Secure key functions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.2 Additional functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3 CPACF protected key . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4 PKCS #11 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.1 The PKCS #11 model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4.2 z/OS PKCS #11 implementation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
138 138 138 139 140 140 141 143 144 144 145 146 147 149 153 155 158 159 161 161 166 167 167 167 167 167 169 170 170 171 171 174 174 175 175 176 177 177 178 178 180 181 183 184 184 184 185 187 189 189 190
viii
8049TOC.fm
6.4.3 Secure IBM Enterprise PKCS #11 (EP11) Coprocessor. . . . . . . . . . . . . . . . . . . 6.5 Cryptographic feature codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.6 CP Assist for Cryptographic Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.7 Crypto Express4S . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.7.1 Configuration rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8 Crypto Express3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.8.1 Configuration rules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.9 Tasks performed by PCIe Crypto Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.9.1 PCIe Crypto Express as a CCA coprocessor . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.9.2 PCIe Crypto Express as an EP11 coprocessor . . . . . . . . . . . . . . . . . . . . . . . . . 6.9.3 PCIe Crypto Express as an Accelerator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.9.4 IBM Common Cryptographic Architecture (CCA) Enhancements. . . . . . . . . . . . 6.10 TKE workstation feature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.10.1 Logical partition, TKE host, and TKE target . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.10.2 Optional smart card reader . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.11 Cryptographic functions comparison . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.12 Software support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 7. zBX zEnterprise BladeCenter Extension Model 003 . . . . . . . . . . . . . . . . . 7.1 zBX concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2 zBX hardware description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.1 zBX racks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.2 Top of rack (TOR) switches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.3 zBX BladeCenter chassis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.4 zBX blades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2.5 Power distribution unit (PDU) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3 zBX entitlements, firmware, and upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.1 zBX management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.2 zBX firmware. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4 zBX connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.1 Intranode management network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.2 Primary and alternate HMCs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.3 Intraensemble data network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.4 Network connectivity rules with zBX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.5 Network security considerations with zBX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4.6 zBX storage connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5 zBX connectivity examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5.1 A single node ensemble with a zBX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5.2 A dual node ensemble with a single zBX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5.3 A dual node ensemble with two zBXs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.6 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 8. Software support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1 Operating systems summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2 Support by operating system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.1 z/OS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.2 z/VM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.3 z/VSE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.4 z/TPF . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.5 Linux on System z. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.6 zEC12 functions support summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3 Support by function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.1 Single system image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
192 193 194 194 196 197 198 199 200 201 202 202 203 207 207 208 210 211 212 212 213 214 216 218 223 223 225 225 226 227 229 230 233 233 235 237 238 239 240 241 243 244 244 245 245 245 245 245 246 256 257
Contents
ix
8049TOC.fm
8.3.2 zAAP support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.3 zIIP support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.4 zAAP on zIIP capability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.5 Transactional Execution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.6 Maximum main storage size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.7 Flash Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.8 Large page support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.9 Guest support for execute-extensions facility . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.10 Hardware decimal floating point . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.11 Up to 60 logical partitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.12 Separate LPAR management of PUs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.13 Dynamic LPAR memory upgrade . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.14 Capacity Provisioning Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.15 Dynamic PU add . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.16 HiperDispatch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.17 The 63.75 K Subchannels. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.18 Multiple Subchannel Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.19 Third Subchannel Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.20 IPL from an alternate subchannel set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.21 MIDAW facility. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.22 HiperSockets Completion Queue . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.23 HiperSockets integration with the intraensemble data network (IEDN) . . . . . . 8.3.24 HiperSockets Virtual Switch Bridge. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.25 HiperSockets Multiple Write Facility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.26 HiperSockets IPv6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.27 HiperSockets Layer 2 support. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.28 HiperSockets network traffic analyzer for Linux on System z . . . . . . . . . . . . . . 8.3.29 FICON Express8S . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.30 FICON Express8 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.31 z/OS discovery and autoconfiguration (zDAC) . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.32 High performance FICON (zHPF) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.33 Request node identification data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.34 Extended distance FICON . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.35 Platform and name server registration in FICON channel . . . . . . . . . . . . . . . . 8.3.36 FICON link incident reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.37 FCP provides increased performance. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.38 N_Port ID virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.39 OSA-Express4S 1000BASE-T Ethernet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.40 OSA-Express4S 10 Gigabit Ethernet LR and SR . . . . . . . . . . . . . . . . . . . . . . . 8.3.41 OSA-Express4S Gigabit Ethernet LX and SX. . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.42 OSA-Express3 10 Gigabit Ethernet LR and SR . . . . . . . . . . . . . . . . . . . . . . . . 8.3.43 OSA-Express3 Gigabit Ethernet LX and SX . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.44 OSA-Express3 1000BASE-T Ethernet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.45 Open Systems Adapter for IBM zAware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.46 Open Systems Adapter for Ensemble. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.47 Intranode management network (INMN). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.48 Intraensemble data network (IEDN) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.49 OSA-Express4S NCP support (OSN) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.50 Integrated Console Controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.51 VLAN management enhancements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.52 GARP VLAN Registration Protocol . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.53 Inbound workload queueing (IWQ) for OSA-Express4S and OSA-Express3 . . 8.3.54 Inbound workload queueing (IWQ) for Enterprise Extender . . . . . . . . . . . . . . . x
IBM zEnterprise EC12 Technical Guide
258 259 259 260 260 261 262 262 262 263 263 264 264 264 265 265 265 266 266 266 266 267 267 268 268 268 269 269 270 270 271 273 273 273 274 274 274 274 275 276 276 277 278 279 279 280 280 280 281 281 281 282 283
8049TOC.fm
8.3.55 Query and display OSA configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.56 Link aggregation support for z/VM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.57 QDIO data connection isolation for z/VM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.58 QDIO interface isolation for z/OS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.59 QDIO optimized latency mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.60 Large send for IPv6 packets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.61 OSA-Express4S checksum offload . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.62 Checksum offload for IPv4 packets when in QDIO mode . . . . . . . . . . . . . . . . . 8.3.63 Adapter interruptions for QDIO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.64 OSA Dynamic LAN idle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.65 OSA Layer 3 Virtual MAC for z/OS environments. . . . . . . . . . . . . . . . . . . . . . . 8.3.66 QDIO Diagnostic Synchronization. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.67 Network Traffic Analyzer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.68 Program directed re-IPL . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.69 Coupling over InfiniBand . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.70 Dynamic I/O support for InfiniBand CHPIDs . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4 Cryptographic support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.1 CP Assist for Cryptographic Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.2 Crypto Express4S . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.3 Crypto Express3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.4 Web deliverables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.5 z/OS ICSF FMIDs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4.6 ICSF migration considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5 z/OS migration considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.1 General guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.2 HCD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.3 InfiniBand coupling links . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.4 Large page support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.5 HiperDispatch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.6 Capacity Provisioning Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5.7 Decimal floating point and z/OS XL C/C++ considerations. . . . . . . . . . . . . . . . . 8.5.8 IBM System z Advanced Workload Analysis Reporter (IBM zAware). . . . . . . . . 8.6 Coupling facility and CFCC considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7 MIDAW facility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.1 MIDAW technical description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.2 Extended format data sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.7.3 Performance benefits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.8 IOCP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.9 Worldwide portname (WWPN) tool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.10 ICKDSF . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.11 zEnterprise BladeCenter Extension (zBX) Model 003 software support . . . . . . . . . . 8.11.1 IBM Blades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.11.2 IBM WebSphere DataPower Integration Appliance XI50 for zEnterprise . . . . . 8.12 Software licensing considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.12.1 MLC pricing metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.12.2 Advanced Workload License Charges (AWLC) . . . . . . . . . . . . . . . . . . . . . . . . 8.12.3 System z New Application License Charges (zNALC) . . . . . . . . . . . . . . . . . . . 8.12.4 Select Application License Charges (SALC). . . . . . . . . . . . . . . . . . . . . . . . . . . 8.12.5 Midrange Workload Licence Charges (MWLC). . . . . . . . . . . . . . . . . . . . . . . . . 8.12.6 Parallel Sysplex Licence Charges (PSLC) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.12.7 System z International Program License Agreement (IPLA). . . . . . . . . . . . . . . 8.13 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
283 283 283 284 284 284 284 285 285 286 286 286 286 287 287 287 288 288 289 289 289 290 294 295 295 295 295 295 296 296 296 297 297 299 299 301 302 302 303 303 304 304 304 305 306 307 307 308 308 308 309 309
Contents
xi
8049TOC.fm
Chapter 9. System upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.1 Upgrade types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.1.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.1.2 Permanent upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.1.3 Temporary upgrades. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2 Concurrent upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2.1 Model upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2.2 Customer Initiated Upgrade facility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2.3 Summary of concurrent upgrade functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3 MES upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.1 MES upgrade for processors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.2 MES upgrades for memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.3 MES upgrades for I/O . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.4 MES upgrades for the zBX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.5 Summary of Plan-ahead features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.4 Permanent upgrade through the CIU facility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.4.1 Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.4.2 Retrieval and activation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5 On/Off Capacity on Demand . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5.2 Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5.3 On/Off CoD testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5.4 Activation and deactivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5.5 Termination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.5.6 z/OS capacity provisioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.6 Capacity for Planned Event. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.7 Capacity Backup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.7.1 Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.7.2 CBU activation and deactivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.7.3 Automatic CBU enablement for GDPS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.8 Nondisruptive upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.8.1 Components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.8.2 Concurrent upgrade considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.9 Summary of Capacity on Demand offerings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.10 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 10. RAS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.1 zEC12 availability characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.2 zEC12 RAS functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.2.1 Scheduled outages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.2.2 Unscheduled outages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.3 zEC12 Enhanced book availability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.3.1 EBA planning considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.3.2 Enhanced book availability processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.4 zEC12 Enhanced driver maintenance. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.5 RAS capability for the HMC and SE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.6 RAS capability for zBX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.7 Considerations for PowerHA in zBX environment. . . . . . . . . . . . . . . . . . . . . . . . . . . 10.8 IBM System z Advanced Workload Analysis Reporter (IBM zAware). . . . . . . . . . . . 10.9 RAS capability for Flash Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
311 312 312 315 316 316 317 319 322 322 323 325 326 327 328 328 330 331 332 332 333 336 337 338 338 342 344 344 346 347 347 348 349 352 353 355 356 357 358 359 360 360 362 368 369 370 371 373 374
Chapter 11. Environmental requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375 11.1 zEC12 power and cooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376 11.1.1 Power consumption . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376 xii
IBM zEnterprise EC12 Technical Guide
8049TOC.fm
11.1.2 Internal Battery Feature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.1.3 Emergency power-off . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.1.4 Cooling requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.2 IBM zEnterprise EC12 physical specifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3 IBM zEnterprise EC12 physical planning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3.1 Raised floor or non-raised floor. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3.2 Top Exit Power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3.3 Top Exit I/O Cabling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3.4 Weight distribution plate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3.5 3-in-1 bolt down kit for raised floor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.4 zBX environmental requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.4.1 zBX configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.4.2 zBX power components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.4.3 zBX cooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.4.4 zBX physical specifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.5 Energy management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.5.1 Power estimation tool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.5.2 Query maximum potential power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.5.3 System Activity Display and Monitors Dashboard. . . . . . . . . . . . . . . . . . . . . . . 11.5.4 IBM Systems Director Active Energy Manager . . . . . . . . . . . . . . . . . . . . . . . . . 11.5.5 Unified Resource Manager: Energy Management . . . . . . . . . . . . . . . . . . . . . . Chapter 12. Hardware Management Console and Support Element . . . . . . . . . . . . . 12.1 Introduction to HMC and SE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.1 HMC and SE enhancements and changes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.2 HMC media support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.3 Tree Style User Interface and Classic Style User Interface . . . . . . . . . . . . . . . 12.2 HMC and SE connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.1 Hardware prerequisites news . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.2 TCP/IP Version 6 on HMC and SE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.3 Assigning addresses to HMC and SE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.3 Remote Support Facility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4 HMC and SE remote operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5 HMC and SE key capabilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.1 CPC management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.2 LPAR management. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.3 Operating system communication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.4 HMC and SE microcode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.5 Monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.6 Capacity on Demand support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.7 Server Time Protocol support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.8 NTP client/server support on HMC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.9 Security and user ID management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.10 System Input/Output Configuration Analyzer on the SE and HMC . . . . . . . . . 12.5.11 Automated operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.12 Cryptographic support. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.13 z/VM virtual machine management. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.5.14 Installation support for z/VM using the HMC. . . . . . . . . . . . . . . . . . . . . . . . . . 12.6 HMC in an ensemble. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.6.1 Unified Resource Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.6.2 Ensemble definition and management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.6.3 HMC availability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.6.4 Considerations for multiple HMCs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
377 378 378 381 382 382 383 384 385 385 385 385 386 387 389 390 391 391 392 392 393 397 398 399 401 401 401 403 404 404 405 406 406 406 407 407 407 409 412 413 414 417 418 419 419 421 421 422 422 424 426 426
Contents
xiii
8049TOC.fm
12.6.5 HMC browser session to a primary HMC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 426 12.6.6 HMC ensemble topology. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 427 Appendix A. IBM zAware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . A.1 Troubleshooting in Complex IT Environments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . A.1.1 Smarter computing needs smarter monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . A.2 Introducing the IBM zAware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . A.2.1 Value of IBM zAware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . A.2.2 IBM z/OS Solutions to Improve Problem Diagnostics. . . . . . . . . . . . . . . . . . . . . A.3 IBM zAware Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . A.4 IBM zAware PreRequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . A.5 Configuring and using the IBM zAware virtual appliance . . . . . . . . . . . . . . . . . . . . . . 429 430 430 431 431 431 432 438 439
Appendix B. Channel options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 441 Appendix C. Flash Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C.1 Flash Express overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C.2 Using Flash Express. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C.3 Security on Flash Express . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C.3.1 Integrated Key Controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C.3.2 Key Serving Topology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C.3.3 Error recovery scenarios . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Other publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 445 446 448 451 451 453 454 455 455 455 455 455
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 457
xiv
8049spec.fm
Notices
This information was developed for products and services offered in the U.S.A. IBM may not offer the products, services, or features discussed in this document in other countries. Consult your local IBM representative for information on the products and services currently available in your area. Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and verify the operation of any non-IBM product, program, or service. IBM may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not grant you any license to these patents. You can send license inquiries, in writing, to: IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A. IBM's statements regarding its plans, directions, and intent are subject to change or withdrawal without notice at IBM's sole discretion. Information regarding potential future products is intended to outline our general product direction and it should not be relied on in making a purchasing decision. The information mentioned regarding potential future products is not a commitment, promise, or legal obligation to deliver any material, code, or functionality. The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. Any references in this information to non-IBM websites are provided for convenience only and do not in any manner serve as an endorsement of those websites. The materials at those websites are not part of the materials for this IBM product and use of those websites is at your own risk. IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. Any performance data contained herein was determined in a controlled environment. Therefore, the results obtained in other operating environments may vary significantly. Some measurements may have been made on development-level systems and there is no guarantee that these measurements will be the same on generally available systems. Furthermore, some measurements may have been estimated through extrapolation. Actual results may vary. Users of this document should verify the applicable data for their specific environment. Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. To illustrate them as completely as possible, the examples include the names of individuals, companies, brands, and products. All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. COPYRIGHT LICENSE: This information contains sample application programs in source language, which illustrate programming techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
Copyright IBM Corp. 2010. All rights reserved.
xv
8049spec.fm
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of these programs.
xvi
8049spec.fm
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or both. These and other IBM trademarked terms are marked on their first occurrence in this information with the appropriate symbol ( or ), indicating US registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in other countries. A current list of IBM trademarks is available on the Web at http://www.ibm.com/legal/copytrade.shtml The following terms are trademarks of the International Business Machines Corporation in the United States, other countries, or both:
AIX BladeCenter CICS DataPower DB2 Connect DB2 Distributed Relational Database Architecture Domino DRDA DS8000 ECKD ESCON FICON FlashCopy GDPS Geographically Dispersed Parallel Sysplex HACMP HiperSockets IBM Systems Director Active Energy Manager IBM IMS Language Environment Lotus MQSeries OMEGAMON Parallel Sysplex Passport Advantage Power Systems POWER6 POWER7 PowerHA PowerPC PowerVM POWER PR/SM Processor Resource/Systems Manager RACF Redbooks Redbooks (logo) Resource Link Resource Measurement Facility RETAIN RMF Sysplex Timer System p System Storage System x System z10 System z9 System z Tivoli WebSphere z/Architecture z/OS z/VM z/VSE z10 z9 zEnterprise
The following terms are trademarks of other companies: Intel, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. Linux is a trademark of Linus Torvalds in the United States, other countries, or both. Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Java, and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its affiliates. UNIX is a registered trademark of The Open Group in the United States and other countries. Other company, product, or service names may be trademarks or service marks of others.
Notices
xvii
8049spec.fm
xviii
8049pref.fm
Preface
The popularity of the Internet and the affordability of IT hardware and software have resulted in an explosion of applications, architectures, and platforms. Workloads have changed. Many applications, including mission-critical ones, are deployed on a variety of platforms, and the System z design has adapted to this change. It takes into account a wide range of factors, including compatibility and investment protection, to match the IT requirements of an enterprise. This IBM Redbooks publication discusses the new IBM zEnterprise System, which consists of the IBM zEnterprise EC12 (zEC12), an updated IBM zEnterprise Unified Resource Manager, and the IBM zEnterprise BladeCenter Extension (zBX) Model 003. The zEC12 is designed with improved scalability, performance, security, resiliency, availability, and virtualization. The superscalar design allows the zEC12 to deliver a record level of capacity over the prior System z servers. It is powered by 120 of the worlds most powerful microprocessors running at 5.5 GHz and is capable of executing more than 75,000 millions of instructions per second (MIPS). The zEC12 Model HA1 is estimated to provide up to 50% more total system capacity than the z196 Model M80. The zBX Model 003 infrastructure works with the zEC12 to enhance System z virtualization and management through an integrated hardware platform that spans mainframe, POWER7, and System x technologies. Through the Unified Resource Manager, the zEnterprise System is managed as a single pool of resources, integrating system and workload management across the environment. This book provides information about the zEnterprise System and its functions, features, and associated software support. Greater detail is offered in areas relevant to technical planning. This book is intended for systems engineers, consultants, planners, and anyone wanting to understand the zEnterprise System functions and plan for their usage. It is not intended as an introduction to mainframes. Readers are expected to be generally familiar with existing IBM System z technology and terminology.
xix
8049pref.fm
IBM internal and external conferences. His areas of expertise include System z security, Parallel Sysplex, System z hardware, and z/OS. Vicente Ranieri is certified as a Distinguished IT Specialist by the Open group. Vicente is a member of the Technology Leadership Council Brazil (TLC-BR), and he is also a member of the IBM Academy of Technology. Gerard Laumay is a System z IT Specialist in the IBM Products and Solutions Support Centre (PSSC) in Montpellier, France. Part of the System z New Technology Introduction (NTI) team managing System z Early Support Programs (ESP) in Europe IOT, he has more than 26 years of experience in the large systems field, as a consultant with IBMs customers. His areas of expertise include IBM System z hardware, operating systems (z/OS, z/VM and Linux for System z), IBM middleware, new workloads and solutions on System z platform. Member of the zChampion worldwide technical community, he teaches at numerous IBM external and internal conferences. Gerard participates to international projects, and he has co-authored several IBM Redbook publications. Parwez Hamid is an Executive IT Consultant with the IBM Server and Technology Group. Over the past 38 years he has worked in various IT roles within IBM. Since 1988 he has worked with a large number of IBM mainframe customers, mainly introducing new technology. Currently, he provides pre-sales technical support for the IBM System z product portfolio and is the lead System z technical specialist for the UK and Ireland. Parwez has co-authored a number of ITSO Redbooks publications and he prepares technical material for the worldwide announcement of System z servers. He is the technical lead for the IBM System z zChampions team in IBM Europe. This team is made up of key IBM System z client-facing technical professionals from the Server and Software groups. Parwez works closely with System z product development in Poughkeepsie, and provides input and feedback for future product plans. Parwez is a Technical Staff member of the IBM UK Technical Council composed of senior technical specialists representing all IBM Client, Consulting, Services. and Product groups. Parwez teaches and presents at numerous IBM user group and internal conferences. Swee Seng Ng is a senior IT specialist working as a Technical Sales Support for IBM Singapore. Fernando Nogal is an IBM Certified Consulting IT Specialist working as an STG Technical Consultant for the Spain, Portugal, Greece, and Israel IMT. He specializes in advanced infrastructures and architectures. In his 30+ years with IBM, he has held a variety of technical positions, mainly providing support for mainframe clients. Previously, he was on assignment to the Europe Middle East and Africa (EMEA) Ssytem z Technical Support group, working full-time on complex solutions for e-business. His job included, and still does, presenting and consulting in architectures and infrastructures, and providing strategic guidance to System z clients regarding the establishment and enablement of advanced technologies on System z, including the z/OS, z/VM, and Linux environments. He is a zChampion and a member of the System z Business Leaders Council. An accomplished writer, he has authored and co-authored over 28 IBM Redbooks publications and several technical papers. Other activities include serving as a University Ambassador. He travels extensively on direct client engagements and as a speaker at IBM and client events and trade shows. Frank Packheiser is a Senior zIT Specialist at the Field Technical Sales Support office in Germany. He has 20 years of experience in zEnterprise, System z, zSeries, and predecessor mainframe servers. He has worked for 10 years for the IBM education center in Germany, developing and providing professional training. He also provides professional services to System z and mainframe clients. In the years 2008 and 2009 he supported clients in Middle East / North Africa (MENA) as a zIT Architect. Besides co-authoring several Redbooks publications since 1999, he has been an ITSO guest speaker on ITSO workshops for the last two years. xx
IBM zEnterprise EC12 Technical Guide
8049pref.fm
Karan Singh is a Project Leader at International Technical Support Organization, Global Content Services, Poughkeepsie Center. Andr Spahni is a Senior System Service Representative working for IBM Global Technology Services in Switzerland. He has 10 years of experience working with and supporting System z customers. Andr has been working for the Technical Support Competence Center (TSCC) Hardware FE System z for Switzerland, Germany, and Austria since 2008. His areas of xpertise include System z hardware, Parallel Sysplex, and connectivity. Esra Ufacik is a System z Client Technical Specialist in Systems and Technology Group. She holds a B.Sc. degree in Electronics and Telecommunication Engineering from Istanbul Technical University. Her IT career started with a Turkish Bank as a z/OS systems programmer. Her responsibilities included maintaining a parallel sysplex environment with DB2 data sharing; planning and executing hardware migrations; installing new operating system releases, middleware releases, and system software maintenance; and generating performance reports for capacity planning. Esra joined IBM in 2007 as a Software Support Specialist in Integrated Technology Services, where she work with mainframe customers. Acting as their account advocate, Esra was assigned to a Software Support Team Leader position in ITS and became involved in several projects. Since 2010 she has been with STG, where her role covers pre-sales technical consultancy, conducting System z competitive assessment studies, presenting the value of System z to various audiences, and assuring technical feasibility of proposed solutions. Esra is also a guest lecturer for "System z and large scale computing" classes, which are given to undergraduate Computer Engineering students within the scope of Academic Initiative. Hans Wijngaard is a Client Technical Specialist, who has been with IBM for over 26 years. He started as a hardware Customer Engineer, but worked most of the time as a z/OS systems programmer at IBM client sites. For more than 20 years he gained a broad experience in z/OS performance and tuning, as well as installing, migrating, implementing, and customzing a variety of both IBM- and non-IBM software. The past four years Hans has been working in the System z hardware area, as a Client Technical Specialist, and he is currently supporting IBM Sales Representatives, IBM Business Partners, and clients in various System z business engagements. Wolfgang Fries is a Senior Consultant in the System z HW Support Center in Germany. He spent several years at the European Support Center in Montpellier, France, providing international support for System z servers. Wolfgang has 35 years of experience in supporting large System z customers. His area of expertise include System z servers and connectivity. Zhaoxu Zhang is a Senior System Service Representative at the IBM Global Technology Services in Beijing, China. He joined IBM in 1998 to support and maintain System z products for clients throughout China. Zhaoxu has been working in the Technical Support Group (TSG) providing second level support to System z clients since 2006. His areas of expertise include System z hardware, Parallel Sysplex, and FICON connectivity. Thanks to the following people for their contributions to this project: This project has been coordinated by Bill White, ITSO Global Content Services System z Portfolio Manager, Poughkeepsie Center.
Preface
xxi
8049pref.fm
Comments welcome
Your comments are important to us! We want our books to be as helpful as possible. Send us your comments about this book or other IBM Redbooks publications in one of the following ways: Use the online Contact us review Redbooks form found at: ibm.com/redbooks Send your comments in an email to: redbooks@us.ibm.com Mail your comments to: IBM Corporation, International Technical Support Organization Dept. HYTD Mail Station P099 2455 South Road Poughkeepsie, NY 12601-5400
xxii
8049ch01.fm
Chapter 1.
8049ch01.fm
zEnterprise EC12
Hypervisors
Energy
Operations
Performance
Networks
Virtual Servers
Workloads continue to change. Multi-tier application architectures and their deployment on heterogeneous infrastructures are common today. But what is uncommon is the infrastructure setup needed to provide the high qualities of service required by mission critical applications 2
IBM zEnterprise EC12 Technical Guide
8049ch01.fm
Creating and maintaining these high-level qualities of service while using a large collection of distributed components takes a great amount of knowledge and effort. It implies acquiring and installing extra equipment and software to ensure availability and security, monitoring, and managing. Additional manpower is required to configure, administer, troubleshoot, and tune such a complex set of separate and diverse environments. Due to platform functional differences, the resulting infrastructure will not be uniform, regarding those qualities of service or serviceability. The zEC12 lies on an evolutionary path that directly addresses those infrastructure problems. The zEC12 design can simultaneously support a large number of diverse workloads while providing the highest qualities of service. The IBM holistic approach to System z design includes hardware, software, and procedures. It takes into account a wide range of factors, including compatibility and investment protection, thus ensuring a tighter fit with the IT requirements of the entire enterprise.
8049ch01.fm
The zEC12 expands the subcapacity settings offer with three subcapacity levels for the first 20 processors, giving a total of 161 distinct capacity settings in the system, and providing for a range of over 1:320 in processing power. The zEC12 delivers scalability and granularity to meet the needs of medium-sized enterprises, while also satisfying the requirements of large enterprises having very demanding, mission-critical transaction and data processing requirements. The zEC12 continues to offer all the specialty engines available on previous System z systems.
Virtualization
Processor Resource/Systems Manager (PR/SM) manages all the installed and enabled resources (processors and memory) as a single large SMP system. It enables the configuration and operation of up to 60 logical partitions, which have processors, memory, and I/O resources assigned from the installed books. zEC12 provides improvements to the PR/SM HiperDispatch function. HiperDispatch provides work alignment to logical processors, and alignment of logical processors to physical processors. This alignment optimizes cache utilization, minimizes inter-book communication, and optimizes z/OS work dispatching, with the end result of increasing throughput. zEC12 provides for the definition of up to 32 HiperSockets. HiperSockets provide for memory communication across logical partitions, without the need of any I/O adapters, with VLAN capability. HiperSockets have been extended to bridge to an ensemble internode data network.
8049ch01.fm
8049ch01.fm
IBM System z Advanced Workload Analysis Reporter (IBM zAware) is a feature introduced with the zEC12 which embodies the next generation of system monitoring. IBM zAware is designed to offer a near real-time, continuous learning, diagnostics and monitoring capability, to help pinpoint and resolve potential problems quickly enough to minimize business impacts. The ability to tolerate service disruptions is diminishing. In a 24 x 7 environment any disruption can have grave consequences, special when it last days, even hours. But increased system complexity makes it more probable that errors occur, and those errors are, too, increasingly complex. Some incidents early symptoms go undetected for long periods of time and can grow to large problems. Systems often experience soft failures (sick but not dead) which are much more difficult or unusual to detect. IBM zAware is designed to help in those circumstances. For more information see Appendix A, IBM zAware on page 429.
1.3.1 Models
The zEC12 has a machine type of 2827. Five models are offered: H20, H43, H66, H89, and HA1. The model name indicates the maximum number of PUs available for purchase (withA1 standing for 101). A PU is the generic term for the z/Architecture processor on the Multi-Chip Module (MCM) that can be characterized as any of the following items: Central Processor (CP). Internal Coupling Facility (ICF) to be used by the Coupling Facility Control Code (CFCC). Integrated Facility for Linux (IFL) Additional System Assist Processor (SAP) to be used by the channel subsystem. System z Application Assist Processor (zAAP). One CP must be installed with or prior to the installation of any zAAPs. System z Integrated Information Processor (zIIP). One CP must be installed with or prior to installation of any zIIPs. In the five-model structure, only one CP, ICF, or IFL must be purchased and activated for any model. PUs can be purchased in single PU increments and are orderable by feature code. The total number of PUs purchased cannot exceed the total number available for that model. The number of installed zAAPs cannot exceed the number of installed CPs. The number of installed zIIPs cannot exceed the number of installed CPs. The multi-book system design provides an opportunity to concurrently increase the capacity of the system in three ways: Add capacity by concurrently activating more CPs, IFLs, ICFs, zAAPs, or zIIPs on an existing book. Add a new book concurrently and activate more CPs, IFLs, ICFs, zAAPs, or zIIPs. Add a new book to provide additional memory or one or more adapters to support a greater number of I/O features.
8049ch01.fm
upgrade). Any z196 or z10 EC model can be upgraded to any zEC12 model. Figure 1-2 presents a diagram of the upgrade path. Note: An air-cooled zEC12 cannot be converted to an water cooled zEC12 or vice-versa.
Not offered
Downgrades within the zEC12 models Upgrade from a z114 with zBX to zEC12 Removal of a zBX without a controlling system Upgrades from IBM System z9 or earlier systems zBX Model 002 attachment to zEC12
z196 zBX Model 002
z10 EC
zEC12
1.3.3 Frames
The zEC12 has two frames, bolted together, and known as the A frame and the Z frame. The frames contain the CPC components, including these: The processor cage, with up to four books PCIe I/O drawers, I/O drawers, and I/O cage Power supplies An optional internal battery feature (IBF) Cooling units for either air or water cooling Support elements
8049ch01.fm
MCM technology
The zEC12 is built on a proven superscalar microprocessor architecture of its predecessor and provides several enhancements over the z196. In each book there is one MCM. The MCM has six PU chips and two SC chips. The PU chip has six cores, with four, five or six active cores, which can be characterized as CPs, IFLs, ICFs, zIIPs, zAAPs, or SAPs. Two MCM sizes are offered, which are 27 or 30 cores. The MCM provides a significant increase in system scalability and an additional opportunity for server consolidation. All books are interconnected with very high-speed internal communication links, in a full star topology through the L4 cache, which allows the system to be operated and controlled by the PR/SM facility as a memory- and cache-coherent symmetric multiprocessor (SMP). The PU configuration is made up of two spare PUs per CPC and a variable number of system assist processors (SAPs), which scale with the number of books installed in the server, such as three SAPs with one book installed and up to 16 when four books are installed. In addition, one PU is reserved and is not available for customer use. The remaining PUs can be characterized as central processors (CPs), Integrated Facility for Linux (IFL) processors, System z Application Assist Processors (zAAPs), System z Integrated Information Processors (zIIPs), Internal Coupling Facility (ICF) processors, or additional SAPs. The zEC12 offers a water cooling option for increased system and data center energy efficiency. In a a water cooled system, the MCM is cooled by a cold plate connected to the internal water cooling loop. In a air-cooled system, radiator units (RU) exchange the heat from internal water loop with air, and air backup. Both cooling options are fully redundant.
Processor features
The processor chip has a six-core design, with either four, five or six active cores, and operates at 5.5 GHz. Depending on the MCM version (27 PU or 30 PU), from 27 to 120 PUs are available, on one to four books. Each core on the PU chip includes a dedicated coprocessor for data compression and cryptographic functions, such as the CP Assist for Cryptographic Function (CPACF). This is an improvement over z196, where two cores shared a coprocessor. Hardware data compression can play a significant role in improving performance and saving costs over doing compression in software. Having standard clear key cryptographic coprocessors integrated with the processor provides high-speed cryptography for protecting data. Each core has its own hardware decimal floating point unit designed according to a standardized, open algorithm. Much of today's commercial computing is decimal floating point, so on-core hardware decimal floating point meets the requirements of business and user applications, and provides improved performance, precision, and function. In the unlikely case of a permanent core failure, each core can be individually replaced by one of the available spares. Core sparing is transparent to the operating system and applications.
8049ch01.fm
scalable multi-threaded execution, and is known in the academia as hardware transactional memory.
Out-of-order execution
The zEC12 has a superscalar microprocessor with out-of-order (OOO) execution to achieve faster throughput. With OOO, instructions might not execute in the original program order, although results are presented in the original order. For instance, OOO allows a few instructions to complete while another instruction is waiting. Per machine cycle, up to three instructions can be decoded and up to seven instructions can be in execution.
8049ch01.fm
The zEC12 takes advantage of PCIe Generation 2 to implement the following features: An I/O bus, which implements the PCIe infrastructure. This is a preferred infrastructure and can be used alongside InfiniBand. PCIe fanouts, which provide 8 GBps connections to the PCIe I/O features. The zEC12 takes advantage of InfiniBand to implement the following features: An I/O bus, which includes the InfiniBand infrastructure. This replaces the self-timed interconnect bus found in System z systems prior to z9. Parallel Sysplex coupling links using InfiniBand (IFB): 12x InfiniBand coupling links for local connections and 1x InfiniBand coupling links for extended distance connections between any two of: zEnterprise CPCs, and z10 CPCs. The 12x IB link has a bandwidth of 6 GBps. Host Channel Adapters for InfiniBand (HCA3) can deliver up to 40% faster coupling link service times than HCA2.
I/O drawer
On the zEC12 I/O drawers are only supported when carried forward on upgrades from z196 or z10. The zEC12 can have up to two I/O drawers in the Z frame. I/O drawers can accommodate up to eight I/O features in any combination. Based on the amount of legacy I/O features carried forward, the configurator will determine the number of required I/O drawers.
I/O cage
On the zEC12 a maximum of one I/O cage is supported and is only available on upgrades from z196 or z10 to zEC12. The I/O cage in housed in the A frame. The I/O cage can accommodate up to 28 I/O features in any combination. Based on the amount of legacy I/O features carried forward, the configurator will determine the required number of I/O drawers and I/O cages.
10
8049ch01.fm
I/O features
The zEC12 supports the following PCIe features, which can be installed only in the PCIe I/O drawers: FICON Express8S SX and 10 KM LX (Fibre Channel connection) OSA-Express4S 10 GbE LR and SR, GbE LX and SX, 1000BASE-T Crypto Express4S Flash Express When carried-forward on an upgrade, the zEC12 also supports up to one I/O cage and up to two I/O drawers on which the following I/O features can be installed: FICON Express8 FICON Express4 10 KM LX and SX (four port cards only) OSA-Express3 10 GbE LR and SR OSA-Express3 GbE LX and SX OSA-Express3 1000BASE-T Crypto Express3 ISC-3 coupling links (peer-mode only) In addition, InfiniBand coupling links, which attach directly to the processor books, are supported.
FICON channels
Up to 160 features with up to 320 FICON Express8S channels are supported. The FICON Express8S features support a link data rate of 2, 4, or 8 Gbps. Up to 44 features with up to 176 FICON Express8 or FICON Express4 channels are supported: The FICON Express8 features support a link data rate of 2, 4, or 8 Gbps. The FICON Express4 features support a link data rate of 1, 2, or 4 Gbps. The zEC12 FICON features support the following protocols: FICON (FC) and High Performance FICON for System z (zHPF). zHPF offers improved access to data, of special importance to OLTP applications Channel-to-channel (CTC) Fibre Channel Protocol (FCP). FICON also offers the following capabilities: Modified Indirect Data Address Word (MIDAW) facility: provides more capacity over native FICON channels for programs that process data sets, which exploit striping and compression (such as DB2, VSAM, PDSE, HFS, and zFS) by reducing channel, director, and control unit overhead. Enhanced problem determination, analysis, and manageability of the storage area network (SAN) by providing registration information to the fabric name server for both FICON and FCP
11
8049ch01.fm
The maximum number of combined OSA Express3 and OSA Expres4S features can not exceed 48.
HiperSockets
The HiperSockets function, also known as internal queued direct input/output (internal QDIO or iQDIO), is an integrated function of the zEC12 that provides users with attachments to up to 32 high-speed virtual LANs with minimal system and network overhead. HiperSockets can be customized to accommodate varying traffic sizes. Because HiperSockets does not use an external network, it can free up system and network resources, eliminating attachment costs while improving availability and performance. For communications between logical partitions in the same zEC12 server, HiperSockets eliminates the need to use I/O subsystem features and to traverse an external network. Connection to HiperSockets offers significant value in server consolidation by connecting many virtual servers, and can be used instead of certain coupling link configurations in a Parallel Sysplex.
1.3.7 Cryptography
Integrated cryptographic features provide leading cryptographic performance and functionality. Reliability, availability, and serviceability (RAS) support is unmatched in the industry and the cryptographic solution has received the highest standardized security certification (FIPS 140-2 Level 42). The crypto cards were enhanced with additional
2
Federal Information Processing Standards (FIPS)140-2 Security Requirements for Cryptographic Modules
12
8049ch01.fm
capabilities to dynamically add or move crypto coprocessors to logical partitions without pre-planning. The zEC12 implements the PKCS #11, one of the industry-accepted standards called Public Key Cryptographic Standards (PKCS) provided by RSA Laboratories of RSA Security Inc., and the IBM Common Cryptographic Architecture (CCA) in its cryptographic features.
13
8049ch01.fm
Crypto Express3 Coprocessor is for secure key encrypted transactions (default). Crypto Express3 Accelerator is for Secure Sockets Layer/Transport Layer Security (SSL/TLS) acceleration.
14
8049ch01.fm
Support for an optional Smart Card Reader attached to the TKE workstation allows for the use of smart cards that contain an embedded microprocessor and associated memory for data storage. Access to and the use of confidential data on the smart cards is protected by a user-defined personal identification number (PIN). When Crypto Express4S is configured as a Secure IBM Enterprise PKCS #11 (EP11) coprocessor the TKE workstation is required to manage the Crypto Express4S feature. If the smart card reader feature is installed in the TKE workstation, the new smart card part 74Y0551 is required for EP11 mode.
15
8049ch01.fm
Statement Of Direction: Removal of ISC-3 support on System z: The IBM zEnterprise EC12 is planned to be the last high-end System z server to offer support of the InterSystem Channel-3 (ISC-3) for Parallel Sysplex environments at extended distances. ISC-3 will not be supported on future high-end System z servers as carry forward on an upgrade. Previously we announced that the IBM zEnterprise 196 (z196) and IBM zEnterprise 114 (z114) servers were the last to offer ordering of ISC-3. Enterprises should continue migrating from ISC-3 features (#0217, #0218, #0219), to 12x InfiniBand (#0171 - HCA3-O fanout) or 1x InfiniBand (#0170 - HCA3-O LR fanout) coupling links.
16
8049ch01.fm
and so on). NTP can only be used as ETS for an STP-only CTN where no server can have an active connection to a Sysplex Timer. The time accuracy of an STP-only CTN is further improved by adding an NTP server with the pulse per second output signal (PPS) as the ETS device. This type of ETS is available from various vendors that offer network timing solutions. Improved security can be obtained by providing NTP server support on the Hardware Management Console ((HMC) for the Support Element (SE), as the HMC is normally attached to the private dedicated LAN for System z maintenance and support. For zEC12, authentication support is added to the HMCs NTP communication with NTP time servers. A zEC12 cannot be connected to a Sysplex Timer. It is preferable to migrate to an STP-only Coordinated Time Network (CTN) for existing environments. It is possible to have a zEC12 as a Stratum 2 or Stratum 3 server in a Mixed CTN, as long as there are at least two System z10s attached to the Sysplex Timer operating as Stratum 1 servers.
The maximum number of blades varies according to the blade type and blade function.
17
8049ch01.fm
1.4.1 Blades
There are two types of blades that can be installed and operated in the IBM zEnterprise BladeCenter Extension (zBX): Optimizer Blades: IBM WebSphere DataPower Integration Appliance XI50 for zEnterprise blades. IBM Blades: A selected subset of IBM POWER7 blades A selected subset of IBM BladeCenter HX5 blades These blades have been thoroughly tested to ensure compatibility and manageability in the IBM zEnterprise System environment: IBM POWER7 blades are virtualized by PowerVM Enterprise Edition, and the virtual servers run the AIX operating system. IBM BladeCenter HX5 blades are virtualized using an integrated hypervisor for System x and the virtual servers run Linux on System x (Red Hat Enterprise Linux - RHEL and SUSE Linux Enterprise Server - SLES) operating systems. Enablement for the blades is specified with an entitlement feature code to be configured on the zEC12s.
18
8049ch01.fm
5. Network management (Networks): Creates and manages virtual networks, including access control, which allows virtual servers to be connected together. 6. Workload Awareness and platform performance management (Performance): Management of CPU resource across virtual servers hosted in the same hypervisor instance to achieve workload performance policy objectives. The Unified Resource Manager provides energy monitoring and management, goal-oriented policy management, increased security, virtual networking, and storage configuration management for the physical and logical resources of a given ensemble.
6 7 8 9
From Version 2.11, see12.6, HMC in an ensemble on page 422. The z/OS V1 Release 10 will require IBM Lifecycle Extension after September 2011. SLES is the abbreviation for SUSE Linux Enterprise Server. RHEL is the abbreviation for Red Hat Enterprise Linux.
19
8049ch01.fm
Windows Server 2008 R2 and Windows Server 2008 SP2 (Datacenter Edition recommended), 64-bit only With support for IBM WebSphere software, full support for SOA, Web services, J2EE, Linux, and Open Standards, the zEC12 is intended to be a platform of choice for integration of a new generation of applications with existing applications and data.
IBM compilers
The latest versions of the compilers are required in order to exploit the new and enhanced instructions introduced with each system. Enterprise COBOL, Enterprise PL/I, and XL C/C++ are leading-edge, z/OS-based compilers that maximize middleware by providing access to IBM DB2, CICS, and IMS systems, and allow integration with web services, XML, and Java. Such interoperability enables you to capitalize on existing IT investments while smoothly incorporating new, web-based applications into your organizations infrastructure. z/OS XL C/C++ helps you to create and maintain critical business applications written in C or C++ to maximize application performance and improve developer productivity. z/OS XL C/C++ can transform C or C++ source code to fully exploit System z hardware (including the zEC12), through hardware-tailored optimizations, built-in functions, performance-tuned libraries, and language constructs that simplify system programming and boost application runtime performance.
8049ch01.fm
Enhanced book availability Hot pluggable PCIe I/O drawers and I/O drawers Redundant I/O interconnect Concurrent PCIe fanout and Host Channel Adapter (HCA-O, HCA-C) fanout card hot-plug Enhanced driver maintenance For a more detailed description, see Chapter 10, RAS on page 355
1.9 Performance
The zEC12 Model HA1 is designed to offer approximately 1.5 times more capacity than the z196 Model M80 system. Uniprocessor performance has also increased significantly. A zEC12 Model 701 offers, on average, performance improvements of about 1.25 times over the z196 Model 701. Figure 1-3 shows the estimated capacity ratios for zEC12, z196, z10 EC, and z9 EC. Notice that LSPR numbers given for z9 EC, z10 EC and z196 systems were obtained with the z/OS 1.11 operating system, and with z/OS 1.13 for the zEC12 system.
z9 EC z10 EC z196 zEC12
1) 2)
up to 54-way: 193 to 18,505 PCIs 1 up to 64-way: 214 to 31,826 PCIs 1 up to 80-way: 240 to 52,286 PCIs 1 up to 101-way: 240 to 78,500 PCIs 2
LSPR data based on z/OS 1.11 LSPR data based on z/OS 1.13
zEC12
On average, the zEC12 can deliver up to 50% more performance in an 101-way configuration than an IBM System z196 80-way. However, variations on the observed performance increase are dependent upon the workload type. Consult the Large System Performance Reference (LSPR) when you consider performance on the zEC12. The range of performance ratings across the individual LSPR workloads is likely to have a large spread. More performance variation of individual logical partitions exists because the impact of fluctuating resource requirements of other partitions can be more pronounced with the increased numbers of partitions and additional PUs available. For more information, see 1.9.6, Workload performance variation on page 28.
21
8049ch01.fm
For detailed performance information, see the LSPR web site: https://www-304.ibm.com/servers/resourcelink/lib03060.nsf/pages/lsprindex The MSU ratings are available from the following web site: http://www-03.ibm.com/systems/z/resources/swprice/reference/exhibits/
Traditional online transaction processing workload (formerly known as IMS) Commercial batch with long-running jobs Low I/O Content Mix Workload Transaction Intensive Mix Workload
22
8049ch01.fm
Instruction complexity
The type of instructions and the sequence in which they are executed will interact with the design of a microprocessor to affect a performance component we can define as instruction complexity. There are many design alternatives that affect this component, such as these: Cycle time (GHz) Instruction architecture Pipeline Superscalar Out-of-order execution Branch prediction As workloads are moved between microprocessors with various designs, performance will likely vary. However, when on a processor, this component tends to be quite similar across all models of that processor.
The Nest
Memory L4 Cache
L3 Cache
L2
L1
Memory L4 Cache
L3 Cache L3 Cache
L2
L1
...
L2
L1
...
L2
L1
L3 Cache
L2
L1
CPU1
....
L2
L1
CPU6
CPU1
....
L2
L1
CPU6
CPU1
....
CPU6
CPU1
....
L2
L1
CPU6
Book 1
Figure 1-4 Memory hierarchy on the zEC12 2-book system
Book 2
23
8049ch01.fm
Workload capacity performance will be quite sensitive to how deep into the memory hierarchy the processor must go to retrieve the workloads instructions and data for execution. Best performance occurs when the instructions and data are found in the cache(s) nearest the processor so that little time is spent waiting prior to execution; as instructions and data must be retrieved from farther out in the hierarchy, the processor spends more time waiting for their arrival. As workloads are moved between processors with various memory hierarchy designs, performance will vary as the average time to retrieve instructions and data from within the memory hierarchy will vary. Additionally, when on a processor, this component will continue to vary significantly, because the location of a workloads instructions and data within the memory hierarchy is affected by many factors including, but not limited to, these: Locality of reference I/O rate Competition from other applications and/or LPARs.
24
8049ch01.fm
Low
Batch Low Single Intensive Low High locality Simple Extensive
High
Transactional High Many Light High Diverse Complex Limited
Figure 1-5 The traditional factors that have been used to categorize workloads.
Note that one can do little to affect most of these factors. An application type is whatever is necessary to do the job. Data reference pattern and CPU usage tend to be inherent to the nature of the application. LPAR configuration and application mix are mostly a function of what needs to be supported on a system. I/O rate can be influenced somewhat through buffer pool tuning. However, one factor that can be affected, software configuration tuning, is often overlooked but can have a direct impact on RNI. Here we refer to the number of address spaces (such as CICS AORs or batch initiators) that are needed to support a workload. This factor has always existed but its sensitivity is higher with todays high frequency microprocessors. Spreading the same workload over a larger number of address spaces than necessary can raise a workloads RNI as the working set of instructions and data from each address space increases the competition for the processor caches. Tuning to reduce the number of simultaneously active address spaces to the proper number needed to support a workload can reduce RNI and improve performance. In the LSPR, the number of address spaces for each processor type and Nway configuration is tuned to be consistent with what is needed to support the workload. Thus, the LSPR workload capacity ratios reflect a presumed level of software configuration tuning. This suggests that re-tuning the software configuration of a production workload as it moves to a bigger or faster processor might be needed in order to achieve the published LSPR ratios.
25
8049ch01.fm
A workload category representing average use of the memory hierarchy. This is similar to the past LoIO-mix workload and is expected to represent the majority of production workloads.
14
IBM Processor Capacity Reference: A no-cost tool that reflects the latest IBM LSPR measurements, and is available at http://www-03.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/PRS1381
26
8049ch01.fm
Low
High
LSPR Primitives
CB-L
WASDB
OLTP-T
OLTP-W
LSPR Mixes
LoIO-Mix
TM-Mix
TI-Mix DI-Mix
CB-Mix
TD-Mix
LSPR Categories
Low
LowAverage
Average
AverageHigh
High
However, as discussed in the LSPR Workload Categories section, the underlying performance sensitive factor is how a workload interacts with the processor hardware. These past techniques were simply trying to approximate the hardware characteristics that were not available through software performance reporting tools. Beginning with the z10 processor, the hardware characteristics can now be measured using CPU MF (SMF 113) counters data. To reflect the memory hierarchy changes in the new zEC12 system, the number of counters has been increased to 183. Thus, there is an opportunity to be able to match a production workload to an LSPR workload category through these hardware characteristics (see the LSPR Workload Categories section for a discussion about RNI Relative Nest Intensity). The AVERAGE RNI LSPR workload is intended to match the majority of customer workloads. When no other data is available, use it for a capacity analysis. DASD I/O rate has been used for many years to separate workloads into two categories: those whose DASD I/O per MSU (adjusted) is <30 (or DASD I/O per PCI <5) and those higher than these values. The majority of production workloads fell into the low I/O category and a LoIO-mix workload was used to represent them. Using the same I/O test, these workloads now use the AVERAGE RNI LSPR workload. Workloads with higher I/O rates can use the HIGH RNI workload or the AVG-HIGH RNI workload that is included with zPCR. Low-Average and Average-High categories allow better granularity for workload characterization. For z10 and newer processors, the CPU MF data can be used to provide an additional hint as to workload selection. When available, this data allows the RNI for a production workload to be calculated. Using the RNI and another factor from CPU MF, the L1MP (percentage of data and instruction references that miss the L1 cache), a workload can be classified as LOW, AVERAGE or HIGH RNI. This classification and resulting hint is automated in the zPCR tool. It is best to use zPCR for capacity sizing. The LSPR workloads, updated for EC12 are considered to reasonably reflect current and growth workloads of the customer. The set contains three generic workload categories based on z/OS 1.13 supporting up to 101 processors in a single image.
27
8049ch01.fm
8049ch01.fm
Improve fetch and store conflict scheme Enhance branch prediction structure and sequential instruction fetching Millicode performance improvements Optimized floating-point performance Faster engine for fixed-point division New second level branch prediction array One cryptographic/compression co-processor per core Cryptography support of UTF8<>UTF16 conversions Higher clock frequency at 5.5 GHz IBM CMOS 13S 32nm SOI technology with IBM eDRAM technology.
29
8049ch01.fm
30
8049ch02.fm
Chapter 2.
31
8049ch02.fm
Processor Books with Memory, HCAand PCIe-Fanout cards InfiniBand and PCIe I/O Interconnects
Support Elements
N+1 Radiators
Figure 2-1 CPC cage, I/O drawers, and I/O cage locations - air-cooled system - front view
32
8049ch02.fm
Power Supplies Processor Books with Memory, HCAand PCIe-Fanout cards Support Elements InfiniBand and PCIe I/O Interconnects I/O cage Carried Forward PCIe I/O drawer N+1 Water Cooling Units
Figure 2-2 CPC cage, I/O drawers, and I/O cage locations - water cooled system - front view
2.1.1 Frame A
As shown in Figure 2-1 on page 32 and Figure 2-2, frame A has these main components: Two optional Internal Battery Features (IBFs), which provide the function of a local uninterrupted power source. The IBF further enhances the robustness of the power design, increasing power line disturbance immunity. It provides battery power to preserve processor data in case of a loss of power on all connected AC or DC feeds from the utility provider. The IBF provides battery power to preserve full system function despite the loss of power at all system power cords. It allows continuous operation through intermittent losses, brownouts, and power source switching, or can provide time for an orderly shutdown in case of a longer outage. The IBF provides up to 10 minutes of full power, depending on the I/O configuration. The batteries are installed in pairs. Two to six battery units can be installed. The number is based on the zEC12 model and configuration. Two, fully redundant radiator units, containing water pumps to feeding the internal closed water loops for MCM cooling, heat exchanger, manifold assembly and blowers to provide the cooling of the MCMs. Instead of the radiator units, the customer can specify two Water Conditioning Units (WCUs) connected to a customer chilled water supply. When the WCUs option is used for cooling the books an additional exhaust air heat exchanger is installed in the rear of the frame. Processor cage, which contains up to four books, connected to the internal water cooling system.
Chapter 2. CPC Hardware Components
33
8049ch02.fm
Depends on the configuration the following I/O assembly can be found (where any combination of up to two drawer or the I/O cage is possible): up to two PCI/e I/O drawer for installation of new form factor cards for FICON8S, OSA-Express4S, Crypto Express4S and Flash Express. The PCIe I/O drawer is used for new installation or can be carried forward from z196 MES and is equipped with a maximum of 32 features. one I/O drawer each containing up to eight I/O cards of any type of FICON Express8, FICON Express4, OSA- Express3, ISC-3 and Crypto Express3 features. The I/O drawer itself and all the current installed features can only be carry forwarded with an MES from a z10 or a z196. one I/O cage, which can house 28 I/O card slots for installation of FICON Express8, FICON Express4, OSA- Express3, ISC-3 and Crypto Express3 features. One I/O cages is supported, the I/O cage itself and all the current installed features can only carried forward with an MES from a z10 or a z196. Air-moving devices (AMD), which provide N+1 redundant cooling for the fanouts, memory, and DCAs.
2.1.2 Frame Z
As shown in Figure 2-1 on page 32, frame Z has these main components: Two optional Internal Battery Features (IBFs). Bulk Power Assemblies (BPAs). The number of BPAs vary and depend on the configuration of the zEC12. Detailed information about the required number of BPAs can be found in 2.9.1, Power consumption on page 68 The Support Element (SE) tray, located in front of I/O drawer slots, contains the two SEs. Up two 4 drawers, which can be all PCIe I/O drawer or a combination of up to two I/O drawer and PCIe drawer. I/O cage is not supported in Frame Z. When the WCUs option is used for cooling the books an additional exhaust air heat exchanger is installed in the rear of the frame. An optional overhead power cable feature (shown in Figure 2-1 on page 32) and an also optional top exit I/O cabling feature which is, when ordered present at frame A and frame Z.
8049ch02.fm
Crypto Express4S. Each Crypto Express4S feature holds one PCI Express cryptographic adapter. Flash Express. Each Flash Express feature occupies two I/O slots but does not have a CHPID type. Logical partitions in all CSSs have access to the features. InfiniBand I/O infrastructure uses the HCA2-C fanout to connect to I/O drawer(s) or I/O cage(s) that can contain a variety of channel, Coupling Link, OSA-Express, and Cryptographic feature cards: FICON channels (FICON or FCP modes) FICON Express4 channels (four port cards) FICON Express8 channels (four port cards) ISC-3 links (up to four coupling links, two links per daughter card). Two daughter cards (ISC-D) plug into one mother card (ISC-M). OSA-Express channels: OSA-Express3 10 Gb Ethernet (two ports cards, LR or SR, two CHPIDs) OSA-Express3 Gb Ethernet (four port cards, LX and SX, two CHPIDs) OSA-Express3 1000BASE-T Ethernet (four port cards, two CHPIDs) OSA-Express2 Gb Ethernet (two port cards, SX, LX, two CHPIDs) OSA-Express2 1000BASE-T Ethernet (two port card, two CHPIDs) Crypto Express3 feature (FC 0864) has two PCI Express adapters per feature. A PCI Express adapter can be configured as a cryptographic coprocessor for secure key operations or as an accelerator for clear key operations. InfiniBand coupling to a coupling facility is achieved directly from the HCA2-O (12xIFB) fanout and HCA3-O (12xIFB) fanout to the coupling facility with a bandwidth of 6 GBps. The HCA2-O LR (1xIFB) fanout and HCA3-O LR (1xIFB) fanout support long distance coupling links for up to 10 km or 100 km when extended by using System z qualified DWDM1 equipment. Supported bandwidths are 5 Gbps and 2.5 Gbps, depending on the DWDM equipment used.
35
8049ch02.fm
Rear
16 D I MM s 100m m H ig h M C M @ 1800W inte rn al system wa ter cool ed
F ront
M MC
Memo ry
M emory
14 D IM Ms 10 0m m H i gh
Each book contains the following components: One multi-chip module (MCM) with six hex-core microprocessor chips, having either 27 or 30 processor units PUs), depending on the model, and two storage control chips with 384 MB of Level 4 cache. Memory DIMMs plugged into 30 available slots, providing from 60 GB to 960 GB of physical memory installed in a book. A combination of up to eight (host channel adapter (HCA) or PCIe) fanout cards. HCA2-Copper connections are for 6GBps links to the I/O cage or I/O drawers in the CPC. PCIe fanouts are used for 8GBps links to the PCIe I/O drawers, and the HCA-Optical fanouts connect to external CPCs (coupling links). Three distributed converter assemblies (DCAs) that provide power to the book. Loss of a DCA leaves enough book power to satisfy the books power requirements (n+1 redundancy). The DCAs can be concurrently maintained. Two flexible service processor (FSP) cards for system control. Figure 2-4 depicts the book logical structure, showing its component connections, including the PUs on MCM.
36
8049ch02.fm
Mem2
Mem1
Mem0
MCU
MCU
MCU
PU
PU
PU
FBC
FBC
SC
SC
PSI
PU
PU
PU
GX
Logical Node
Memory is connected to MCM through three memory control units (MCUs). GX0 to GX7 are the I/O buses interfaces to HCAs, with full store buffering, maximum of 10 GB/s per bus direction, and support for Infiniband and PCIe. Processor support interfaces (PSIs) are used to communicate with FSP cards for system control. Fabric book connectivity (FBC) provides the point-to-point connectivity between books.
Book
FBC FBC
FBC
Book
FBC FBC
Book
FBC
Book
Up to four books can reside in the processor cage. Books slide into a mid-plane card that supports up to four books and is located in the top of frame A. The mid-plane card is also the location of two oscillator cards.
Chapter 2. CPC Hardware Components
37
8049ch02.fm
The location of books is as follows: In a one-book model, the first book slides in the second slot from the left (processor cage slot location LG06). In a two-book model, the second book slides in the right-most slot (processor cage slot location LG15). In a three-book model, the third book slides in the third slot from the left (processor cage slot location LG10). In a four-book model, the fourth book slides into the left-most slot (processor cage slot location LG01). Table 2-1 indicates the order of book installation and position in the processor cage.
Table 2-1 Book installation order and position in the processor cage Book Installation order Position in cage (LG) Book0 Fourth 01 Book1 First 06 Book2 Third 10 Book3 Second 15
Book installation is concurrent, except for the upgrade to the model HA1. Concurrent book replacement requires a minimum of two books. Tip: The processor cage slot locations are important in the sense that in the physical channel ID (PCHID) report, resulting from the IBM configurator tool, locations 01, 06, 10, and 15 are used to indicate whether book features such as fanouts and AID assignments relate to the first, second, third, or fourth book in the processor cage.
Dual external clock facility: Two external clock facility (ECF) cards are already installed and shipped with the CPC and provide a dual-path interface for pulse per second (PPS). This redundancy allows continued operation even if a single ECF card fails.
This redundant design also allows concurrent maintenance. The two connectors that connect to the PPS output of an NTP server are located above the books and are connected on the mid-plane to which the books are connected. The support element provides a simple network time protocol (SNTP) client. When server time protocol (STP) is used, the time of an STP-only coordinated timing network (CTN) can be synchronized with the time provided by a network time protocol (NTP) server, allowing a heterogeneous platform environment to have the same time source. The time accuracy of an STP-only CTN is improved by adding an NTP server with the pulse per second output signal (PPS) as the external time source (ETS) device. NTP server devices with pulse per second output (PPS) are available from several vendors that offer network timing solutions. A cable connection from the PPS port on the ECF card to the PPS output of the NTP server is required when the zEC12 is using STP and configured in an STP-only CTN using NTP with pulse per second as the external time source. STP tracks the highly stable and accurate PPS signal from the NTP server and maintains an accuracy of 10 s to the ETS, as measured at the PPS input of the IBM zEnterprise EC12. If STP uses an NTP server without PPS, a time accuracy of 100 ms to the ETS is maintained. Figure 2-6 shows the location of the two ECF cards on the CPC, above book 0 and book 3 locations.
38
8049ch02.fm
ECF
OSC
OSC
ECF
Book 0
Book 1
Book 2
Book 3
LG01
LG06
LG10
LG15
STP: Server time protocol (STP) is available as FC 1021. STP is implemented in the Licensed internal code (LIC) and is designed for multiple servers to maintain time synchronization with each other and synchronization to an External Time Source. For more information, see the following publications: Server Time Protocol Planning Guide, SG24-7280 Server Time Protocol Implementation Guide, SG24-7281 Server Time Protocol Recovery Guide, SG24-7380
2.2.2 Oscillator
The zEC12 has two oscillator cards (OSC), a primary and a backup. Although not part of the book design, they are found above the books, connected to the same mid-plane to which the books are connected. If the primary fails, the secondary detects the failure, takes over transparently, and continues to provide the clock signal to the CPC. Figure 2-6 on page 39 shows the location of the two OSC cards on the CPC, above book 1 and book 2 locations.
39
8049ch02.fm
FSP
Bulk Power
FSP
Bulk Power
Primary SE
Ethernet 'hub' 1
Ethernet 'hub' 2
Alternate SE
HMC
FSP External LAN FSP FSP FSP FSP FSP FSP FSP
Book
Book
.....
I/O cage
I/O cage
A typical FSP operation is to control a power supply. An SE sends a command to the FSP to bring up the power supply. The FSP (using SSI connections) cycles the various components of the power supply, monitors the success of each step and the resulting voltages, and reports this status to the SE. Most system elements are duplexed (n+1), and each element has an FSP. Two internal Ethernet LANs and two SEs, for redundancy, and crossover capability between the LANs are available so that both SEs can operate on both LANs. The SEs, in turn, are connected to one or two (external) LANs (Ethernet only), and the Hardware Management Consoles (HMCs) are connected to these external LANs. One or more HMCs can be used, but two (a primary and an alternate2) are mandatory with an ensemble. Additional HMCs can operate a zEC12 when it is not a member of an ensemble.
40
8049ch02.fm
processor unit (PU) chips and two storage control (SC) chips. Figure 2-8 illustrates the chip locations. The total number of transistors on all chips on the MCM is more than 20 billion.
92 mm PU2 PU1 PU0
SC1
SC0
PU3
PU4
PU5
92 mm
The MCM plugs into a card that is part of the book packaging. The book itself is plugged into the mid-plane board to provide interconnectivity between the books, so that a multibook system appears as a symmetric multiprocessor (SMP) system.
41
8049ch02.fm
Memory 2 GX 2 & 6
6 PU c or es 6 L1 c ac he s 6 L2 c ac he s 1 L3 c ache 6 COP MC, GX
Memory 0 PSI GX 0 P SI GX 5
6 PU core s 6 L1 c aches 6 L2 c aches 1 L3 ca che 6 C OP MC, GX
GX 4
6 PU c or es 6 L1 cac he s 6 L2 cac he s 1 L3 c ache 6 COP MC, GX
GX 3
6 PU core s 6 L1 c aches 6 L2 c aches 1 L3 ca che 6 C OP MC, GX
192MB L4 SC
192MB L4 SC
2.4.1 PU chip
The zEC12 PU chip is an evolution of the z196 core design, using CMOS 13S technology, out-of-order instruction processing, higher clock frequency, and redesigned and larger caches. Compute intensive workloads can achieve additional performance improvements through compiler enhancements, and larger caches can improve system performance on many production workloads. Each PU chip has up to six cores running at 5.5 GHz, which means the cycle time is slightly shorter then 0,18 ns. There are six PU chips on each MCM. The PU chips come in three versions, having four, five, or six active cores. For models H20, H43, H66, and H89, the processor units in the MCM in each book are implemented with 27 active cores per MCM. This means that model H20 has 27, model H43 has 54, model H66 has 81, and model H89 has 108 active cores. The model HA1 has six 30 active cores per MCM. This means that there are 120 active cores on model HA1. A schematic representation of the PU chip is shown in Figure 2-10.
42
8049ch02.fm
G X
i/o
CORE
CORE
CORE
M C
i/o
SC i/o
SC i/o
G X G X
i/o
L3 Cache
SC i/o
L3 Control
L3 Cache
SC i/o
M C
CORE
CORE
CORE
M C
i/o
Each PU chip has 2.75 billion transistors. Each one of the six cores has its own L1 cache with 64 KB for instructions and 96 KB for data. Next to each core resides its private L2 cache, with 1 MB for instructions and 1 MB for data respectively. There is one L3 cache, with 48 MB. This 48 MB L3 cache is a store-in shared cache across all cores in the PU chip. It has 192 x 512Kb eDRAM macros, dual address-sliced and dual store pipe support, an integrated on-chip coherency manager, cache and cross-bar switch. The L3 directory filters queries from local L4. Both L3 slices can deliver up to 160 GB/s bandwidth to each core simultaneously. The L3 cache interconnects the six cores, GX I/O buses, and memory controllers (MCs) with storage control (SC) chips. The memory controller (MC) function controls access to memory. The GX I/O bus controls the interface to the host channel adapters (HCAs) accessing the I/O. The chip controls traffic between the cores, memory, I/O, and the L4 cache on the SC chips. There are also one dedicated co-processors (CoP) for data compression and encryption functions for each core. The compression unit is integrated with the CP assist for cryptographic function (CPACF), benefiting from combining (or sharing) the use of buffers and interfaces. The assist provides high-performance hardware encrypting and decrypting support for clear key operations. For more information, see 3.3.3, Compression and cryptography accelerators on a chip on page 86.
43
8049ch02.fm
accesses appear in order to software. Not all instructions are directly executed by the hardware. This is the case of several complex instructions: some are execute by millicode and some are cracked into multiple operation, which are then executed by the hardware. The following functional areas are on each core, as shown in Figure 2-11: Instruction sequence unit (ISU): This unit enables the out-of-order (OOO) pipeline. It keeps track of register names, OOO instruction dependency, and handling of instruction resource dispatch. This unit is also central to performance measurement through a function called instrumentation. Instruction fetching unit (IFU) (prediction): These unit contain the instruction cache, branch prediction logic, instruction fetching controls, and buffers. Its relative size is the result of the elaborate branch prediction design, which is further described in 3.3.2, Superscalar processor on page 86. Instruction decode unit (IDU): The IDU is fed from the IFU buffers and is responsible for parsing and decoding of all z/Architecture operation codes. Load-store unit (LSU): The LSU contains the data cache and is responsible for handling all types of operand accesses of all lengths, modes and formats as defined in the z/Architecture. Translation unit (XU): The XU has a large translation look-aside buffer (TLB) and the Dynamic Address Translation (DAT) function that handles the dynamic translation of logical to physical addresses. Fixed-point unit (FXU): The FXU handles fixed point arithmetic. Binary floating-point unit (BFU): The BFU handles all binary and hexadecimal floating-point, and fixed-point multiplication and division operations. Decimal unit (DU): The DU executes both floating- point and fixed-point decimal operations. Recovery unit (RU): The RU keeps a copy of the complete state of the system, including all registers, collects hardware fault signals, and manages the hardware recovery actions. Dedicated Co-Processor (COP): The dedicated coprocessor is responsible for data compression and encryption functions for each core.
44
8049ch02.fm
IFU
RU
I D U
ISU
BFU FXU DF U
I-cache XU
LSU
Instr. L2
L2 Control
Data-L2 COP
2.4.3 PU characterization
In each MCM, some PUs can be characterized for customer use. The characterized PUs can be used for general purpose to run supported operating systems (as z/OS, z/VM, Linux on System z), or specialized to run specific workloads (as Java, XML services, IPSec, some DB2 workloads) or functions (as Coupling Facility Control Code). For more information about PU characterization, see 3.4, Processor unit functions on page 91. The maximum number of characterized PUs depends on the zEC12 model. Some PUs are characterized by the system as standard system assist processors (SAPs), to run the I/O processing. Also as standard, there are at least two spare PUs per system, which are used to assume the function of a failed PU. The remaining installed PUs can be characterized for customer use. A zEC12 model nomenclature includes a number which represents this maximum number of PUs that can be characterized for customer use, as shown in Table 2-2.
Table 2-2 Number of PUs per zEC12 model Model H20 H43 H66 H89 HA1 Books 1 2 3 4 4 Installed PUs 27 (1 x 27) 54 (2 x 27) 81 (3 x 27) 108 (4 x 27) 120 (4 x 30) Standard SAPs 4 8 12 16 16 Min Spare PUs 2 2 2 2 2 Max characterized PUs 20 43 66 89 101
45
8049ch02.fm
Fabric
IOs
T OD
Clk
Repower
Data BitStack
L4 Controller
F abric
IOs
Most of the SC chip space is taken by the L4 controller and the 192 MB L4 cache, which consists of four 48 MB quadrants with 256 (1.5 Mb) eDRAM macros per quadrant. The L4 cache is logically organized as 16 address sliced banks, with 24-way set associatively. The L4 cache controller is a single pipeline with multiple individual controllers, sufficient to handle 125 simultaneous cache transactions per chip. The L3 caches on PU chips communicate with the L4 caches via the attached SC chip using uni-diretional busses. L3 is divided into two logical slices, each slice is 24 MB consisting of two 12 MB banks. L3 is 12-way set associative, each bank has 4K sets, and the cache line size is 256B. The bus/clock ratio (2:1) between the L4 cache and the PU is controlled by the storage controller on the SC chip. The SC chip also acts as an L4 cache cross-point switch for L4-to-L4 traffic to up to three remote books by three bidirectional data buses. The integrated SMP fabric transport and system coherency manager use the L4 directory to filter snoop traffic from remote books, with a enhanced synchronous fabric protocol for improved latency and cache management. There are two clock domains and the clock function is distributed among both SC chips.
46
8049ch02.fm
PU Chip 6 Cores
L1 + + + + L1 2MB + + + + 2MB L2 L2 48 MB eDRAM Inclusive L3
PU Chip 6 Cores
L1 + + + + L1 2MB + + + + 2MB L2 L2 48 MB eDRAM Inclusive L3
PU Chip 6 Cores
L1 + + + + L1 2MB + + + + 2MB L2 L2 48 MB eDRAM Inclusive L3
PU Chip 6 Cores
L1 + + + + L1 2MB + + + + 2MB L2 L2 48 MB eDRAM Inclusive L3
PU Chip 6 Cores
L1 + + + + L1 2MB + + + + 2MB L2 L2 48 MB eDRAM Inclusive L3
PU Chip 6 Cores
L1 + + + + L1 2MB + + + + 2MB L2 L2 48 MB eDRAM Inclusive L3
+ Cach e for cores 1 to 6 LRU Cast- Out CP Stores Data Fetch Return
Each core has its own 160 KB cache Level 1 (L1), split into 96 KB for data (D-cache) and 64 KB for instructions (I-cache). The L1 cache is designed as a store-through cache, meaning that altered data is also stored to the next level of memory. The next level is the private cache Level 2 (L2) located on each core, having 2 MB, spitted into 1 MB D-cache and 1 MB I-cache and also designed as a store-through cache. The cache Level 3 (L3) is also located on the PUs chip and shared by the six cores, having 48 MB and designed as a store-in cache. Cache levels L2 and L3 are implemented on the PU chip to reduce the latency between the processor and the large shared cache L4, which is located on the two SC chips. Each SC chip has 192 MB, resulting in 384 MB of L4 cache, which is shared by all PUs on the MCM. The L4 cache uses a store-in design.
47
8049ch02.fm
2.5 Memory
Maximum physical memory size is directly related to the number of books in the system. Each book can contain up to 960 GB of physical memory, for a total of 3840 GB (3.75TB) of installed memory per system. AzEC12 has more memory installed than ordered. Part of the physical installed memory is used to implement the redundant array of independent memory (RAIM) design, resulting on up to 768 GB of available memory per book and up to 3072 GB (3 TB) per system. Table 2-3 on page 48 shows the maximum and minimum memory sizes customer can order for each zEC12 model.
Table 2-3 zEC12 memory sizes Model H20 H43 H66 H89 HA1 Number of books 1 2 3 4 4 Customer memory (GB) 32 - 704 32 - 1392 32 - 2272 32 - 3040 32 - 3040
The minimum physical installed memory is 80 GB per book, and the minimum initial amount of memory that can be ordered is 32 GB for all zEC12 models. The maximum customer memory size is based on the physical installed memory minus RAIM and minus HSA memory, which has a fixed amount of 32 GB. Table 2-4 shows the memory granularity based on the installed customer memory.
Table 2-4 Memory granularity Granularity (GB) 32 64 96 112 128 256 Customer memory (GB) 32 - 256 320 - 512 608 - 896 1008 1136 - 1520 1776 - 3056
With the zEC12 the memory granularity varies from 32 GB (for customer memory sizes from 32 to 256 GB) up to 256 GB (for CPCs having from 1776 GB to 3056 GB of customer memory).
48
8049ch02.fm
MCU1
MD11 MD12 MD13 MD14 MD15 MD16 MD17 MD18 Master (a) Slave (b) MD19 MD20
MCU0
MD01 MD02 MD03 MD04 MD05 MD06 MD07 MD08 MD09 MD10 Channel 0
Channel 1
DIMMs
Channel 2
Channel 3
Channel 4
MCU2
PU2
MCU1
PU1
MCU0
PU0
GX2 GX6
GX1 GX7
GX0 PSI
MCM
SC1 PU3
GX3
GX4
GX5 PSI
Each book has from 10 to 30 dual in-line memory modules (DIMMs). DIMMs are connected to the MCM through three memory control units (MCUs) located on PU0, PU1 and PU2. Each MCU uses five channels, one of them for RAIM implementation, on a 4 +1 (parity) design. Each channel has one or two chained DIMMs, so a single MCU can have five or ten DIMMs. Each DIMM has 4, 16 or 32 GB, and there is no mixing of DIMM sizes on a book.
49
8049ch02.fm
The RAIM design requires the addition of one memory channel that is dedicated for RAS, as shown in Figure 2-15.
Level 4 Cache
16B 16B 16B 16B 16B 16B
Key Cache
Key Cache
Key Cache
MCU 0
2B
MCU 1
2B
MCU 2
2B
DATA CHECK
The parity of the four data DIMMs are stored in the DIMMs attached to the fifth memory channel. Any failure in a memory component can be detected and corrected dynamically. This design takes the RAS of the memory subsystem to another level, making it essentially a fully fault tolerant N+1 design.
50
8049ch02.fm
Memory (GB) 160 192 224 256 320 384 448 512 608 704 800 896 1008 1136 1264 1392 1520 1776 2032 2288 2544 2800 3056
Model H20 Book 1 240 320 320 400 480 640 640 800 800 960 N/A N/A N/A N/A N/A N/A N/A N/A N/A N/A N/A N/A N/A
Model H43 Book 1 120 160 160 240 240 320 320 400 400 480 640 640 640 800 800 960 960 N/A N/A N/A N/A N/A N/A Book 3 100 120 160 160 240 240 320 320 400 480 480 640 640 640 800 800 960 N/A N/A N/A N/A N/A N/A
Model H66 Book 1 80 100 100 120 160 240 240 240 320 320 400 400 480 480 640 640 640 800 960 960 N/A N/A N/A Book 2 80 80 100 120 160 160 240 240 240 320 320 400 400 480 480 640 640 800 800 960 N/A N/A N/A Book 3 60 80 100 100 100 160 160 240 240 320 320 400 400 480 480 480 640 640 800 960 N/A N/A N/A
Model H89 Model HA1 Book 0 60 80 80 100 120 160 160 240 240 240 320 320 320 400 400 480 480 640 640 800 800 960 960 Book 1 60 60 80 80 100 120 160 160 240 240 240 320 320 400 400 480 480 640 640 800 800 960 960 Book 2 60 60 80 80 100 120 160 160 160 240 240 320 320 320 400 400 480 480 640 640 800 800 960 Book 3 40 60 60 80 100 100 100 160 160 240 240 240 320 320 400 400 480 480 640 640 800 800 960
a. 80 GB for a one book system. However if System is ordered with >1 book, 40 GB installed
Physically, memory is organized as follows: A book always contains a minimum of 10 DIMMs of 4 GB each (40 GB). A book has more memory installed than enabled. The amount of memory that can be enabled by the customer is the total physically installed memory minus the RAIM amount and minus the 32 GB HSA memory. A book can have available unused memory, which can be ordered on a memory upgrade.
51
8049ch02.fm
Figure 2-16 illustrates how the physical installed memory is allocated on a zEC12, showing HSA memory, RAIM, customer memory, and the remaining available unused memory that can be enabled by licensed internal code (LIC) when required.
C usto mer
RAIM
RAIM
HSA 32 GB)
H SA (1 6 GB) (
As an example, a zEC12 model H43 (two books) ordered with 192 GB of memory would have memory sizes as follows. See Figure 2-16. Physical installed memory is 280 GB: 160 GB on book 1 and 120 GB on book 3. Book 1 has the 32 GB HSA memory and up to 96 GB for customer memory, while book 2 has up to 96 GB for customer memory, resulting in 192 GB of available memory for the customer. As the customer has ordered 176 GB, provided the granularity rules are met, there are still 16 GB (192 - 176 GB) available to be used in conjunction with additional memory for future upgrades by LIC. Memory upgrades are satisfied from already installed unused memory capacity until it is exhausted. When no more unused memory is available from the installed memory cards (DIMMs), one of the following additions must occur: Memory cards have to be upgraded to a higher capacity. An additional book with additional memory is necessary. Memory cards (DIMMs) must be added. A memory upgrade is concurrent when it requires no change of the physical memory cards. A memory card change is disruptive when no use is made of enhanced book availability. See 2.7.2, Enhanced book availability on page 60. When activated, a logical partition can use memory resources located in any book. No matter in which book the memory resides, a logical partition has access to that memory for up to a maximum of 1 TB. Despite the book structure, thezEC12 is still a symmetric multiprocessor (SMP). For more information, see 3.6, Logical partitioning on page 107.
52
8049ch02.fm
53
8049ch02.fm
Table 2-7 shows the memory granularity for the Flexible Memory Option.
Table 2-7 Flexible memory granularity Granularity (GB) 32 64 96 112 128 256 Flexible memory (GB) 32 - 256 320 - 512 608 - 896a 1008 1136 - 1520b 1776 - 2288
Flexible memory can be purchased but cannot be used for normal everyday use. For that reason, a different purchase price for the flexible memory is offered to increase the overall availability of the system.
54
8049ch02.fm
Table 2-8 Feature codes for plan-ahead memory Memory Pre-planned memory Charged when physical memory is installed. Used for tracking the quantity of physical increments of plan-ahead memory capacity. Pre-planned memory activation Charged when plan-ahead memory is enabled. Used for tracking the quantity of increments of plan-ahead memory that is being activated. zEC12 feature code FC 1996
FC 1901
Installation of pre-planned memory is done by ordering FC 1996. The ordered amount of plan-ahead memory is charged with a reduced price compared to the normal price for memory. One FC 1996 is needed for each 16 GB of usable memory (20 GB RAIM). Activation of installed pre-planned memory is achieved by ordering FC 1901, which causes the other portion of the previously contracted charge price to be invoiced. One FC 1901 is needed for each additional 16 GB to be activated. Memory: Normal memory upgrades use up the plan-ahead memory first.
Actually, it is highly unlikely that either one of the above examples could occur on a zEC12 server. The System z hardware has gone through decades of intense engineering, and this resulted in a very robust and reliable platform. Not only the hardware has a lot of RAS features built-in, but also the software has. IBMs parallel sysplex architecture is a good example of this.
55
8049ch02.fm
Improved error detection for L3/L4 eDRAM configuration latches. Improved error detection for L3/L4 wordline failures. Improved recovery for unresponsive memory DIMM ASIC. Improved recovery of memory command words by changing code points. Improved chip marking technology (marking a chip as defective).
Taken together, this provides the most robust memory subsystem of all generations of System z server.
IBM zEnterprise EC12 continues to deliver robust server designs through exciting new technologies, hardening both new- and classic redundancy. For more information see Chapter 10, RAS on page 355.
56
8049ch02.fm
2.7 Connectivity
Connections to I/O cages, I/O drawers and Parallel Sysplex InfiniBand coupling (PSIFB) are driven from the host channel adapter (HCA) fanouts that are located on the front of the book, while connections to PCIe I/O drawers are driven from the PCIe fanouts also located on the front of the books. Figure 2-17 shows the location of the fanouts and connectors for a two-book system. In the figure, ECF is the External Clock Facility card for the Pulse Per Second (PPS) connectivity, OSC is the oscillator card; FSP is the flexible service processor; and LG is the location code for logic card.
ECF
OSC D1 I/O D2 I/O D3 FSP D4 FSP D5 I/O D6 I/O D7 I/O D8 I/O D9 I/O DA I/O
OSC
ECF D1 I/O D2 I/O D3 FSP D4 FSP D5 I/O D6 I/O D7 I/O D8 I/O D9 I/O DA I/O
filler
filler
LG01
LG06
LG10
LG15
Each book has up to eight fanouts (numbered D1, D2, and D5 through DA). A fanout can be repaired concurrently with the use of redundant I/O interconnect. See 2.7.1, Redundant I/O interconnect on page 59. These are the six types of fanouts: Host Channel Adapter2-C (HCA2-C): Copper connections for InfiniBand I/O interconnect to all FICON, OSA, and Crypto cards in I/O cages and I/O drawers. PCIe fanout: Copper connections for PCIe I/O interconnect to all FICON, OSA, Crypto and Flash Express in PCIe I/O drawers. Host Channel Adapter2-O (HCA2-O): Optical connections for 12x InfiniBand for coupling links (IFB). The HCA2-O (12xIFB) provides a point-to-point connection over a distance of up to 150 m using OM3 fiber optic cables (50/125 m). zEC12 to z196, z114 or System z10 connections use 12-lane InfiniBand link at 6 GBps. Host Channel Adapter2-O Long Reach (HCA2-O LR):
57
8049ch02.fm
Optical connections for 1x IFB and supports IFB Long Reach (IFB LR) coupling links for distances of up to 10 km and up to 100 km when repeated through a System z qualified DWDM. IFB LR coupling links operate at up to 5.0 Gbps between two CPCs, or automatically scale down to 2.5 Gbps depending on the capability of the attached equipment. Host Channel Adapter3-O (HCA3-O): Optical connections for 12x IFB or 12x IFB3 for coupling links. For details, see 12x IFB and 12x IFB3 protocols on page 140. The HCA3-O (12xIFB) provides a point-to-point connection over a distance of up to 150 m using OM3 fiber optic cables (50/125 m). Host Channel Adapter3-O Long Reach (HCA3-O LR): Optical connections for 1x InfiniBand and supports IFB Long Reach (IFB LR) coupling links for distances of up to 10 km and up to 100 km when repeated through a System z qualified DWDM. IFB LR coupling links operate at up to 5.0 Gbps between two CPCs, or automatically scale down to 2.5 Gbps depending on the capability of the attached equipment. Fanout positions: On a model H20 and a model H43, all fanout positions can be populated. On a model H66, all fanout positions can be populated only on the first book. Positions D1 and D2 must remain free of fanouts on both the second and third books. On models H89 and HA1, all D1 and D2 positions must remain free of any fanout. When configuring for availability, the channels, coupling links, and OSAs should be balanced across books. In a system configured for maximum availability, alternate paths maintain access to critical I/O devices, such as disks, networks, and so on. Enhanced book availability (EBA) allows a single book in a multibook CPC to be concurrently removed and reinstalled for an upgrade or a repair. Removing a book means that the connectivity to the I/O devices connected to that book is lost. To prevent connectivity loss, the redundant I/O interconnect feature allows you to maintain connection to critical devices, except for Parallel Sysplex InfiniBand coupling, when a book is removed.
58
8049ch02.fm
...
HCA-C
...
...
...
Slot 5
Interconnect
B andwidth to I/O card depends o n card type 2 GB /sec 1 GB/sec 500 MB/sec 333 MB/sec
Domain 0 Domain 1
Normally, the HCA2-C fanout in the first book connects to the IFB-MP (A) card and services domain 0 in a I/O cage or I/O drawer. In the same fashion, the HCA2-C fanout of the second book connects to the IFB-MP (B) card and services domain 1 in a I/O cage or I/O drawer. If the second book is removed, or the connections from the second book to the cage or drawer are removed, connectivity to domain 1 is maintained by guiding the I/O to domain 1 through the interconnect between IFB-MP (A) and IFB-MP (B).
59
8049ch02.fm
To support Redundant I/O Interconnect (RII) between front to back domain pairs 0,1 and 2,3, the two interconnects to each pair must be driven from 2 different PCIe fanouts. Normally, each PCIe interconnect in a pair supports the eight I/O cards in its domain. In backup operation mode, one PCIe interconnect supports all 16 I/O cards in the domain pair.
Book 0 Memor y
PU PU PU
Book 1 Memory
PU PU PU
Fanouts
PCIe (8x)
x16 PCIe Gen2 8 GBp s
PCIe (8x)
PCIe switch
PCIe switch
PCIe switch
PCIe switch
FICON Express8S
OSA-Express4S
60
8049ch02.fm
Capacity marked CP A processor purchased for future use as a CP is marked as available capacity. It is offline and not available for use until an upgrade for the CP is installed. It does not affect software licenses or maintenance charges. IFL The Integrated Facility for Linux is a processor that is purchased and activated for use by z/VM for Linux guests and Linux on System z operating systems. A processor purchased for future use as an IFL. It is offline and cannot be used until an upgrade for the IFL is installed. It does not affect software licences or maintenance charges. An Internal coupling facility (ICF) processor purchased and activated for use by the Coupling Facility Control Code. A z196 Application Assist Processor (zAAP) purchased and activated to run eligible workloads such as Java code under control of z/OS JVM or z/OS XML System Services. A z196 Integrated Information Processor (zIIP) purchased and activated to run eligible workloads such as DB2 DRDA or z/OS3 Communication Server IPSec. An optional processor that is purchased and activated for use as a system assist processor (SAP).
Unassigned IFL
ICF zAAP
zIIP
Additional SAP
A minimum of one PU characterized as a CP, IFL, or ICF is required per system. The maximum number of CPs, IFLs, and ICFs is 101, The maximum number of zAAPs is 50, but requires an equal or greater number of characterized CPs. The maximum number of zIIPs is also 40 and requires an equal or greater number of characterized CPs. The sum of all zAAPs and zIIPs cannot be larger than two times the number of characterized CPs. Not all PUs on a given model are required to be characterized. The zEC12 model nomenclature is based on the number of PUs available for customer use in each configuration. The models are summarized in Table 2-9.
z/VM V5R4 and above support zAAP and zIIP processors for guest configurations.
61
8049ch02.fm
Table 2-9 z196 configurations Spares / reserved 2/ 1 2/ 1 2/ 1 2/ 1 2/ 1 Yes Yes Yes Yes Model Books PUs per MCM 27 54 81 108 120 CPs IFLs/ uIFL ICFs zAAPs zIIPs Add. SAPs Std. SAPs
1 2 3 4 4
020 019 043 042 066 065 089 088 0101 0100
4 8 12 16 16
A capacity marker identifies that a certain number of CPs have been purchased. This number of purchased CPs is higher than or equal to the number of CPs actively used. The capacity marker marks the availability of purchased but unused capacity intended to be used as CPs in the future. They usually have this status for software-charging reasons. Unused CPs are not a factor when establishing the MSU value that is used for charging MLC software, or when charged on a per-processor basis.
2.8.1 Upgrades
Concurrent upgrades of CPs, IFLs, ICFs, zAAPs, zIIPs, or SAPs are available for the zEC12. However, concurrent PU upgrades require that additional PUs are installed (at a prior time), but not activated. Spare PUs are used to replace defective PUs. There are always two spare PUs on a zEC12. Note that in the rare event of a PU failure, a spare PU is concurrently and transparently activated, and assigned the characteristics of the failing PU. If an upgrade request cannot be accomplished within the given configuration, a hardware upgrade is required. The upgrade enables the addition of one or more books to accommodate the desired capacity. Additional books can be installed concurrently. Although upgrades from one zEC12 model to another zEC12 model are concurrent, meaning that one or more books can be added, there is one exception. Upgrades from any zEC12 (model H20, H43, H66, H89) to a model HA1 is disruptive, because this upgrade requires the replacement of all four books. Table 2-10 shows the possible upgrades within the zEC12 configuration range.
Table 2-10 zEC12 to zEC12 upgrade paths To 2827 From 2827 Model H20 Model H43 Model H66 Model H89 Model H20 Model H43 Yes Model H66 Yes Yes Model H89 Yes Yes Yes Model HA1a
62
8049ch02.fm
a. disruptive upgrade
You can also upgrade a z10 EC or a z196 to a zEC12, while preserving the CPC serial number (S/N). The I/O cards can also be carried forward (with certain restrictions) to the zEC12. Attention: Upgrades from System z10 and System z196 are always disruptive. Upgrade paths from any z10 Enterprise Class to any zEC12 are supported as listed in Table 2-11.
Table 2-11 z10 EC to zEC12 upgrade paths To 2827 From 2097 Model E12 Model E26 Model E40 Model E56 Model E64 Model H20 Yes Yes Yes Yes Yes Model H43 Yes Yes Yes Yes Yes Model H66 Yes Yes Yes Yes Yes Model H89 Yes Yes Yes Yes Yes Model HA1 Yes Yes Yes Yes Yes
Upgrades from any z196 to any zEC12 are supported as listed in Table 2-12.
Table 2-12 z196 to z2120 upgrade paths To 2827 From 2817 Model M15 Model M32 Model M49 Model M66 Model M80 Model H20 Yes Yes Yes Yes Yes Model H43 Yes Yes Yes Yes Yes Model H66 Yes Yes Yes Yes Yes Model H89 Yes Yes Yes Yes Yes Model HA1 Yes Yes Yes Yes Yes
63
8049ch02.fm
Granular capacity
The zEC12 offers 60 capacity settings at the low end of the processor. Only 20 CPs can have granular capacity. When subcapacity settings are used, other PUs, beyond 20, can only be characterized as specialty engines. The three defined ranges of subcapacity settings have model capacity identifiers numbered from 401 to 420, 501 to 520, and 601 to 620. CPs: Within a zEC12, all CPs have the same capacity identifier. Specialty engines (IFLs, zAAPs, zIIPs, ICFs) operate at full speed.
64
8049ch02.fm
Model capacity identifier 701789, 601620, 501520, 401420 7017A1, 601620, 501520, 401420
Attention: On zEC12, model capacity identifier 400 is used for IFL or ICF only configurations.
65
8049ch02.fm
prevent sudden deactivation when the 90-day period expires. The contract duration can be set from one to five years. The CBU record describes the following properties related to the CBU: Number of CP CBUs allowed to be activated Number of IFL CBUs allowed to be activated Number of ICF CBUs allowed to be activated Number of zAAP CBUs allowed to be activated Number of zIIP CBUs allowed to be activated Number of SAP CBUs allowed to be activated Number of additional CBU tests allowed for this CBU record Number of total CBU years ordered (duration of the contract) Expiration date of the CBU contract The record content of the CBU configuration is documented in IBM configurator output, shown in Example 2-1. In the example, one CBU record is made for a 5-year CBU contract without additional CBU tests for the activation of one CP CBU.
Example 2-1 Simple CBU record and related configuration features
On Demand Capacity Selecions: NEW00001 - CBU - CP(1) - Years(5) - Tests(5) Resulting feature numbers in configuration: 6817 6818 6820 Total CBU Years Ordered CBU Records Ordered Single CBU CP-Year 5 1 5
In Example 2-2, a second CBU record is added to the configuration in Example 2-1for two CP CBUs, two IFL CBUs, two zAAP CBUs, and two zIIP CBUs, with five additional tests and a 5-year CBU contract. The result is now a total number of 10 years of CBU ordered, which is the five years in the first record and five years in the second record. The two CBU records are independent and can be activate individually. Five additional CBU tests have been requested, and because there is a total of five years contracted for a total of three CP CBUs, two IFL CBUs, two zAAPs, and two zIIP CBUs, they are shown as 15, 10, 10, and 10 CBU years for their respective types.
Example 2-2 Second CBU record and resulting configuration features
NEW00001 - CBU - Replenishment is required to reactivate Expiration(06/21/2017) NEW00002 - CBU - CP(2) - IFL(2) - zAAP(2) - zIIP(2) Total Tests(10) - Years(5)
Resulting cumulative feature numbers in configuration: 6805 6817 6818 6820 6822 6826 6828 5 Additional CBU Tests Total CBU Years Ordered CBU Records Ordered Single CBU CP-Year Single CBU IFL-Year Single CBU zAAP-Year Single CBU zIIP-Year 5 10 2 15 10 10 10
66
8049ch02.fm
Model H20 Model H43 Model H66 Model H89 Model HA1
Unassigned IFLs are ignored. They are considered spares and are available for use as CBU. When an unassigned IFL is converted to an assigned IFL, or when additional PUs are characterized as IFLs, the number of CBUs of any type that can be activated is decreased.
67
8049ch02.fm
when a model capacity identifier 504 has two CP5s added temporarily, it becomes a model capacity identifier 506. When the addition of temporary capacity requested by On/Off CoD for CPs results in a cross-over from one capacity identifier range to another, the total number of CPs active when the temporary CPs are activated is equal to the number of temporary CPs ordered. For example, when a CPC with model capacity identifier 504 specifies six CP6 temporary CPs through On/Off CoD, the result is a CPC with model capacity identifier 606. A cross-over does not necessarily mean that the CP count for the additional temporary capacity will increase. The same 504 could temporarily be upgraded to a CPC with model capacity identifier 704. In this case, the number of CPs does not increase, but additional temporary capacity is achieved.
68
8049ch02.fm
Tr an sfer Sw itch
UPS
Step dow n
AC/DC
480VAC
DC/AC
PDU
208 VA C 3 phase
AC/DC
4 10V DC
DC/DC
12V D C
Batteries
PSU
Lo sses ~ 13%
Server
480VAC 3 phas e
12 V DC
520VDC
Batteri es
AC/DC
PSU
Lo sses ~4%
Server
UPS
D Input C
DC/DC
12V DC
DC/ DC
AC/ DC
PSU
Server
69
8049ch02.fm
2.9.6 Cooling
Water cooling technology is fully utilized inzEC12 MCM cooling. ForzEC12, it still has air cooled models and water cooled models.
70
8049ch02.fm
In addition to the MCMs, the internal water loop also circulates through two heat exchangers located in the path of the exhaust air in the rear of the frames. These heat exchangers will remove approximately 60% to 65% of the residual heat from the I/O drawers, the air cooled logic in the books and the heat generated within the power enclosures. Almost 2/3 of the total heat generated to be removed from the room by the chilled water. Selection of air cooled models or water cooled models is done at ordering and the appropriate equipment is factory installed. MES from air cooled model to water cooled model is not allowed, vice versa.
Blow er s
Mani fo ld Assemb ly
Pum ps
The radiator unit can connect to all four books and cool all MCMs simultaneously. The cooling capability is redundant design, a single working pump and blower can support entire load. Replacement of pump or blower is fully redundant, without performance impact. Figure 2-22 shows the closed water loop in radiator unit. The warm water exiting from the MCM code plates enters pumps through a common manifold, being pumped through a heat exchanger where heat extracted by the air flowing across the heat exchanger fins. The cooled water is then recirculated back into the MCM cold plates. The radiator air cooling system is open loop, that is to say there is no specific target temperature but only an estimated range based upon MCM power and ambient temperature. However the radiator blowers will be speed up in increment of air flow when MCM temperature increasing.
71
8049ch02.fm
MC M Cold Plate
Book 1
Blower
Same as MRU design in z196, backup blowers is the redundant design of radiator unit in zEC12. If MCM temperature increases to threshold limit for any abnormal situations, such as radiator unit fails or environment temperature is too high, the backup blowers will be turned on to strengthen MCM cooling, and cycle steering mode is required. Please reference section 2.9.9, Backup air cooling system to learn more about backup blower and cycle steering mode. In zEC12 backup blowers are also provided primarily to permit concurrent replacement of the radiator (heat exchanger) or the return manifold. At this time, cycle steering mode may required.
MCM Co ld Plate
Chilled
Cold
Warmer
Pump
Warm er
72
8049ch02.fm
The water in the closed loop within the system exchanges heat with the continuous supply of building chilled water. The internal water loop contains approximately 23 l (6 gallons) of conditioned water which circulates between the MCM cold plate and a heat exchanger within the WCU. Here the system water comes in 'contact' with the customer supplied chilled water. Heat from the MCM transfers to the cold plate where it is in turn transferred to the circulating system water. Finally the system water gives up its heat to the building chilled water within the WCU's heat exchanger and the MCM is efficiently cooled in this manner. The zEC12 operates with two fully redundant water cooling units (WCUs). These water cooling units each have their own facility feed and return water connections. If water is interrupted to one of the water cooling units, the other water cooling unit will pick up the entire load and the server will continue to operate without interruption. The customer is required to provide independent redundant water loops to the water cooling units in order to obtain the full redundancy. In the event of a total loss of building chilled water or in the unlikely event that both water cooling units fail, the backup blowers will be turned on to keep server running, same as air cooled models. At this time, cycle time degradation is required, please reference 2.9.9, Backup air cooling system for more details. The internal circulating water in water cooling unit is conditioned water and will be added to the two WCUs during machine installation with specific Fill and Drain Tool (FC 3378), same as radiator unit. Water and Drain Tool will be shipped with new zEC12.
73
8049ch02.fm
In addition to the MCM cold plates, the internal water loop also circulates through these two heat exchangers located in the path of the exhaust air in the rear of the frames. These two heat exchangers will remove approximately 65% of the residual heat from the I/O drawers, the air cooled logic in the book and the heat generated within the power enclosures. The goal is for 2/3 of the total heat generated to be removed from the room by the chilled water. A diagrammatic view of the entire circulation system is shown in Figure 2-25.
Chil led water feed Nod Feed Manifold WCU 1 Chil led water return Chil led water feed WCU 2 Chil led water return Boo k 3 Boo k 0 Retur n Manifo ld A F rame Heat Exchanger
Boo k 1
Boo k 2
Above figure shows the complete water loop of WCU. Two customer water feeds connect to the two redundant WCUs. The WCUs feed water to a common feed manifold which supplies water to the frame heat exchangers. Two frame heat exchangers are fed in parallel from this manifold. The node feed manifold feeds the four book positions in parallel. The heat is exchanged with the customer supplied chilled water in the WCUs which also pump the system water around the loop. If one customer water supply or one WCU fails, the remaining feed will maintain MCM cooling. WCUs and the associated drive card are concurrently replaceable. Also, the heat exchangers can be disconnected and removed from the system concurrently.
Water cooled system power in normal room/hot room -est. 12.9 kW / 14.1 kW 17.4 kW / 19.0 kW 24.7 kW / 26.3 kW
74
8049ch02.fm
Temperature
Inlet air temperature Heat to water and as % of total system heat load 18 C 23 C 27 C 32 C (hot room) 7.3 kW (57%) 9.5 kW (74%) 11.5 kW (89%) 14.8 kW (105%) 9.8 kW (56%) 12.6 kW (72%) 14.8 kW (85%) 18.2 kW (96%) 12.6 kW (51%) 15.6 kW (63%) 18.0 kW (73%) 21.6 kW (82%)
75
8049ch02.fm
76
8049ch03.fm
Chapter 3.
77
8049ch03.fm
Federal Information Processing Standards (FIPS)140-2 Security Requirements for Cryptographic Modules
78
8049ch03.fm
Have a balanced system design, providing large data rate bandwidths for high performance connectivity along with processor and system capacity. The following sections describe the zEC12 system structure, showing a logical representation of the data flow from PUs, caches, memory cards, and a variety of interconnect capabilities.
79
8049ch03.fm
DIMMs One cache per Book over two SC Chip (fully shared) SC Chips One per PU Chip (shared by 6 cores) up to 30 (PUs)
384 MB
Level 4 eDRAM
up to four (books)
Level 3
48 MB
eDRAM
Level 2 Level 1
1 MB / 1 MB 64 KB I-cache 96 KB D-cache
SRAM
S RAM
The 4-level cache structure is implemented within the MCM. The first three levels (L1, L2, and L3) are located on each PU chip and the last level (L4) resides on SC chips: L1 and L2 caches use static random access memory (SRAM) and are private for each core. L3 cache uses embedded dynamic static random access memory (eDRAM) and is shared by all six cores within the PU chip. Each book has six L3 caches and a four-book system has 24 of them, resulting in 1152 MB (48 x 24 MB) of this shared PU chip level cache. L4 cache also uses eDRAM and is shared by all PU chips on the MCM. A four-book system has 1536 MB (4 x 384 MB) of shared L4 cache. Main storage has up to 0.75 TB per book, using up to 30 DIMMs. A four-book system can have up to 3 TB of main storage. Cache sizes are being limited by ever-diminishing cycle times because they must respond quickly without creating bottlenecks. Access to large caches costs more cycles. Instruction and data caches (L1) sizes must be limited because larger distances must be traveled to reach long cache lines. This L1 access time should occur in one cycle, avoiding increased latency. Also, the distance to remote caches as seen from the microprocessor becomes a significant factor. An example of this is the L4 cache that is not on the microprocessor (and might not even be in the same book). Although the L4 cache is rather large, the reduced cycle time has the effect that more cycles are needed to travel the same distance. In order to overcome this and avoid potential latency, zEC12 uses two additional cache levels (L2 and L3) within the PU chip, with denser packaging. This design reduces traffic to and from the shared L4 cache, which is located on another chip (SC chip). Only when there is a cache miss in L1, L2, and L3, the request is sent to L4. L4 is the coherence manager, meaning that all memory fetches must be in the L4 cache before that data can be used by the processor.
80
8049ch03.fm
Another approach is available for avoiding L4 cache access delays (latency) as much as possible. The L4 cache straddles up to four books. This means relatively large distances exist between the higher-level caches in the processors and the L4 cache content. To overcome the delays that are inherent to the book design and to save cycles to access the remote L4 content, it is beneficial to keep instructions and data as close to the processors as possible by directing as much work of a given logical partition workload on the processors located in the same book as the L4 cache. This is achieved by having the PR/SM scheduler and the z/OS dispatcher work together to keep as much work as possible within the boundaries of as few processors and L4 cache space (which is best within a book boundary) as can be achieved without affecting throughput and response times. Preventing PR/SM and the dispatcher from scheduling and dispatching a workload on any processor available, and keeping the workload in as small a portion of the system as possible, contributes to overcoming latency in a high-frequency processor design such as the zEC12. The cooperation between z/OS and PR/SM has been bundled in a function called HiperDispatch. HiperDispatch exploits the zEC12 cache topology, with reduced cross-book help, and better locality for multi-task address spaces. For more information on HiperDispatch, see 3.6, Logical partitioning on page 107. Figure 3-2 compares the cache structures of the zEC12 with the System z previous generation system, the z196.
z196
4 L4 Cac hes 192 MB Shared eDRAM L4
2 4 MB Sh r eD RAM L3
L2 L2 L2 L2 L1 L1 L1 L1
EC12
4 L4 Caches 384 MB Shar ed eDRAM L4
48 MB Shr eDR AM L 3
L2 L2 L2 L2 L1 L1 L1 L1 L2 L1 L2 L1
6 L3s, 3 6 L1 / L2s
48 MB Shr eD RAM L3
L2 L2 L1 L1 L2 L2 L2 L2 L1 L1 L1 L1
L2 L2 L1 L1
L1 :
64 K I + 12 8 KD 8w DL1, 4 w IL1 25 6 B line size Private 1 .5 MB Inclus ive of L1s 12 w Set As soc iative 25 6 B c ac he line s ize Sha red 24 MB Inc lusive of L2s 12 w Set As soc iative 25 6 B c ac he line s ize 19 2 MB Inclus iv e 24 w Set As soc iative 25 6 B c ac he line s ize
L1:
64 K I + 96 K D 8w DL1, 4 w IL1 25 6 B line size Private 1 MB Inclusive of DL1 Private 1 MB Inclusive of IL1 8w Set Ass oc ia tiv e 25 6 B c ac he line s ize Sha red 48 MB Inc lusiv e of L2s 12 w Set As sociative 25 6 B c ac he line s ize 38 4 MB Inclus ive 24 w Set As sociative 25 6 B c ac he line s ize
L2
L2
L3
L3
L4
L4
Compared to z196, the zEC12 cache design has much larger cache level sizes, except for the L1 private cache on each core (as its access time should occur in one cycle). The zEC12 cache level structure is focused on keeping more data closer to the processor unit. This design can improve system performance on many production workloads.
81
8049ch03.fm
6PU
6PU
6PU
6PU
6PU
6PU
L4 cache
6PU 6PU 6PU
L4 cache
6PU 6PU 6PU
6PU
6PU
6PU
6PU
6PU
6PU
L4 cache
6PU 6PU 6PU
L4 cache
6PU 6PU 6PU
Inter-book communication takes place at the L4 cache level, which is implemented on Storage Controller (SC) cache chips in each MCM. The SC function regulates coherent book-to-book traffic.
8049ch03.fm
Numerous improvements in the out of order design Enhanced instruction dispatch and grouping efficiency Enhanced branch prediction structure and sequential instruction fetching Millicode improvements Transactional execution (TX) facility Runtime instrumentation (RI) facility Enhanced DAT-2 for 2GB pages support Decimal floating point (DFP) improvements. The zEC12 enhanced instruction set architecture (ISA) includes a set of instructions added to improve compiled code efficiency. This results on optimized processor units to meet the demands of a wide variety of business workload types without compromising the performance characteristics of traditional workloads.
83
8049ch03.fm
Increased OOO resources (Global Completion Table entries, physical GPR entries, physical FPR entries) Improved completion rate Reduced cache/TLB miss penalty Improved execution of D-Cache store and reload and new Fixed-point divide. New OSC (load-hit-store conflict) avoidance scheme Enhanced branch prediction structure and sequential instruction fetching.
Program results
The OOO execution does not change any program results. Execution can occur out of (program) order, but all program dependencies are honored, ending up with same results of the in order (program) execution. This implementation requires special circuitry to make execution and memory accesses appear in order to software. The logical diagram of a zEC12 core is shown in Figure 3-4.
Out Of Order (OOO) addition
Decode/Dispatch1 InstQ GCT Rename VBQ VBU Issue Queue Fixed Pt GPR Fixed Pt
Group of micro-ops (up to 3)
0cycE
Group of micro-ops (up to 3)
VBQ
Load/ Store2
VBU
on-the-fly OSC detection bubble-free instruction expansion 2) banking for concurrent load/stores prefetching enhancements
1)
L1 Dir
L2 Dir
FPR
Memory address generation and memory accesses can occur out of (program) order. This capability can provide a greater exploitation of the zEC12 superscalar core, and can improve system performance. Figure 3-5 shows how OOO core execution can reduce the execution time of a program.
84
8049ch03.fm
3 4 5 6 7 Time
The left side of the example shows an in-order core execution. Instruction 2 has a big delay due to an L1 cache miss, and the next instructions wait until instruction 2 finishes. In the usual in-order execution, the next instruction waits until the previous one finishes. Using OOO core execution, shown on the right side of the example, instruction 4 can start its storage access and execution while instruction 2 is waiting for data, if no dependencies exist between both instructions. So, when the L1 cache miss is solved, instruction 2 can also start its execution while instruction 4 is executing, Instruction 5 might need the same storage data required by instruction 4, and as soon as this data is on L1 cache, instruction 5 starts execution at the same time. The zEC12 superscalar PU core can have up to seven instructions/operations in execution per cycle. Compared to the z196, the first IBM System z CPC that used the OOO technology, further enhancements to the execution cycle are integrated in the cores, which results in a shorter execution time.
85
8049ch03.fm
Challenges of creating a superscalar processor: Many challenges exist in creating an efficient superscalar processor. The superscalar design of the PU has made big strides in avoiding address generation interlock (AGI) situations. Instructions requiring information from memory locations can suffer multi-cycle delays to get the desired memory content, and because high-frequency processors wait faster, the cost of getting the information might become prohibitive.
Coprocessor units
There is one coprocessor (CoP) unit for compression and cryptography on each core in the chip (see Figure 3-6 on page 87). The compression engine uses static dictionary compression and expansion. The dictionary size is up to 64KB, with 8K entries, and has a local 16 KB cache for dictionary data. The cryptography engine is used for CP assist for cryptographic function (CPACF), which offers a set of symmetric cryptographic functions for high encrypting and decrypting performance of clear key operations.
86
8049ch03.fm
IF U
I D U
RU IS U
F XU D F U B FU
I-c a c h e XU In s tr. L2
L SU
L2 C o n tro l
D a ta -L 2
CO P
Federal Information Processing Standards (FIPS)140-2 Security Requirements for Cryptographic Modules
87
8049ch03.fm
decimal floating point computations, and an instruction that performs data conversions to and from the decimal floating point representation. The decimal floating point (DFP) architecture on zEC12 has been improved to facilitate better performance on traditional zoned-decimal operations for Cobol programs. Additional instructions are provided to convert zoned-decimal data into DFP format in FPR registers.
Software support
Decimal floating point is supported in various programming languages and products such as: Release 4 and later of the High Level Assembler C/C++ (requires z/OS 1.10 with program temporary fixes, PTFs, for full support) Enterprise PL/I Release 3.7 and Debug Tool Release 8.1 Java Applications using the BigDecimal Class Library SQL support as of DB2 Version 9
88
8049ch03.fm
Soft Error
Instruction State Checkpoint
Hard Error
PU x
PU y
89
8049ch03.fm
90
8049ch03.fm
3.4.1 Overview
All PUs on a zEC12 are physically identical. When the system is initialized, PUs can be characterized to specific functions: CP, IFL, ICF, zAAP, zIIP, or SAP. The function assigned to a PU is set by the licensed internal code (LIC), which is loaded when the system is initialized (at power-on reset) and the PU is characterized. Only characterized PUs have a designated function. Non-characterized PUs are considered spares. At least one CP, IFL, or ICF must be ordered on a zEC12.
Chapter 3. CPC System design
91
8049ch03.fm
This design brings outstanding flexibility to the zEC12, because any PU can assume any available characterization. This also plays an essential role in system availability, because PU characterization can be done dynamically, with no system outage, allowing the actions discussed in the following sections. Also see Chapter 8, Software support on page 243 for information about software level support on functions and features.
Concurrent upgrades
Except on a fully configured model, concurrent upgrades can be done by the licensed internal code (LIC), which assigns a PU function to a previously non-characterized PU. Within the book boundary or boundary of multiple books, no hardware changes are required, and the upgrade can be done concurrently through the following facilities: Customer Initiated Upgrade (CIU) facility for permanent upgrades On/Off Capacity on Demand (On/Off CoD) for Temporary upgrades Capacity Backup (CBU) for Temporary upgrades Capacity for Planned Event (CPE) for temporary upgrades. If the MCMs in the installed books have no available remaining PUs, an upgrade results in a model upgrade and the installation of an additional book (up to the limit of 4 books). Book installation is nondisruptive, but can take more time than a simple LIC upgrade. For more information about Capacity on Demand, see Chapter 9, System upgrades on page 311.
PU sparing
In the rare event of a PU failure, the failed PUs characterization is dynamically and transparently reassigned to a spare PU. The zEC12 has two spare PUs. PUs not characterized on a CPC configuration can also be used as additional spare PUs. More information about PU sparing is provided in 3.4.11, Sparing rules on page 102.
PU pools
PUs defined as CPs, IFLs, ICFs, zIIPs, and zAAPs are grouped together in their own pools, from where they can be managed separately. This significantly simplifies capacity planning and management for logical partitions. The separation also has an effect on weight management because CP, zAAP, and zIIP weights can be managed separately. For more information, see PU weighting on page 93. All assigned PUs are grouped together in the PU pool. These PUs are dispatched to online logical PUs. As an example, consider a zEC12 with ten CPs, three zAAPs, two IFLs, two zIIPs, and one ICF. This system has a PU pool of 18 PUs, called the pool width. Subdivision of the PU pool defines: A CP pool of ten CPs An ICF pool of one ICF An IFL pool of two IFLs A zAAP pool of three zAAPs A zIIP pool of two zIIPs. PUs are placed in the pools according to the following occurrences: When the system is power-on reset At the time of a concurrent upgrade As a result of an addition of PUs during a CBU 92
IBM zEnterprise EC12 Technical Guide
8049ch03.fm
Following a capacity on demand upgrade, through On/Off CoD or CIU. PUs are removed from their pools when a concurrent downgrade takes place as the result of removal of a CBU, and through On/Off CoD and conversion of a PU. Also, when a dedicated logical partition is activated, its PUs are taken from the proper pools, as is the case when a logical partition logically configures a PU on, if the width of the pool allows. By having different pools, a weight distinction can be made between CPs, zAAPs, and zIIPs. On earlier machines specialty engines such as zAAPs automatically received the weight of the initial CP. For a logical partition, logical PUs are dispatched from the supporting pool only. This means that logical CPs are dispatched from the CP pool, logical zAAPs are dispatched from the zAAP pool, logical zIIPs from the zIIP pool, logical IFLs from the IFL pool, and the logical ICFs from the ICF pool.
PU weighting
Because zAAPs, zIIPs, IFLs, and ICFs have their own pools from where they are dispatched, they can be given their own weights. For more information about PU pools and processing weights, see zEnterprise System Processor Resource/Systems Manager Planning Guide, SB10-7156.
Granular capacity
The zEC12 recognizes four distinct capacity settings for CPs. Full-capacity CPs are identified as CP7. In addition to full-capacity CPs, three subcapacity settings (CP6, CP5, and CP4), each for up to 20 CPs, are offered. The four capacity settings appear in hardware descriptions, as follows: CP7 feature code 1908 CP6 feature code 1907 CP5 feature code 1906 CP4 feature code 1905 Granular capacity adds 60 subcapacity settings to the 101 capacity settings that are available with full capacity CPs (CP7). Each of the 60 subcapacity settings applies only to up to 20 CPs, independently of the model installed.
Chapter 3. CPC System design
93
8049ch03.fm
Information about CPs in the remainder of this chapter applies to all CP capacity settings, CP7, CP6, CP5, and CP4, unless indicated otherwise. See 2.8, Model configurations on page 61, for more details about granular capacity.
IFL pool
All PUs characterized as IFLs within a configuration are grouped into the IFL pool. The IFL pool can be seen on the HMC workplace. IFLs do not change the model capacity identifier of the zEC12. Software product license charges based on the model capacity identifier are not affected by the addition of IFLs.
Unassigned IFLs
An IFL that is purchased but not activated is registered as an unassigned IFL (FC 1914). When the system is subsequently upgraded with an additional IFL, the system recognizes that an IFL was already purchased and is present.
8049ch03.fm
Shared ICFs add flexibility. However, running only with shared coupling facility PUs (either ICFs or CPs) is not a desirable production configuration. It is preferable for a production CF to operate by using dedicated ICFs. In Figure 3-8, the CPC on the left has two environments defined (production and test), each having one z/OS and one coupling facility image. The coupling facility images are sharing the same ICF.
Test Sysplex
ICF
z/OS Test
z/OS
Prod
CF CF
z/OS Test
z/OS CF Prod
HMC
Setup
Figure 3-8 ICF options; shared ICFs
The logical partition processing weights are used to define how much processor capacity each coupling facility image can have. The capped option can also be set for a test coupling facility image to protect the production environment. Connections between these z/OS and coupling facility images can use internal coupling links (ICs) to avoid the use of real (external) coupling links and to get the best link bandwidth available.
95
8049ch03.fm
Potential cost savings. Simplification of infrastructure as a result of the co-location and integration of new applications with their associated database systems and transaction middleware (such as DB2, IMS, or CICS). Simplification can happen, for example, by introducing a uniform security environment, reducing the number of TCP/IP programming stacks and system interconnect links. Prevention of processing latencies that would occur if Java application servers and their database servers were deployed on separate server platforms. One CP must be installed with or prior to installing a zAAP. The number of zAAPs in a CPC cannot exceed the number of purchased CPs. Within the capacity of the sum of all unassigned PUs in up to four books, up to 50 zAAPs on a model HA1 can be characterized. Table 3-1 shows the allowed number of zAAPs for each model.
Table 3-1 Number of zAAPs per model Model zAAPs H20 0 10 H43 0 21 H66 0 33 H89 0 44 HA1 0 50
The number of permanent zAAPs plus temporary zAAPs cannot exceed the number of purchased (permanent plus unassigned) CPs plus temporary CPs. Also, the number of temporary zAAPs cannot exceed the number of permanent zAAPs. PUs characterized as zAAPs within a configuration are grouped into the zAAP pool. This allows zAAPs to have their own processing weights, independent of the weight of parent CPs. The zAAP pool can be seen on the hardware console. zAAPs are orderable by feature code (FC 1912). Up to one zAAP can be ordered for each CP or marked CP configured in the CPC.
96
8049ch03.fm
Availability is treated as follows: If a zAAP is available (not busy), the dispatcher suspends the JVM task on the CP, and assigns the Java task to the zAAP. When the task returns control to the JVM, it passes control back to the dispatcher that reassigns the JVM code execution to a CP. If no zAAP is available (all busy) at that time, the z/OS dispatcher can allow a Java task to run on a standard CP, depending on the option used in the OPT statement in the IEAOPTxx member of SYS1.PARMLIB.
Standard Processor
WebSphere WebSphere Execute JAVA Code Execute JAVACode
zAAP z/OS Dispatcher z/OS z/OS Dispatcher Dispatcher Dispatch JVM task on z/OS Dispatch JVM task on z/OS zAAP logical processor zAAP processor logical logical processor
JVM JVM
z/OS Dispatcher z/OS Dispatcher Suspend JVM task on z/OS Suspend JVM task on z/OS standard logical processor standard logical processor z/OS Dispatcher z/OS Dispatcher z/OS Dispatcher Java Application Code Java ApplicationCode Executing on a zAAP Executing on a zAAP logical processor logical processor
Dispatch JVM task on z/OS Dispatch JVM task on z/OS standard logical processor standard logical processor
JVM
JVM JVM
JVM JVM JVM Switch to Switch to standard processor Switch to standard processor
z/OS Dispatcher z/OS Dispatcher Suspend JVM task on z/OS Suspend JVM task on z/OS zAAP logical processor zAAP logical processor
WebSphere WebSphere
A zAAP executes only IBM authorized code. This includes the z/OS JVM in association with parts of system code, such as the z/OS dispatcher and supervisor services. A zAAP is not able to process I/O or clock comparator interruptions and does not support operator controls such as IPL. Java application code can either run on a CP or a zAAP. The installation can manage the use of CPs such that Java application code runs only on CPs, only on zAAPs, or on both. Three execution options for Java code execution are available. These options are user specified in IEAOPTxx and can be dynamically altered by the SET OPT command. The options that are currently supported for z/OS V1R10 and later releases are as follows. Option 1: Java dispatching by priority (IFAHONORPRIORITY=YES): This is the default option and specifies that CPs must not automatically consider zAAP-eligible work for dispatching on them. The zAAP-eligible work is dispatched on the zAAP engines until Workload Manager (WLM) considers that the zAAPs are overcommitted. WLM then requests help from the CPs. When help is requested, the CPs consider dispatching zAAP-eligible work on the CPs themselves based on the dispatching priority relative to other workloads. When the zAAP engines are no longer overcommitted, the CPs stop considering zAAP-eligible work for dispatch.
97
8049ch03.fm
This option has the effect of running as much zAAP-eligible work on zAAPs as possible and only allowing it to spill over onto the CPs when the zAAPs are overcommitted. Option 2: Java dispatching by priority (IFAHONORPRIORITY=NO): zAAP-eligible work executes on zAAPs only while at least one zAAP engine is online. zAAP-eligible work is not normally dispatched on a CP, even if the zAAPs are overcommitted and CPs are unused. The exception to this is that zAAP-eligible work may sometimes run on a CP to resolve resource conflicts, and other reasons. Therefore, zAAP-eligible work does not affect the CP utilization that is used for reporting through SCRT, no matter how busy the zAAPs are. Option 3: Java discretionary crossover (IFACROSSOVER=YES or NO): As of z/OS V1R8 (and the IBM zIIP Support for z/OS V1R7 Web deliverable), the IFACROSSOVER parameter is no longer honored. If zAAPs are defined to the logical partition but are not online, the zAAP-eligible work units are processed by CPs in order of priority. The system ignores the IFAHONORPRIORITY parameter in this case and handles the work as though it had no eligibility to zAAPs.
98
8049ch03.fm
z/OS Communications Server exploits the zIIP for eligible Internet protocol security (IPSec) network encryption workloads. This requires z/OS V1R10 or later releases. Portions of IPSec processing take advantage of the zIIPs, specifically end-to-end encryption with IPSec. The IPSec function moves a portion of the processing from the general-purpose processors to the zIIPs. In addition to performing the encryption processing, the zIIP also handles cryptographic validation of message integrity and IPSec header processing. z/OS Global Mirror, formerly known as Extended Remote Copy (XRC), exploits the zIIP too. Most z/OS DFSMS system data mover (SDM) processing associated with zGM is eligible to run on the zIIP. This requires z/OS V1R10 or later releases. The first IBM exploiter of z/OS XML system services is DB2 V9. With regard to DB2 V9 prior to the z/OS XML system services enhancement, z/OS XML system services non-validating parsing was partially directed to zIIPs when used as part of a distributed DB2 request through DRDA. This enhancement benefits DB2 V9 by making all z/OS XML system services non-validating parsing eligible to zIIPs when processing is used as part of any workload running in enclave SRB mode. z/OS Communications Server also allows the HiperSockets Multiple Write operation for outbound large messages (originating from z/OS) to be performed by a zIIP. Application workloads based on XML, HTTP, SOAP, Java, etc as well as traditional file transfer, can benefit. For business intelligence, IBM Scalable Architecture for Financial Reporting provides a high-volume, high performance reporting solution by running many diverse queries in z/OS batch and can also be eligible for zIIP. For more information, see the IBM zIIP website: http://www-03.ibm.com/systems/z/hardware/features/ziip/about.html
zIIPs are orderable by feature code (FC 1913). Up to one zIIP can be ordered for each CP or marked CP configured in the system. If the installed books have no remaining unassigned PUs, the assignment of the next zIIP might require the installation of an additional book. PUs characterized as zIIPs within a configuration are grouped into the zIIP pool. By doing this, zIIPs can have their own processing weights, independent of the weight of parent CPs. The zIIP pool can be seen on the hardware console. The number of permanent zIIPs plus temporary zIIPs cannot exceed the number of purchased CPs plus temporary CPs. Also, the number of temporary zIIPs cannot exceed the number of permanent zIIPs.
99
8049ch03.fm
The zAAP on zIIP capability is available to z/OS when running as a guest of z/VM on machines with zAAPs installed, provided that no zAAPs are defined to the z/VM LPAR. This allows, for instance, testing this capability to estimate usage before committing to production.
100
8049ch03.fm
SAP configuration
A standard SAP configuration provides a very well-balanced system for most environments. However, there are application environments with very high I/O rates (typically various TPF environments). In this case, optional additional SAPs can be ordered. Assignment of additional SAPs can increase the capability of the channel subsystem to perform I/O operations. In zEC12 systems, the number of SAPs can be greater than the number of CPs.
101
8049ch03.fm
Reserved processors can be dynamically configured online by an operating system that supports this function, if enough unassigned PUs are available to satisfy this request. The PR/SM rules regarding logical processor activation remain unchanged. Reserved processors provide the capability to define to a logical partition more logical processors than the number of available CPs, IFLs, ICFs, zAAPs, and zIIPs in the configuration. This makes it possible to configure online, nondisruptively, more logical processors after additional CPs, IFLs, ICFs, zAAPs, and zIIPs have been made available concurrently with one of the Capacity on Demand options. The maximum number of reserved processors that can be defined to a logical partition depends on the number of logical processors that are already defined. The maximum number of logical processors plus reserved processors is 101. Do not define more active and reserved processors than the operating system for the logical partition can support. For more information about logical processors and reserved processors and their definition, see 3.6, Logical partitioning on page 107.
102
8049ch03.fm
Systems with a failed PU for which no spare is available will call home for a replacement. A system with a failed PU that has been spared and requires an MCM to be replaced (referred to as a pending repair) can still be upgraded when sufficient PUs are available.
Application preservation
If no spare PU is available, application preservation (z/OS only) is invoked. The state of the failing processor is passed to another active processor used by the operating system and, through operating system recovery services, the task is resumed successfully (in most cases, without customer intervention).
3.5.1 Overview
The zEC12 memory design also provides flexibility and high availability, allowing upgrades: Concurrent memory upgrades (if the physically installed capacity is not yet reached):
103
8049ch03.fm
The zEC12 can have more physically installed memory than the initial available capacity. Memory upgrades within the physically installed capacity can be done concurrently by LIC, and no hardware changes are required. Note that memory upgrades cannot be done through Capacity BackUp (CBU) or On/Off CoD. Concurrent memory upgrades (if the physically installed capacity is reached): Physical memory upgrades require a book to be removed and re-installed after having replaced the memory cards in the book. Except for a model H20, the combination of enhanced book availability and the flexible memory option allow you to concurrently add memory to the system. For more information, see 2.5.5, Book replacement and memory on page 53, and 2.5.6, Flexible Memory Option on page 53. When the total capacity installed has more usable memory than required for a configuration, the licensed internal code configuration control (LICCC) determines how much memory is used from each card. The sum of the LICCC provided memory from each card is the amount available for use in the system.
Memory allocation
Memory assignment or allocation is done at power-on reset (POR) when the system is initialized. PR/SM is responsible for the memory assignments. PR/SM has knowledge of the amount of purchased memory and how it relates to the available physical memory in each of the installed books. PR/SM has control over all physical memory and therefore is able to make physical memory available to the configuration when a book is nondisruptively added. PR/SM also controls the reassignment of the content of a specific physical memory array in one book to a memory array in another book. This is known as the memory copy/reassign function, which is used to reallocate the memory content from the memory in a book to another memory location. It is used when enhanced book availability (EBA) is applied to concurrently remove and re-install a book in case of an upgrade or repair action. Because of the memory allocation algorithm, systems that undergo a number of miscellaneous equipment specification (MES) upgrades for memory can have a variety of memory mixes in all books of the system. If, however unlikely, memory fails, it is technically feasible to power-on reset the system with the remaining memory resources. After power-on reset, the memory distribution across the books is now different, and so is the amount of available memory.
104
8049ch03.fm
It is very desirable to have one's addresses in the TLB. With 4 K pages, holding all the addresses for 1 MB of storage takes 256 TLB lines. When using 1 MB pages, it takes only 1 TLB line. This means that large page size exploiters have a much smaller TLB footprint. Large pages allow the TLB to better represent a large working set and suffer fewer TLB misses by allowing a single TLB entry to cover more address translations. Exploiters of large pages are better represented in the TLB and are expected to see performance improvement in both elapsed time and CPU time. This is because DAT and memory operations are part of CPU busy time even though the CPU waits for memory operations to complete without processing anything else in the meantime. Overhead is associated with creating a 1 MB page. To overcome that overhead, a process has to run for a period of time and maintain frequent memory access to keep the pertinent addresses in the TLB. Very short-running work does not overcome the overhead; short processes with small working sets are expected to receive little or no improvement. Long-running work with high memory-access frequency is the best candidate to benefit from large pages. Long-running work with low memory-access frequency is less likely to maintain its entries in the TLB. However, when it does run, a smaller number of address translations is required to resolve all the memory it needs. So, a very long-running process can benefit somewhat even without frequent memory access. You should weight the benefits of whether something in this category should use large pages as a result of the system-level costs of tying up real storage. There is a balance between the performance of a process using large pages, and the performance of the remaining work on the system. On zEC12, 1 MB large pages become pageable if Flash Express is enabled. They are only available for 64-bit virtual private storage such as virtual memory located above 2 GB. One would be inclined to think, that increasing the TLB size is a feasible option to deal with TLB-miss situations. However, this is not as straightforward as it seems. As the size of the TLB increases, so does the overhead involved in managing the TLBs contents. Correct sizing of the TLB is subject to very complex statistical modelling in order to find the optimal trade-off between size and performance.
105
8049ch03.fm
Storage considerations
Except for z/VM, z/Architecture operating systems do not use expanded storage. Because they operate in 64-bit addressing mode, they can have all the required storage capacity allocated as central storage. z/VM is an exception because, even when operating in 64-bit mode, it can have guest virtual machines running in 31-bit addressing mode, which can use expanded storage. In addition, z/VM exploits expanded storage for its own operations. Defining expanded storage to a coupling facility image is not possible. However, any other image type can have expanded storage defined, even if that image runs a 64-bit operating system and does not use expanded storage. The zEC12 only runs in LPAR mode. Storage is placed into a single storage pool called LPAR single storage pool, which can be dynamically converted to expanded storage and back to central storage as needed when partitions are activated or de-activated.
106
8049ch03.fm
3.6.1 Overview
Logical partitioning (LPAR) is a function implemented by the Processor Resource/Systems Manager (PR/SM) on the zEC12. The zEC12 runs only in LPAR mode. This means that all system aspects are controlled by PR/SM functions. PR/SM is aware of the book structure on the zEC12. Logical partitions, however, do not have this awareness. Logical partitions have resources allocated to them from a variety of physical resources. From a systems standpoint, logical partitions have no control over these physical resources, but the PR/SM functions do. PR/SM manages and optimizes allocation and the dispatching of work on the physical topology. Most physical topology that was previously handled by the operating systems is the responsibility of PR/SM. As shown in 3.4.10, Processor unit assignment on page 102, the initial PU assignment is done during power-on-reset (POR), using rules to optimize cache usage. This is the physical step, where CPs, zIIPs, zAAPs, IFLs, ICFs, and SAPs are allocated on books. When a logical partition is activated, PR/SM builds logical processors and allocates memory for the logical partition. Memory allocation is spread across all books, using a round robin algorithm with three increments per book, to match the number of memory controllers (MCs) per book. This memory allocation design is driven by performance results, also minimizing variability for the majority of workloads. Logical processors are dispatched by PR/SM on physical processors. The assignment topology used by PR/SM to dispatch logical on physical PUs is also based on cache usage optimization. Book level assignments are more important, because this optimizes L4 cache usage. So logical processors from a given logical partition are packed into a book (or books) as much as possible. Then PR/SM optimizes chip assignments within the assigned book (or books), to maximize L3 cache efficiency. So logical processors from a logical partition are dispatched on physical processors on the same PU chip as much as possible. Note that the number of processors per chip (six) matches the number of z/OS processor affinity queues (also six), used by HiperDispatch, achieving optimal cache usage within an affinity node. PR/SM also tries to redispatch a logical processor on the same physical processor to optimize private caches (L1 and L2) usage.
HiperDispatch
PR/SM and z/OS work in tandem to more efficiently use processor resources. HiperDispatch is a function that combines the dispatcher actions and the knowledge that PR/SM has about the topology of the system.
107
8049ch03.fm
Performance can be optimized by redispatching units of work to same processor group, keeping processes running near their cached instructions and data, and minimizing transfers of data ownership among processors/books. The nested topology is returned to z/OS by the Store System Information (STSI) instruction, and HiperDispatch utilizes the information to concentrate logical processors around shared caches (L3 at PU chip level, and L4 at book level), and dynamically optimizes assignment of logical processors and units of work. z/OS dispatcher manages multiple queues, called affinity queues, with a target number of six processors per queue, which fits nicely into a single PU chip. These queues are used to assign work to as few logical processors as are needed for a given logical partition workload. So, even if the logical partition is defined with a large number of logical processors, HiperDispatch optimizes this number of processors nearest to the required capacity. The optimal number of processors to be used are kept within a book boundary where possible.
Logical partitions
PR/SM enables the zEC12 to be initialized for a logically partitioned operation, supporting up to 60 logical partitions. Each logical partition can run its own operating system image in any image mode, independently from the other logical partitions. A logical partition can be added, removed, activated, or deactivated at any time. Changing the number of logical partitions is not disruptive and does not require power-on reset (POR). Certain facilities might not be available to all operating systems, because the facilities might have software co-requisites. Each logical partition has the same resources as a real CPC. Those are processors, memory, and channels: Processors: Called logical processors, they can be defined as CPs, IFLs, ICFs, zAAPs, or zIIPs. They can be dedicated to a logical partition or shared among logical partitions. When shared, a processor weight can be defined to provide the required level of processor resources to a logical partition. Also, the capping option can be turned on, which prevents a logical partition from acquiring more than its defined weight, limiting its processor consumption. Logical partitions for z/OS can have CP, zAAP, and zIIP logical processors. All three logical processor types can be defined as either all dedicated or all shared. The zAAP and zIIP support is available in z/OS. The weight and the number of online logical processors of a logical partition can be dynamically managed by the LPAR CPU Management function of the Intelligent Resource Director (IRD) to achieve the defined goals of this specific partition and of the overall system. The provisioning architecture of the zEC12, described in Chapter 9, System upgrades on page 311, adds another dimension to the dynamic management of logical partitions. For z/OS Workload License Charge (WLC) pricing metric, and metrics based on it such as AWLC, a logical partition defined capacity can be set, enabling the soft capping function. Workload charging introduces the capability to pay software license fees based on the processor utilization of the logical partition on which the product is running, rather than on the total capacity of the system, as follows: In support of WLC, the user can specify a defined capacity in millions of service units (MSUs) per hour. The defined capacity sets the capacity of an individual logical partition when soft capping is selected. The defined capacity value is specified on the Options tab on the Customize Image Profiles panel 108
IBM zEnterprise EC12 Technical Guide
8049ch03.fm
WLM keeps a 4-hour rolling average of the CPU usage of the logical partition, and when the 4-hour average CPU consumption exceeds the defined capacity limit, WLM dynamically activates LPAR capping (soft capping). When the rolling 4-hour average returns below the defined capacity, the soft cap is removed. For more information regarding WLM, see System Programmer's Guide to: Workload Manager, SG24-6472. For a review of software licensing see 8.12, Software licensing considerations on page 305. Weight settings: When defined capacity is used to define an uncapped logical partitions capacity, looking carefully at the weight settings of that logical partition is important. If the weight is much smaller than the defined capacity, PR/SM will use a discontinuous cap pattern to achieve the defined capacity setting. This means PR/SM will alternate between capping the LPAR at the MSU value corresponding to the relative weight settings, and no capping at all. It is best to avoid this case, and try to establish a defined capacity which is equal or close to the relative weight. Memory: Memory, either central storage or expanded storage, must be dedicated to a logical partition. The defined storage must be available during the logical partition activation. Otherwise, the activation fails.
Modes of operation
Table 3-5 shows the nodes of operation, summarizing all available mode combinations: operating modes and their processor types, operating systems, and addressing modes. Only the currently in-support versions of operating systems are considered.
109
8049ch03.fm
Table 3-5 zEC12 modes of operation Image mode ESA/390 PU type CP and zAAP/zIIP CP CP ESA/390 TPF Coupling facility LINUX-only CP only ICF or CP IFL or CP Operating system z/OS z/VM z/VSE and Linux on System z (64-bit) Linux on System z (31-bit) z/TPF CFCC Linux on System z (64-bit) z/VM Linux on System z (31-bit) z/VM zAware CP, IFL, zIIP, zAAP, ICF IFL or CP z/VM zAware 31-bit 64-bit 64-bit Addressing mode 64-bit 64-bit 31-bit 64-bit 64-bit 64-bit
The 64-bit z/Architecture mode has no special operating mode because the architecture mode is not an attribute of the definable images operating mode. The 64-bit operating systems are IPLed in 31-bit mode and, optionally, can change to 64-bit mode during their initialization. The operating system is responsible for taking advantage of the addressing capabilities provided by the architectural mode. For information about operating system support, see Chapter 8, Software support on page 243.
8049ch03.fm
Dedicated or shared CPs Dedicated or shared ICFs Linux-only mode, to run these systems: A Linux on System z operating system, on either: Dedicated or shared IFLs Dedicated or shared CPs Dedicated or shared IFLs Dedicated or shared CPs
z/VM-mode to run z/VM on dedicated or shared CPs or IFLs, plus zAAPs, zIIPs, and ICFs. zAware mode, to run by loading the zAware code into the logical partition defined as: Dedicated or shared CPs Dedicated or shared IFLs Table 3-6 shows all LPAR modes, required characterized PUs, operating systems, and the PU characterizations that can be configured to a logical partition image. The available combinations of dedicated (DED) and shared (SHR) processors are also shown. For all combinations, a logical partition can also have reserved processors defined, allowing nondisruptive logical partition upgrades.
Table 3-6 LPAR mode and PU usage LPAR mode ESA/390 PU type CPs Operating systems z/Architecture operating systems ESA/390 operating systems Linux on System z z/OS z/VM (V5R4 and later for guest exploitation) z/TPF CFCC Linux on System z z/VM z/VM (V5R4 and later) PUs usage CPs DED or CPs SHR
CPs and zAAPs or zIIPs ESA/390 TPF Coupling facility LINUX only z/VM CPs ICFs or CPs IFLs or CPs CPs, IFLs, zAAPs, zIIPs, ICFs IFLs, or CPs
CPs DED and zAAPs DED, and (or) zIIPs DED or CPs SHR and zAAPs SHR or zIIPs SHR CPs DED or CPs SHR ICFs DED or ICFs SHR, or CPs DED or CPs SHR IFLs DED or IFLs SHR, or CPs DED or CPs SHR All PUs must be SHR or DED
zAware
zAware
111
8049ch03.fm
commands using the Hardware Configuration Definition (HCD). At the same time, required channels have to be defined for the new logical partition. Partition profile: Cryptographic coprocessors are not tied to partition numbers or MIF IDs. They are set up with Adjunct Processor (AP) numbers and domain indices. These are assigned to a partition profile of a given name. The customer assigns these AP and domains to the partitions and continues to have the responsibility to clear them out when their profiles change.
112
8049ch03.fm
Before activating a logical partition, central storage (and, optionally, expanded storage) must be defined to the logical partition. All installed storage can be configured as central storage. Each individual logical partition can be defined with a maximum of 1 TB of central storage. Central storage can be dynamically assigned to expanded storage and back to central storage as needed without a power-on reset (POR). For details, see LPAR single storage pool on page 106. Memory cannot be shared between system images. It is possible to dynamically reallocate storage resources for z/Architecture logical partitions running operating systems that support dynamic storage reconfiguration (DSR). This is supported by z/OS, and z/VM V5R4 and later releases. z/VM in turn virtualizes this support to its guests. For details, see 3.6.5, LPAR dynamic storage reconfiguration on page 116. Operating systems running under z/VM can exploit the z/VM capability of implementing virtual memory to guest virtual machines. The z/VM dedicated real storage can be shared between guest operating systems. Table 3-7 shows the zEC12 storage allocation and usage possibilities, depending on the image mode.
Table 3-7 Storage definition and usage possibilities Image mode Architecture mode (addressability) Maximum central storage Architecture 16 EB 2 GB 2 GB 1.5 TB 16 EB 2 GB 16 EB 16 EB zEC12 definition 1 TB 128 GB 2 GB 1 TB 256 GB 2 GB 256 GB 1 TB Expanded storage zEC12 definable Yes Yes Yes No Yes Yes Yes Yes Operating system usagea Yes Yes No No Only by z/VM Only by z/VM Yes No
ESA/390
z/VMb zAware
a. z/VM supports the use of expanded storage. b. z/VM-mode is supported by z/VM V5R4 and later.
ESA/390 mode
In ESA/390 mode, storage addressing can be 31 or 64 bits, depending on the operating system architecture and the operating system configuration. An ESA/390 mode image is always initiated in 31-bit addressing mode. During its initialization, a z/Architecture operating system can change it to 64-bit addressing mode and operate in the z/Architecture mode. Certain z/Architecture operating systems, such as z/OS, always change the 31-bit addressing mode and operate in 64-bit mode. Other z/Architecture operating systems, such as z/VM, can be configured to change to 64-bit mode or to stay in 31-bit mode and operate in the ESA/390 architecture mode.
113
8049ch03.fm
The following modes are provided: z/Architecture mode: In z/Architecture mode, storage addressing is 64-bit, allowing for virtual addresses up to 16 exabytes (16 EB). The 64-bit architecture theoretically allows a maximum of 16 EB to be used as central storage. However, the current central storage limit for logical partitions is 1 TB of central storage. The operating system that runs in z/Architecture mode has to be able to support the real storage. Currently, z/OS for example, supports up to 4 TB of real storage (z/OS V1R10 and higher releases). Expanded storage can also be configured to an image running an operating system in z/Architecture mode. However, only z/VM is able to use expanded storage. Any other operating system running in z/Architecture mode (such as a z/OS or a Linux on System z image) does not address the configured expanded storage. This expanded storage remains configured to this image and is unused. ESA/390 architecture mode: In ESA/390 architecture mode, storage addressing is 31-bit, allowing for virtual addresses up to 2 GB. A maximum of 2 GB can be used for central storage. Because the processor storage can be configured as central and expanded storage, memory above 2 GB can be configured as expanded storage. In addition, this mode permits the use of either 24-bit or 31-bit addressing, under program control. Because an ESA/390 mode image can be defined with up to 128 GB of central storage, the central storage above 2 GB is not used, but remains configured to this image. Storage resources: Either a z/Architecture mode or an ESA/390 architecture mode operating system can run in an ESA/390 image on a zEC12. Any ESA/390 image can be defined with more than 2 GB of central storage and can have expanded storage. These options allow you to configure more storage resources than the operating system is capable of addressing.
Linux-only mode
In Linux-only mode, storage addressing can be 31-bit or 64-bit, depending on the operating system architecture and the operating system configuration, in exactly the same way as in ESA/390 mode.
114
8049ch03.fm
Only Linux and z/VM operating systems can run in Linux-only mode. Linux on System z 64-bit distributions (SUSE SLES 10 and later, Red Hat RHEL 5 and later) use 64-bit addressing and operate in the z/Architecture mode. z/VM also uses 64-bit addressing and operates in the z/Architecture mode.
z/VM mode
In z/VM mode, certain types of processor units can be defined within one LPAR. This increases flexibility and simplifies systems management by allowing z/VM to perform the following tasks all in the same z/VM LPAR: Manage guests to operate Linux on System z on IFLs. Operate z/VSE and z/OS on CPs. Offload z/OS system software overhead, such as DB2 workloads on zIIPs. Provide an economical Java execution environment under z/OS on zAAPs or on zIIPs.
zAware mode
In IBM zAware mode, storage addressing is 64-bit for a IBM zAware image running IBM System z Advanced Workload Analysis Reporter firmware, allowing for an addressing range up to 16 EB. However, the current zEC12 definition limit for logical partitions is 1 TB of storage. IBM zAware feature, which is exclusive to zEC12, allows the following capabilities: Help in detecting and diagnosing unusual behavior of z/OS images in near real time. Reduces problem determination time and improve service availability beyond standard z/OS features. Provides an easy to use graphical user interface with quick drill-down capabilities to view analytical data about z/OS behavior. For details, see Appendix A, IBM zAware on page 429. Only IBM zAware firmware can run in IBM zAware mode.
115
8049ch03.fm
Any unused available storage Another partition that has released storage A memory upgrade. A concurrent logical partition storage upgrade uses dynamic storage reconfiguration (DSR). z/OS uses the reconfigurable storage unit (RSU) definition to add or remove storage units in a nondisruptive way. z/VM V5R4 and later releases support the dynamic addition of memory to a running logical partition by using reserved storage, and also virtualizes this support to its guests. Removal of storage from the guests or z/VM is disruptive. SUSE Linux Enterprise Server (SLES) 11 supports both concurrent add and remove.
The granularity applies across all central storage defined, both initial and reserved. For example, for a logical partition with an initial storage amount of 30 GB and a reserved storage amount of 48 GB, the central storage granularity of both initial and reserved central storage is 256 MB. Expanded storage granularity is fixed at 256 MB. Logical partition storage granularity information is required for logical partition image setup and for z/OS Reconfigurable Storage Units (RSU) definition. Logical partitions are limited to a maximum size of 1 TB of central storage. For z/VM V5R4 and later the limitation is 256 GB.
116
8049ch03.fm
PR/SM dynamically takes offline a storage increment and makes it available to other partitions when an operating system running in a logical partition releases a storage increment.
LPAR Clust er
z/OS
z/OS
z/OS
Linux
S Y S P L E X
z/OS
CF
System z
Figure 3-10 IRD LPAR cluster example
IRD addresses three separate but mutually supportive functions: LPAR CPU management: WLM dynamically adjusts the number of logical processors within a logical partition and the processor weight based on the WLM policy. The ability to move the CPU weights across an LPAR cluster provides processing power to where it is most needed, based on WLM goal mode policy.
117
8049ch03.fm
This function is automatically deactivated when HiperDispatch is active. HiperDispatch was introduced in 3.6, Logical partitioning on page 107. HiperDispatch manages the number of logical CPs in use. It adjusts the number of logical processors within a logical partition in order to achieve the optimal balance between CP resources and the requirements of the workload in the logical partition. HiperDispatch also adjusts the number of logical processors. The goal is to map the logical processor to as few physical processors as possible, attempting use the processor resources more efficiently by trying to stay within the local cache structure, making efficient use of the advantages of the high-frequency microprocessors and improving throughput and response times. Dynamic channel path management (DCM): DCM moves FICON channel bandwidth between disk control units to address current processing needs. The zEC12 supports DCM within a channel subsystem. Channel subsystem priority queuing: This function on the zEC12 and System z allows the priority queuing of I/O requests in the channel subsystem and the specification of relative priority among logical partitions. WLM, executing in goal mode, sets the priority for a logical partition and coordinates this activity among clustered logical partitions. For information about implementing LPAR CPU management under IRD, see z/OS Intelligent Resource Director, SG24-5952.
118
8049ch03.fm
z10 EC
CF01
ICF
zEC12
z/OS PSIFB or IC
PS
B IF
PS
IS
* -3
e (p
IF B
IS
C -3
z196
(p ee r)
Sysplex LPARs
z/OS
Sysplex LPARs
CF02 ICF
Disks
Figure 3-11 Sysplex hardware overview
Disks
Disks
Figure 3-11 on page 119 shows a zEC12 system containing multiple z/OS sysplex partitions and an internal coupling facility (CF02), a z10 EC system containing a stand-alone CF (CF01), and a z196 containing multiple z/OS sysplex partitions. STP over coupling links provides time synchronization to all systems. Appropriate CF link technology (1x IFB or 12x IFB) selection depends on the system configuration and how distant they are physically located. ISC-3 links can be carried forward to zEC12 only when upgrading from either z196 or z10 EC. The ICB-4 coupling link is not supported on both zEC12 and zEnterprise CPCs. Link technologies are described in 4.10.1, Coupling links on page 161. Parallel Sysplex technology is an enabling technology, allowing highly reliable, redundant, and robust System z technology to achieve near-continuous availability. A Parallel Sysplex comprises one or more (z/OS) operating system images coupled through one or more Coupling Facilities. The images can be combined together to form clusters. A properly configured Parallel Sysplex cluster maximizes availability, as follows: Continuous (application) availability: Changes can be introduced, such as software upgrades, one image at a time, while the remaining images continue to process work. For details, see Parallel Sysplex Application Considerations, SG24-6523. High capacity: Scales can be from 2 to 32 images. Dynamic workload balancing: Viewed as a single logical resource, work can be directed to any similar operating system image in a Parallel Sysplex cluster having available capacity. Systems management: Architecture provides the infrastructure to satisfy customer requirements for continuous availability, and provides techniques for achieving simplified systems management consistent with this requirement.
119
8049ch03.fm
Resource sharing: A number of base (z/OS) components exploit coupling facility shared storage. This exploitation enables sharing of physical resources with significant improvements in cost, performance, and simplified systems management. Single system image: The collection of system images in the Parallel Sysplex appears as a single entity to the operator, the user, the database administrator, and so on. A single system image ensures reduced complexity from both operational and definition perspectives. N-2 support Multiple hardware generations (normally three) are supported in the same Parallel Sysplex. This provides for a gradual evolution of the systems in the Sysplex, without forcing changing all simultaneously. Similarly, software support for multiple releases or versions is supported. Through state-of-the-art cluster technology, the power of multiple images can be harnessed to work in concert on common workloads. The System z Parallel Sysplex cluster takes the commercial strengths of the platform to improved levels of system management, competitive price for performance, scalable growth, and continuous availability.
Performance improvements
CFCC Level 18 introduces improvements in cache structure management: Dynamic structure size alter is enhanced to improve the performance of changing cache structure size. DB2 Global Buffer Pool (GBP) write-around (cache bypass) supports a new conditional write to GBP command. DB2 can use this enhancement during batch update/insert processing to intelligently decide which entries should be written to the GBP cache, and which should just be written around the cache to disk. Before this enhancement, overrunning cache structures with useless directory entries and changed data during batch update/insert jobs (e.g. reorganizations) caused several issues such as CF overhead, thrashing the cache through LRU processing, and leaded to castout processing backlogs and delays. CF castout class contention avoidance reduces latch contention with more granular class assignments. CF storage class contention avoidance improves response time by changing the latching from a suspend lock to a spin lock. CF Engine performance is improved by more efficient use of shared-processor CF images with good service times; as well as latency reduction for asynchronous CF operations and asynchronous CF notifications.
8049ch03.fm
Added to the Subchannel Activity and CF To CF Activity sections of the RMF Postprocessor Coupling Facility Activity report Provided on the Subchannels Details panel of the RMF Monitor III Coupling Facility Systems report.
Serviceability enhancements
Serviceability enhancements provides help in debugging problems. Additional structure control info in CF dumps Prior to CFCC Level 18, only CF control structures were dumped and no structure-related controls were included. With CFCC Level 18, new structure control info included in CF dumps, though data elements (customer data) are still not dumped Enhanced CFCC tracing support CFCC Level 18 has significantly enhanced trace points, especially in areas like latching, suspend queue management/dispatching, duplexing protocols and sublist notification. Enhanced Triggers for CF non-disruptive dumping for soft-failure cases beyond break-duplexing The Coupling Facility Control Code (CFCC), the CF Operating System, is implemented using the active wait technique. This technique means that the CFCC is always running (processing or searching for service) and never enters a wait state. This also means that the CF Control Code uses all the processor capacity (cycles) available for the coupling facility logical partition. If the LPAR running the CFCC has only dedicated processors (CPs or ICFs), then using all processor capacity (cycles) is not a problem. However, this can be an issue if the LPAR that is running the CFCC also has shared processors. Therefore, it is best to enable dynamic dispatching on the CF LPAR. CF structure sizing changes are expected when going from CFCC Level 17 (or below) to CFCC Level 18. We suggest reviewing the CF LPAR size by using the CFSizer tool available at the following web page: http://www.ibm.com/systems/z/cfsizer/
121
8049ch03.fm
ICF
CP
Partition Image Profile
With Dynamic CF dispatching enabled, backup CF becomes active only for new work
ACTIVE CF
z/OS
CF
z/OS
BACK UP CF
z/OS
CF
z/OS
HMC
Setup
DYNDISP ON (CFCC cmd) IMAGE Profile setup
Function
For additional details regarding CF configurations, see Coupling Facility Configuration Options, GF22-5042, also available from the Parallel Sysplex website: http://www.ibm.com/systems/z/advantages/pso/index.html
122
8049ch04.fm
Chapter 4.
123
8049ch04.fm
124
8049ch04.fm
4.1.4 PCIe
PCIe is a serial bus with embedded clock and uses 8b/10b encoding, where every eight bits are encoded into a 10-bit symbol that is then decoded at the receiver. Thus, the bus needs to transfer 10 bits to send 8 bits of actual usable data. A PCIe bus generation 2 single lane can transfer 5 Gbps of raw data (duplex connection), which is 10 Gbps of raw data. From these 10 Gbps, only 8 Gbps are actual data (payload). Therefore an x16 (16 lanes) PCIe gen2 bus transfers 160 Gbps encoded, which is 128 Gbps of unencoded data (payload). This is 20 GBps raw data and 16 GBps of encoded data. The new measuring unit for transfer rates for PCIe is GT/s (Giga Transfers per second) which refers to the raw data (even though only 80% of this transfer is actual payload data). The translation between GT/s to GBps is: 5 GT/s equals 20 GBps or 1 GT/s equals 4 GBps. The 16 lanes of the PCIe bus are virtual lanes, always consisting of one transmit and one receive lane. Each of these lanes consists of two physical copper wires. The physical method used to transmit signals is a differential bus, which means that the signal is encoded into the different voltage levels between two wires (as opposed to one voltage level on one wire in
1
125
8049ch04.fm
comparison to the ground signal). Therefore, each of the 16 PCIe lanes uses actually four copper wires for the signal transmissions.
4.2.1 Characteristics
The zEC12 I/O subsystem design provides great flexibility, high availability, and excellent performance characteristics, as follows: High bandwidth: The zEC12 uses PCIe as an internal interconnect protocol to drive PCIe I/O drawers. The I/O bus infrastructure data rate increases up to 8 GBps. The zEC12 uses InfiniBand as the internal interconnect protocol to drive I/O cages and I/O drawers and CPC to CPC connection. InfiniBand supports I/O bus infrastructure data rate up to 6 GBps. Connectivity options: The zEC12 can be connected to a range of interfaces such as FICON/Fibre Channel Protocol for storage area network connectivity, and for CPC to CPC connectivity by FICON CTCs (FCTC), 10 Gigabit Ethernet, Gigabit Ethernet, and 1000BASE-T Ethernet for local area network connectivity. For CPC to CPC connection, zEC12 uses Parallel Sysplex InfiniBand (PSIFB), ISC-3 coupling links. Concurrent I/O upgrade: You can concurrently add I/O cards to the server if an unused I/O slot position is available. Concurrent PCIe I/O drawer upgrade: Additional PCIe I/O drawers can be installed concurrently without preplanning if there is free space in one of the frames. Dynamic I/O configuration: Dynamic I/O configuration supports the dynamic addition, removal, or modification of channel path, control units, and I/O devices without a planned outage. Pluggable optics: The FICON Express8S, FICON Express8, and FICON Express4 features have Small Form Factor Pluggable (SFP) optics to permit each channel to be individually serviced in the event of a fiber optic module failure. The traffic on the other channels on the same feature can continue to flow if a channel requires servicing. Concurrent I/O card maintenance: Every I/O card plugged in an I/O cage, I/O drawer, or PCIe I/O drawer supports concurrent card replacement in case of a repair action.
8049ch04.fm
Up to 176 FICON Express8 channels Up to 320 FICON Express8S channels Up to 96 OSA-Express3 ports Up to 96 OSA-Express4S ports Up to 48 ISC-3 coupling links Up to 16 InfiniBand fanouts: Up to 32 12x InfiniBand coupling links with HCA2-O fanout, or Up to 32 1x InfiniBand coupling links with HCA2-O LR (1xIFB) fanout, or Up to 32 12x InfiniBand coupling links with HCA3-O fanout, or Up to 64 1x InfiniBand coupling links with HCA3-O LR (1xIFB) fanout Coupling links: The maximum number of external coupling links combined (ISC-3 and IFB coupling links) cannot exceed 112 for each zEC12.
Figure 4-2 on page 128 illustrates the I/O structure of an I/O cage. An InfiniBand (IFB) cable connects the HCA2-C fanout to an IFB-MP card in the I/O cage. The passive connection between two IFB-MP cards allows for redundant I/O interconnection. The IFB cable between
127
8049ch04.fm
an HCA2-C fanout in a book and each IFB-MP card in the I/O cage supports a 6 GBps bandwidth.
Processor
Memory
HCA2-C Fanout
IFB-MP
IFB-MP
Crypto
Crypto
OSA
ISC
OSA
I/O Cage
Note: Only one I/O cage is supported in zEC12, carry forward only. The I/O cage domains and their related I/O slots are shown in Figure 4-3 on page 129
128
ISC
8049ch04.fm
RII
B W I R N G D E
25E 26F 27E 28F 17G 18G 19G 20G
IFB-MP
5A 6B 7A 8B 9C 10D 11C 12D
G
21E 22F 23E 24F
C
13C 14D 15C 16D
Front 16
Rear 12
Each I/O domain supports up to four I/O cards (FICON, OSA, Crypto, or ISC). All I/O cards are connected to the IFB-MP cards through the backplane board. Table 4-1 lists the I/O domains and their related I/O slots.
Table 4-1 I/O domains of I/O cage Domain number (name) 0 (A) 1 (B) 2 (C) 3 (D) 4 (E) 5 (F) 6 (G) I/O slot in domain 01, 03, 06, 08 02, 04, 07, 09 10, 12, 15, 17 11, 13, 16, 18 19, 21, 24, 26 20, 22, 25, 27 29, 30, 31, 32
129
8049ch04.fm
Front
L G0 1
J00
L G0 2
L G03 L G04
J01
L G05
I/O cards
Rear
LG06
J00
DCA
IFB-MP
LG10 LG11
The I/O structure in a zEC12 server is illustrated in Figure 4-5 on page 131. An IFB cable connects the HCA fanout card to an IFB-MP card in the I/O drawer. The passive connection between two IFB-MP cards allows redundant I/O interconnection (RII). This provides connectivity between an HCA fanout card, and I/O cards in case of concurrent fanout card or IFB cable replacement. The IFB cable between an HCA fanout card and each IFB-MP card supports a 6 GBps link rate.
130
8049ch04.fm
Processor
Memory
HCA2-C Fanout
IFB-MP
IFB-MP
I/O Drawer
Note: Only two I/O drawers are supported in zEC12, and as carry forward only. The I/O drawer domains and their related I/O slots are shown in Figure 4-6. The IFB-MP cards are installed at slot 09 at the rear side of the I/O drawer. The I/O cards are installed from the front and rear side of the I/O drawer. Two I/O domains (A and B) are supported. Each I/O domain has up to four I/O feature cards (FICON, OSA, Crypto, or ISC). The I/O cards are connected to the IFB-MP card through the backplane boar.
1 2 3 4 5
DCA 1 I/O-A
RII
6 7 8 9 10 11
IFB-MP B IFB-MP A
I/O-A I/O-B
Front - 4
Rear - 4
131
8049ch04.fm
Each I/O domain supports four I/O card slots. Balancing I/O cards across both I/O domains on new build servers, or on upgrades, is automatically done when the order is placed. Table 4-2 lists the I/O domains and their related I/O slots.
Table 4-2 I/O domains of I/O drawer Domain A B I/O slot in domain 02, 05, 08, 10 03, 04, 07, 11
AMD
Front
7U (~311 mm)
FICON Express8S
OSA-Express4S
Rear
DCA
560 mm (max)
132
8049ch04.fm
The I/O structure in a zEC12 server is illustrated in Figure 4-8. The PCIe switch card provides the fanout from the high speed x16 PCIe host bus to eight individual card slots. The PCIe switch card is connected to the processor nest through a single x16 PCIe Gen 2 bus from a PCIe fanout card, which converts the book internal bus into two PCIe buses. A switch card in the front is connected to a switch card in the rear through the PCIe I/O drawer board to provide a failover capability in case of a PCIe fanout card failure, book upgrade, and so on. In PCIe I/O drawer, the 8 I/O cards directly attached to the switch card constitute an I/O domain. The PCIe I/O drawer supports concurrent add and delete to enable a client to increase I/O capability as needed without plan ahead.
B oo k 0 Memo ry
PU PU PU
Bo ok 1 Memo ry
PU PU PU
SC1, S C0 (FBC)
PU PU PU
PC Ie (8x)
x16 P CI e Gen2 8 GBps
PCIe (8x)
PC Ie s witc h
PCIe switc h
P CIe switc h
PC Ie s witc h
F ICON Express8 S
OSA-Express4S 10 GbE
The PCIe I/O Drawer supports up to 32 I/O cards. They are organized in four hardware domains per drawer. Each domain is driven through a PCIe switch card. Always two PCIe switch cards provide a backup path for each other through the passive connection in the PCIe I/O Drawer backplane, so that in case of a PCIe fanout card or cable failure, all 16 I/O cards in the two domains can be driven through a single PCIe switch card. To support Redundant I/O Interconnect (RII) between front to back domain pairs 0-1 and 2-3 the two interconnects to each pair must be from two different PCIe fanouts (all four domains in one of these cages can be activated with two fanouts). The flexible service processors (FSPs) are used for system control. The PCIe I/O drawer domains and their related I/O slots are shown in Figure 4-9 on page 134.
133
8049ch04.fm
0 0 0 0
PCIe interconnect
PCIe interconnect
1 1 1 1 1 1 1 1
38 37 36 35 34 33 32 31 30 29 28 27 26 25 24 23 22 21 20
RII
FSP-1, 1
0 0 0 0 2 2 2 2
FSP-1, 2
PCIe interconnect
PCIe interconnect
3 3 3 3
2 2 2 2 Front 16
3 3 3 3 Rear 16
Each I/O domain supports up to eight I/O cards (FICON, OSA, Crypto, and Flash Express). All I/O cards are connected to the PCIe switch card through the backplane board. Table 4-3 lists the I/O domains and slots.
Table 4-3 I/O domains of PCIe I/O drawer Domain 0 1 2 3 I/O slot in domain 01, 02, 03, 04, 06, 07, 08, 09 30, 31, 32, 33, 35, 36, 37, 38 11, 12, 13, 14, 16, 17, 18, 19 20, 21, 22, 23, 25, 26, 27, 28
Power Sequence Controller (PSC) feature is not supported on PCIe drawer and zEC12.
4.6 I/O cage, I/O drawer and PCIe I/O drawer offerings
A maximum of five PCIe drawers can be installed supporting up to 160 PCIe I/O features. Customers will not order I/O cages or I/O drawers, they will order I/O features and the configurator will determine the proper mix of I/O cage, I/O drawer, and PCIe I/O drawer.
134
8049ch04.fm
Note: On new build zEC12 only PCIe I/O drawers are supported.
A maximum of 44 I/O features can be carried forward. Table 4-5 list the number and mix of I/O cages and I/O drawers depending on amount of legacy I/O features.
Table 4-5 Number and mix of I/O cages and I/O drawers Number of I/O cards carried forward on upgrades 0 1-8 9-16 17-28 29-36 Number of I/O cages 0 0 0 1 1 Number of I/O drawers 0 1 2 0 1
135
8049ch04.fm
4.7 Fanouts
The zEC12 server uses fanout cards to connect the I/O hardware Subsystem to the processor Books and to provide the InfiniBand coupling links for Parallel Sysplex as well. All Fanout cards support concurrent add, delete and move. zEC12 supports two different internal I/O infrastructures for the internal connection. zEC12 uses InfiniBand based infrastructure for the internal connection to I/O cages and I/O drawers and uses PCIe based infrastructure for PCIe I/O drawers in which the cards for the connection to peripheral devices and networks reside. The InfiniBand and PCIe fanouts are located in the front of each book. Each book has eight fanout slots. They are named D1 to DA, top to bottom; slots D3 and D4 are not used for fanouts. Six types of fanout cards are supported by zEC12. Each slot holds one of the following six fanouts: Host Channel Adapter (HCA2-C): This copper fanout provides connectivity to the IFB-MP card in the I/O cage and I/O drawer. PCIe Fanout: This copper fanout provides connectivity to the PCIe switch card in the PCIe I/O drawer. Host Channel Adapter (HCA2-O (12xIFB)): This optical fanout provides 12x InfiniBand coupling link connectivity up to 150 meters distance to a zEC12, z196, z114, and System z10. Host Channel Adapter (HCA2-O LR (1xIFB)): This optical long range fanout provides 1x InfiniBand coupling link connectivity up to 10 km unrepeated distance to a zEC12, z196, z114 and System z10 servers. Host Channel Adapter (HCA3-O (12xIFB)): This optical fanout provides 12x InfiniBand coupling link connectivity up to 150 meters distance to a zEC12, z196, z114 and System z10. Host Channel Adapter (HCA3-O LR (1xIFB)): This optical long range fanout provides 1x InfiniBand coupling link connectivity up to 10 km unrepeated distance to a zEC12, z196, z114 and System z10 servers. The HCA3-O LR (1xIFB) fanout comes with 4 ports and each other fanout comes with two ports. Figure 4-10 on page 137 illustrates the IFB connection from the CPC cage to an I/O cage and an I/O drawer, and the PCIe connection from the CPC cage to a PCIe I/O drawer.
136
8049ch04.fm
Book 0 Memory
PU PU PU PU PU PU
Book 1 Me mory
PU PU PU PU PU PU
Book 2 Memory
PU PU PU PU PU PU
Book 3 Me mory
PU PU PU PU PU PU
Fanouts
PU PU
PU PU
PU PU
Fanouts
PCIe (8x )
x16 PCIe Gen2 8 GBps RII
PCIe (8x)
HCA2 (8x)
6 GBps
HCA2 (8x)
PCIe switch
PCIe switch
PCIe switch
RII
PCIe switch
IFB-MP RII
IFB-MP
IFB-MP
RII
IFB-MP
2 GBpsmSTI
Ports O SA-Express3
12x Infini Band Up to 150 meters 1x Infini Band Up to 10/ 100KM 12x Infini Band Up to 150 meters .. .. .. .. 1x InfiniBand
H CA 3 -O
H CA3 -O LR
Up to 10/100KM
.. .. .. ..
Fanouts
IFB-MP
ISC-3
Up to 10/100 km
E C12, z196 z114, z10 EC, z10 BC
H CA2 -C
F PGA
2GBps mSTI
137
8049ch04.fm
8049ch04.fm
be defined as shared between images within a channel subsystem and they can be also be spanned across multiple CSSs in a server. Each HCA2-O (12xIFB) fanout used for coupling links has an assigned adapter ID (AID) number that must be used for definitions in IOCDS to create a relationship between the physical fanout location and the CHPID number. For details about AID numbering, see Adapter ID number assignment on page 142. For detailed information about how the AID is used and referenced in HCD, see Implementing and Managing InfiniBand Coupling Links on System z SG24-7539. When STP is enabled, IFB coupling links can be defined as timing-only links to other zEC12, z196, z114, and System z10 servers.
139
8049ch04.fm
When STP is enabled, IFB LR coupling links can be defined as timing-only links to other zEC12, z196, z114, and System z10 servers.
140
8049ch04.fm
System z10 servers, or to connect to a coupling port in the same server by using a fiber cable. Up to 16 HCA3-O LR (1xIFB) fanouts are supported by zEC12 and provide up to 64 ports for coupling links. The HCA-O LR fanout supports InfiniBand 1x optical links that offer longer distance of coupling links. The cable has one lane containing two fibers; one fiber is used for transmitting and one fiber is used for receiving data. Each connection supports a link rate of 5 Gbps if connected to a zEC12, z196, a z114, a z10 server, or to a System z qualified DWDM, and a data link rate of 2.5 Gbps when connected to a System z qualified DWDM. The link rate is auto- negotiated to the highest common rate. HCA3-O LR (1xIFB) fanout: Ports on the HCA3-O LR (1xIFB) fanout are used exclusively for coupling links and cannot be used or shared for any other purpose The fiber optic cables are 9 m single mode (SM) optical cables terminated with an LC Duplex connector. The maximum unrepeated distance is 10 km and up to 100 km with System z qualified DWDM.
141
8049ch04.fm
ECF
OSC
ECF
ECF
OSC D1 D2 D3 D4 D5 D6 D7 D8 D9 DA
ECF
filler
filler
filler
filler
filler
FSP FSP
H20
D4 D5 D6 D7 D8 D9 DA
H43
D4 D5 D6 D7 D8 D9 DA
01
06
10
15
01
06
10
15
ECF
OSC
ECF
ECF
OSC
OSC
ECF
filler
D3 D4 D5 D6 D7 D8 D9 DA
FSP FSP
D3 D4 D5 D6 D7 D8 D9 DA
FSP FSP
H66
D4 D5 D6 D7 D8 D9 DA
H89 H89
D3 D4 D5 D6 D7 D8 D9 DA
FSP FSP
D3 D4 D5 D6 D7 D8 D9 DA
FSP FSP
D3
FSP
D3
FSP
D4 FSP D5 D6 D7 D8 D9 DA
D4 FSP D5 D6 D7 D8 D9 DA
01
06
10
15
01
06
10
15
Fanout slots
The fanout slots are numbered D1 to DA top to bottom, as shown in Table 4-7 on page 143. All fanout locations and their AIDs for all four books are shown in the table for reference only. Fanouts in locations D1 and D2 are not available on all models. Slots D3 and D4 will never have a fanout installed (dedicated for FSPs).
142
8049ch04.fm
Attention: Slots D1 and D2 are not used in a 4-book system, and only partially in a 3-book system.
Table 4-7 Fanout AID numbers Fanout location D1 D2 D3 D4 D5 D6 D7 D8 D9 DA Fourth book 00 01 02 03 04 05 06 07 First book 08 09 0A 0B 0C 0D 0E 0F Third book 10 11 12 13 14 15 16 17 Second book 18 19 1A 1B 1C 1D 1E 1F
Important: The AID numbers in Table 4-7 are valid only for a new build system or for new books added. If a fanout is moved, the AID follows the fanout to its new physical location. The AID assigned to a fanout is found in the PCHID REPORT provided for each new server or for MES upgrade on existing servers. Example 4-1 shows part of a report, named PCHID REPORT, for a model M32. In this example, one fanout is installed in the first book (location 06) and one fanout is installed in the second book (location 15), both in location D5. The assigned AID for the fanout in the first book is 0A; the AID assigned to the fanout in the second book is 1A.
Example 4-1 AID assignment in PCHID report
CHPIDSTART 12345675 PCHID Machine: xxxx-H43 SNXXXXXXX - - - - - - - - - - - - - - - - - Source Cage Slot F/C 06/D5 A25B D506 0163 15/D5 A25B D515 0163
REPORT
Jun xx,2012
143
8049ch04.fm
Table 4-8 Fanout summary Fanout feature HCA2-C Feature code 0162 Use Connect to I/O cage or I/O drawer Coupling link Coupling link Connect to PCIe I/O drawer Coupling link Coupling link Cable type Copper Connector type n/a Max. distance 3.5 m Link data rate 6 GBps
0163
MPO
150 m
6 GBps
0168 0169
LC Duplex n/a
10 kma 3m
0171
MPO
150 m
6 GBpsc
0170
LC Duplex
10 kma
a. Up to 100 km with repeaters (System z qualified DWDM) b. Auto-negotiated, depending on DWDM equipment c. When using the 12x IFB3 protocol, synchronous service times are 40% faster than when using the 12x IFB protocol.
144
8049ch04.fm
Channel feature FICON Express8S SX OSA-Express3 GbE LX OSA-Express4S GbE LX OSA-Express3 GbE SX OSA-Express4S GbE SX OSA-Express3 1000BASE-T Ethernet OSA-Express4S 1000BASE-T Ethernet OSA-Express3 10 GbE LR OSA-Express4S 10 GbE LR OSA-Express3 10 GbE SR OSA-Express4S 10 GbE SR ISC-3 ISC-3 up to 20 kma HCA2-O (12xIFB) HCA2-O LR (1xIFB) HCA3-O (12xIFB) HCA3-O LR (1xIFB) Crypto Express3 Crypto Express4S Flash Express
Feature code 0410 3362 0404 3363 0405 3367 0408 3370 0406 3371 0407 0217 (ISC-M) 0218 (ISC-D) RPQ 8P2197 (ISC-D) 0163 0168 0171 0170 0864 0865 0402
New build Y N Y N Y N Y N Y N Y N N N N Y Y N Y Y
a. RPQ 8P2197 enables the ordering of a daughter card supporting 20 km unrepeated distance for 1 Gbps peer mode. RPQ 8P2262 is a requirement for that option, and other than the normal mode, the channel increment is two, meaning that both ports (FC 0219) at the card must be activated.
145
8049ch04.fm
The following list explains the content of the sample PCHID REPORT: Feature code 0170 (HCA3-O LR (1xIFB)) is installed in the first book (cage A25B, slot 06) location D7 and has AID 0C assigned. Feature code 0171 (HCA3-O (12xIFB)) is installed in the second book (cage A25B, slot 15) location D5 and has AID 1A assigned. Feature code 0405 (OSA-Express4S GbE SX) is installed in cage Z15B slot 4 and has PCHIDs 130 and 131 assigned. PCHID 130 is shared by port 00 and 01, PCHID 131 is shared by port 02 and 03. Feature code 0218 (ISC-3) is installed in cage Z15B slot 20 and has PCHID 210 and 211 assigned to the two ports on the upper daughter card, and PCHID 218 and 219 to the two ports on the lower daughter card. Feature code 0409 (FICON Express8S LX 10 km) is installed in drawer Z15B slot 4 and has PCHIDs 520 and 521 assigned. Feature code 0865 (Crypto Express4S) is installed in drawer Z22B slot 2 and has PCHIDs 580 assigned. Feature code 0407 (OSA-Express4S 10 GbE SR) is installed in drawer Z22B slot 3 and has PCHIDs 590 assigned. Feature code 0408 (OSA-Express4S 1000BASE-T) is installed in drawer Z22B slot 4 and has PCHIDs 5A0 and 5A1 assigned. PCHID 5A0 is shared by port 00 and 01, PCHID 5A1 is shared by port 02 and 03. The pre-assigned PCHID number of each I/O port relates directly to its physical location (jack location in a specific slot).
4.9 Connectivity
I/O channels are part of the channel subsystem (CSS). They provide connectivity for data exchange between servers, or between servers and external control units (CU) and devices, or networks. Communication between servers is implemented by using InterSystem Channel-3 (ISC-3), coupling using InfiniBand (IFB), or channel-to-channel connections (CTC).
146
8049ch04.fm
Communication to local area networks (LANs) is provided by the OSA-Express3 and OSA-Express4S features. Connectivity to I/O subsystems to exchange data is provided by FICON channels.
FC, FCP FC, FCP FC, FCP OSD, OSX OSD, OSN OSE, OSD, OSC, OSN, OSM OSD OSD, OSX OSE, OSD, OSC, OSN, OSM CFP CFP CIB CIB CIB CIB
2 1 2
2 1 2
96 48 96
2 / ISC-D 2 / ISC-D 2 2 2 4
1 2 2 2 2 4
48 48 32 32 32 64
12 12 16 16 16 16
Yes Yes No No No No
147
8049ch04.fm
b. Each OSA-Express3 feature installed in an I/O cage/drawer reduces by two (2) the number of OSA-Express4S features allowed
At least one I/O feature (FICON) or one coupling link feature (IFB or ISC-3) must be present in the minimum configuration. A maximum of 256 channels is configurable per channel subsystem and per operating system image. The following channels can be shared and spanned: FICON channels defined as FC or FCP OSA-Express3 defined as OSC, OSD, OSE, OSM, OSN or OSX OSA-Express4S defined as OSC, OSD, OSE, OSM, OSN or OSX Coupling links defined as CFP, ICP, or CIB HiperSockets defined as IQD Each Crypto Express feature occupies an I/O slot but does not have a CHPID type, but logical partitions in all CSSs have access to the features. Each Crypto Express adapter can be defined to up to 16 logical partitions. Each Flash Express feature occupies two I/O slots but does not have a CHPID type, but logical partitions in all CSSs have access to the features. The Flash Express feature can be defined to up to 60 logical partitions.
148
8049ch04.fm
Feature code 0219 3321 3322 3325 3326 0409 0410 3370 3371 3362 3363 3367 0404 0405 0406 0407 0408
Feature name ISC-3 FICON Express4 LX 10 km FICON Express4 SX FICON Express8 LX 10 km FICON Express8 SX FICON Express8S LX 10 km FICON Express8S SX OSA-Express3 10 GbE LR OSA-Express3 10 GbE SR OSA-Express3 GbE LX OSA-Express3 GbE SX OSA-Express3 1000BASE-T OSA-Express4S GbE LX OSA-Express4S GbE SX OSA-Express4S 10 GbE LR OSA-Express4S 10 GbE SR OSA-Express4S 1000BASE-T
Connector type LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex LC Duplex RJ-45 LC Duplex LC Duplex LC Duplex LC Duplex RJ-45
Cable type 9 m SM 9 m SM 50, 62.5 m MM 9 m SM 50, 62.5 m MM 9 m SM 50, 62.5 m MM 9 m SM 50, 62.5 m MM 9 m SM 50, 62.5 m MM Category 5 UTPc 9 m SM 50, 62.5 m MM 9 m SM 50, 62.5 m MM Category 5 UTP
a. MM is multimode fiber. b. SM is single mode fiber. c. UTP is unshielded twisted pair. Consider using Category 6 UTP for 1000 Mbps connections.
149
8049ch04.fm
All FICON Express8S, FICON Express8 and FICON Express4 features use small form-factor pluggable (SFP) optics that allow for concurrent repair or replacement for each SFP. The data flow on the unaffected channels on the same feature can continue. A problem with one FICON port no longer requires replacement of a complete feature. All FICON Express8S, FICON Express8 and FICON Express4 features also support cascading (the connection of two FICON Directors in succession) to minimize the number of cross-site connections and help reduce implementation costs for disaster recovery applications, GDPS, and remote copy. Each FICON Express8S, FICON Express8 and FICON Express4 channel can be defined independently, for connectivity to servers, switches, directors, disks, tapes, and printers as: CHPID type FC - FICON, High Performance FICON for System z (zHPF), and FICON Channel-to-Channel (FCTC). FICON, FCTC, and zHPF protocols are supported simultaneously. CHPID type FCP - Fibre Channel Protocol which supports attachment to SCSI devices directly or through Fibre Channel switches or directors. FICON channels (CHPID type FC or FCP) can be shared among logical partitions and can be defined as spanned. All ports on a FICON feature must be of the same type, either LX or SX. The features are connected to a FICON-capable control unit, either point-to-point or switched point-to-point, through a Fibre Channel switch.
FICON Express8S
The FICON Express8S feature resides exclusively in the PCIe I/O drawer. Each of the two independent ports is capable of 2 gigabits per second (Gbps), 4 Gbps, or 8 Gbps depending upon the capability of the attached switch or device. The link speed is auto-negotiated, point-to-point, and is transparent to users and applications. The two types of FICON Express8S optical transceivers supported are the long wavelength (LX) and the short wavelength (SX): FICON Express8S 10km LX feature FC 0409, with two ports per feature, supporting LC Duplex connectors FICON Express8S SX feature FC 0410, with two ports per feature, supporting LC Duplex connectors Each port of the FICON Express8S 10 km LX feature uses a 1300 nanometer (nm) optical transceiver, supports an unrepeated distance of 10 km using 9 m single-mode fiber. Each port of the FICON Express8S SX feature uses an 850 nanometer (nm) optical transceiver. supports varying distances depending on the fiber used (50 or 62.5 m multimode fiber). Auto-negotiation: FICON Express8S features do not support auto-negotiation to a data link rate of 1 Gbps. Statement of direction: FICON Express8S feature will be the last FICON feature supporting auto-negotiation to a data link rate of 2 Gbps.
FICON Express8
The FICON Express8S feature resides in an I/O cage or I/O drawer. Each of the four independent ports is capable of 2 gigabits per second (Gbps), 4 Gbps, or 8 Gbps depending
150
8049ch04.fm
upon the capability of the attached switch or device. The link speed is auto-negotiated, point-to-point, and is transparent to users and applications. The two types of FICON Express8 optical transceivers supported are the long wavelength (LX) and the short wavelength (SX): FICON Express8 10km LX feature FC 3325, with four ports per feature, supporting LC Duplex connectors FICON Express8 SX feature FC 3326, with four ports per feature, supporting LC Duplex connectors Each port of FICON Express8 10 km LX feature uses a 1300 nanometer (nm) fiber bandwidth transceiver, supports an unrepeated distance of 10 km using 9 m single-mode fiber. Each port of FICON Express8 SX feature uses an 850 nanometer (nm) optical transceiver, supports varying distances depending on the fiber used (50 or 62.5 m multimode fiber). Auto-negotiation: FICON Express8 features do not support auto-negotiation to a data link rate of 1 Gbps.
FICON Express4
The FICON Express4 feature resides in an I/O cage or I/O drawer. Each of the four independent ports is capable of 1 gigabits per second (Gbps), 2Gbps, or 4 Gbps depending upon the capability of the attached switch or device. The link speed is auto-negotiated, point-to-point, and is transparent to users and applications. Statement of Direction: The IBM zEnterprise EC12 is planned to be the last high-end System z server to offer support of the FICON Express4 features (#3321, #3322). FICON Express4 will not be supported on future high-end System z servers as carry forward on an upgrade. Enterprises should continue migrating from the FICON Express4 features to the FICON Express8S features (#0409, #0410). The two types of FICON Express4 optical transceivers supported are the two long wavelength (LX) and one short wavelength (SX): FICON Express4 10km LX feature FC 3321, with four ports per feature, supporting LC Duplex connectors FICON Express4 SX feature FC 3322, with four ports per feature, supporting LC Duplex connectors FICON Express4 4 km LX (FC3324) is not supported on zEC12 The FICON Express4 LX features use 1300 nanometer (nm) optical transceivers. One supports an unrepeated distance of 10 km, and the other an unrepeated distance of 4 km, using 9 m single-mode fiber. Use of MCP cables limits the link speed to 1 Gbps and the unrepeated distance to 550 meters. The FICON Express4 SX feature use 850 nanometer (nm) optical transceivers. supports varying distances depending on the fiber used (50 or 62.5 m multimode fiber). Link speed: FICON Express4 is the last FICON family able to negotiate link speed down to 1 Gbps.
151
8049ch04.fm
2 Gbps
1 Gbps
3325 3326
2, 4, or 8 Gbps 8 Gbps
4 Gbps
2 Gbps
0409 0410
2, 4, or 8 Gbps 8 Gbps
4 Gbps
2 Gbps
a. Minimum fiber bandwidths in MHz/km for multimode fiber optic links are included in parentheses were applicable.
152
8049ch04.fm
OSA-Express4S 10 GbE LR OSA-Express4S 10 GbE SR OSA-Express4S GbE LX OSA-Express4S GbE SX OSA-Express4S 1000BASE-T
1 1 2 2 2
OSD, OSX OSD, OSX OSD OSD OSC, OSD, OSE, OSM, OSN
a. Each OSA-Express3 feature installed in an I/O cage/drawer reduces by two (2) the number of OSA-Express4S features allowed b. Both ports on each feature share one PCHID/CHPID
153
8049ch04.fm
being reused, a pair of Mode Conditioning Patch cables are required, one for each end of the link.
8049ch04.fm
receptacle for cabling to an Ethernet switch. The RJ-45 receptacle is required to be attached using EIA/TIA Category 5 or Category 6 unshielded twisted pair (UTP) cable with a maximum length of 100 meters. The OSA-Express4S 1000BASE-T Ethernet feature supports auto-negotiation when attached to an Ethernet router or switch. If you allow the LAN speed and duplex mode to default to auto-negotiation, the OSA-Express port and the attached router or switch auto-negotiate the LAN speed and duplex mode settings between them and connect at the highest common performance speed and duplex mode of interoperation. If the attached Ethernet router or switch does not support auto-negotiation, the OSA-Express port examines the signal it is receiving and connects at the speed and duplex mode of the device at the other end of the cable. The OSA-Express4S 1000BASE-T Ethernet feature can be configured as CHPID type OSC, OSD, OSE, OSN or OSM. Non-QDIO operation mode requires CHPID type OSE. When defined as CHPID type OSM, the port provides connectivity to the intranode management network (INMN). The following settings are supported on the OSA-Express3 1000BASE-T Ethernet feature port: Auto-negotiate 10 Mbps half-duplex or full-duplex 100 Mbps half-duplex or full-duplex 1000 Mbps full-duplex If you are not using auto-negotiate, the OSA-Express port will attempt to join the LAN at the specified speed and duplex mode. If this does not match the speed and duplex mode of the signal on the cable, the OSA-Express port will not connect. Statement of Direction: The OSA-Express4S 1000BASE-T Ethernet feature is planned to be the last copper Ethernet feature to support half-duplex operation and a 10 Mbps link data rate. The zEnterprise EC12 servers are planned to be the last IBM System z servers to support half-duplex operation and a 10 Mbps link data rate for copper Ethernet environments. Any future 1000BASE-T Ethernet feature will support full-duplex operation and auto-negotiation to 100 or 1000 Mbps exclusively.
155
8049ch04.fm
OSA-Express3 10 Gigabit Ethernet (GbE) Short Reach (SR), feature code 3371 OSA-Express3 Gigabit Ethernet (GbE) Long wavelength (LX), feature code 3362 OSA-Express3 Gigabit Ethernet (GbE) Short wavelength (SX), feature code 3363 OSA-Express3 1000BASE-T Ethernet, feature code 3367 Table 4-14 lists the characteristics of OSA-Express3 features.
Table 4-14 OSA-Express3 features I/O feature Feature code Number of ports per feature 2 2 4 4 4 Port increment Maximum number of ports 48 48 96 96 96 Maximum number of features 24 24 24 24 24 CHPID type
OSA-Express3 10 GbE LR OSA-Express3 10 GbE SR OSA-Express3 GbE LX OSA-Express3 GbE SX OSA-Express3 1000BASE-T
2 2 4 4 4
156
8049ch04.fm
157
8049ch04.fm
8049ch04.fm
connection is from the zEC12 to the IEDN top of rack (TOR) switches on zBX Model 003. Or with a stand-alone zEC12 node (no-zBX) interconnect pairs of OSX ports through LC DUPLEX directly connected cables and not wrap cables as has previously been specified. For detailed information about OSA-Express3 an OSA Express4S in an ensemble network, see zBX connectivity on page 226
4.9.6 HiperSockets
The HiperSockets function of IBM zEnterprise EC12 provides up to 32 high-speed virtual LAN attachments like IBM zEnterprise 196 and IBM zEnterprise 114 servers. Previous servers provided 16 attachments. HiperSockets can be customized to accommodate varying traffic sizes. Because HiperSockets does not use an external network, it can free up system and network resources, which can help eliminate attachment costs, and improve availability and performance. HiperSockets eliminates having to use I/O subsystem operations and having to traverse an external network connection to communicate between logical partitions in the same zEC12 server. HiperSockets offers significant value in server consolidation connecting many virtual servers, and can be used instead of certain coupling link configurations in a Parallel Sysplex. HiperSockets internal networks in the support two transport modes: Layer 2 (link layer) Layer 3 (network or IP layer) Traffic can be IPv4 or IPv6, or non-IP such as AppleTalk, DECnet, IPX, NetBIOS, or SNA. HiperSockets devices are protocol and Layer 3-independent. Each HiperSockets device (Layer 2 and Layer 3 mode) has its own MAC address designed to allow the use of applications that depend on the existence of Layer 2 addresses, such as DHCP servers and
Chapter 4. CPC I/O System Structure
159
8049ch04.fm
firewalls. Layer 2 support helps facilitate server consolidation, can reduce complexity, can simplify network configuration, and allows LAN administrators to maintain the mainframe network environment similarly as for non-mainframe environments. Packet forwarding decisions are based on Layer 2 information instead of Layer 3. The HiperSockets device can perform automatic MAC address generation to create uniqueness within and across logical partitions and servers. The use of Group MAC addresses for multicast is supported as well as broadcasts to all other Layer 2 devices on the same HiperSockets networks. Datagrams are delivered only between HiperSockets devices that use the same transport mode. A Layer 2 device cannot communicate directly to a Layer 3 device in another logical partition network. A HiperSockets device can filter inbound datagrams by VLAN identification, the destination MAC address, or both. Analogous to the Layer 3 functions, HiperSockets Layer 2 devices can be configured as primary or secondary connectors or multicast routers. This enables the creation of high-performance and high-availability link layer switches between the internal HiperSockets network and an external Ethernet network or to connect to the HiperSockets Layer 2 networks of different servers. HiperSockets Layer 2 in the zEC12 is supported by Linux on System z, and by z/VM for Linux guest exploitation. zEC12, and zEnterprise CPCs (z196 and z114 servers) support the HiperSockets Completion Queue function that is designed to allow HiperSockets to transfer data synchronously if possible and asynchronously if necessary. This feature combines ultra-low latency with more tolerance for traffic peaks. With the asynchronous support, during high volume situations, data can be temporarily held until the receiver has buffers available in its inbound queue. HiperSockets Completion Queue function is requires at a minimum: z/OS V1.13 Linux on System z distributions: Red Hat Enterprise Linux (RHEL) 6.2 SUSE Linux Enterprise Server (SLES) 11 SP2 z/VSE 5.1.1 The zEC12 and the zEnterprise servers provide the capability to integrate HiperSockets connectivity to the intraensemble data network (IEDN). This extends the reach of the HiperSockets network outside the CPC to the entire ensemble, appearing as a single, Layer 2. Because HiperSockets and IEDN are both internal System z networks, the combination allows System z virtual servers to use the optimal path for communications. In z/VM 6.2, the virtual switch function is enhanced to transparently bridge a guest virtual machine network connection on a HiperSockets LAN segment. This bridge allows a single HiperSockets guest virtual machine network connection to also directly communicate with the following: Other guest virtual machines on the virtual switch External network hosts through the virtual switch OSA UPLINK port. Statements of Direction: HiperSockets Completion Queue: IBM plans to support in a future deliverable of z/VM, transferring HiperSockets messages asynchronously, in addition to the current synchronous manner, on zEC12, z196 and z114.
160
8049ch04.fm
161
8049ch04.fm
IC: CHPIDs (type ICP) defined for internal coupling can connect a CF to a z/OS logical partition in the same zEC12. IC connections require two CHPIDs to be defined, which can only be defined in peer mode. The bandwidth is greater than 2 GBps. A maximum of 32 IC CHPIDs (16 connections) can be defined. Table 4-15 shows the coupling link options.
Table 4-15 Coupling link options Type Description Use Link rate 2 Gbps Distance zEC12H20 maximum 48 zEC12H43->HA1 maximum 48
ISC-3
InterSystem Channel-3
zEC12 to zEC12, z196 z114, z10 zEC12 to zEC12 z196, z114, z10 zEC12 to zEC12 z196, z114, z10 zEC12 to zEC12 z196, z114, z10 zEC12 to zEC12 z196, z114, z10 Internal communication
IFB
6 GBps
16b
32
6 GBps
16b
32
IFB LR
32b
64
16b
32
IC
Internal speeds
N/A
32
32
a. 12x IFB3 protocol: max 4 CHPIDs and connect to other HCA3-O port, else 12x IFB protocol. Auto configured when conditions are met for IFB3. 4.7.5, HCA3-O (12xIFB) fanout (FC0171) on page 140 b. Uses all available fanout slots. Allows no other I/O or coupling
The maximum for IFB links is 64. The maximum number of external coupling links combined (active ISC-3 links and IFB LR) cannot exceed 112 per server. There is a maximum of 128 coupling CHPIDs limitation, including ICP for IC, CIB (for IFB and IFB LR), and CFP (for ISC-3). The zEC12 supports various connectivity options depending on the connected zEC12, z196, z114, or System z10 server. Figure 4-13 on page 163 shows zEC12 coupling link support for zEC12, z196, z114, and System z10 servers.
162
8049ch04.fm
When defining IFB coupling links (CHPID type CIB), HCD now defaults to 7 subchannels. 32 Subchannels are only supported on HCA2-O LR (1xIFB) and HCA3-O LR (1xIFB) on zEC12 and when both sides of the connection use 32 subchannels. Otherwise you should change the default value from 7 to 32 subchannel on each CIB definition.
EC12
HCA3-O LR OR HCA2-O LR* HCA3-O L R OR HCA2-O L R* HCA 3-O OR HCA2-O* HCA2-O* OR HCA3-O
12 x IF B 3 or IFB 6 GBps 150 m
HCA2-O
HCA3-O OR HCA2-O*
HCA2-O OR HCA3-O
HCA3-O OR HCA2-O*
zNext
z/OS and coupling facility images can be running on the same or on separate servers. There must be at least one CF connected to all z/OS images, although there can be other CFs that are connected only to selected z/OS images. Two coupling facility images are required for system-managed CF structure duplexing and, in this case, each z/OS image must be connected to both duplexed CFs. To eliminate any single-points of failure in a Parallel Sysplex configuration, have at least the following components: Two coupling links between the z/OS and coupling facility images. Two coupling facility images not running on the same server. One stand-alone coupling facility. If using system-managed CF structure duplexing or running with resource sharing only, then a stand-alone coupling facility is not mandatory.
163
8049ch04.fm
12x InfiniBand using HCA3-O and HCA2-O (12xIFB) fanout card at 6 GBps to zEC12, z196, z114 and System z10. 1x InfiniBand using both HCA3-O LR and HCA2-O LR at 5.0 or 2.5 Gbps to zEC12, z196, z114, and System z10 servers
164
8049ch04.fm
linkless connection (implemented in Licensed Internal Code) and so does not require any hardware or cabling. An IC link is a fast coupling link, using memory-to-memory data transfers. IC links do not have PCHID numbers, but do require CHPIDs. IC links require an ICP channel path definition at the z/OS and the CF end of a channel connection to operate in peer mode. They are always defined and connected in pairs. The IC link operates in peer mode and its existence is defined in HCD/IOCP. IC links have the following attributes: On System z servers, operate in peer mode (channel type ICP). Provide the fastest connectivity, significantly faster than any external link alternatives. Result in better coupling efficiency than with external links, effectively reducing the server cost associated with Parallel Sysplex technology. Can be used in test or production configurations, and reduce the cost of moving into Parallel Sysplex technology while enhancing performance and reliability. Can be defined as spanned channels across multiple CSSs. Are free of charge (no feature code). Employing ICFs with IC channels will result in considerable cost savings when configuring a cluster. IC links are enabled by defining channel type ICP. A maximum of 32 IC channels can be defined on a System z server.
165
8049ch04.fm
The zEC12 server does not support attachment to the IBM Sysplex Timer. A zEC12 can be added into a Mixed CTN only when there is a system z10 attached to the Sysplex Timer operating as Stratum 1 server. Connections to two Stratum 1 servers are preferable to provide redundancy and avoid a single point of failure. Important: A Parallel Sysplex in an ETR network must migrate to Mixed CTN or STP-only CTN before introducing a zEC12.
166
8049ch04.fm
167
8049ch04.fm
Flash Express cards are internal to the CPC and are accessible using the new System z architected Extended Asynchronous Data Mover (EADM) Facility. EADM is an extension of the ADM architecture used in the past with expanded storage. EADM access is initiated with a Start Subchannel instruction. zEC12 support a maximum of 4 pairs of Flash Express cards. Only one Flash Express card is allowed per Domain. The PCIe drawer has 4 I/O Domains and can install 2 pairs of Flash Express cards. Each pair is installed either in the front of PCIe I/O drawers at slots 1 and 14 or in the rear at slots 25 and 33. The Flash Express cards are first plug into the fronts slot of installed or being installed PCIe I/O drawer before plugging in rear of drawer. These four slots are reserved for Flash Express and not fill by other types of I/O cards until there is no spare slot. Figure 4-14 shows a PCIe I/O cage fully populated with Flash Express cards.
Top View
Card Slot 01 Card Slot 02 Card Slot 03 Card Slot 04 PCI-IN 0 Card Slot 06 Card Slot 07 Card Slot 08 Card Slot 09 FSP Card Slot 11 Card Slot 12 Card Slot 13 Card Slot 14 PCI-IN 2 D o m a i n 2 D o m a i n 3 D o m a i n 0 Card Slot 38 D Card Slot 37 o m Card Slot 36 a Card Slot 35 i PCI-IN 1 n Card Slot 33 1 Card Slot 32 Card Slot 31 Card Slot 30 FSP Card Slot 28 Card Slot 27 Card Slot 26 Card Slot 25 PCI-IN 3 Card Slot 23 Card Slot 22 Card Slot 21 Card Slot 20
Front
Rear
2 interconnect cables
Figure 4-14 PCIe I/O cage fully populated with Flash Express cards.
168
8049ch05.fm
Chapter 5.
169
8049ch05.fm
EC12
Partitions
Partitions: Support the running of a System Control Program (SCP), such as z/OS and allow CPs, IFLs, memory, and Subchannels access to channels. Subchannels: A subchannel represents an I/O device to the hardware, and is used by the SCP to pass an I/O request from the SCP to the channel subsystem. Channel Subsystem: Controls queuing, de-queuing, and priority management, and I/O identification of all I/O operations performed by any logical partitions in a CSS. Channels: The communication path from the channel subsystem to the I/O network and the connected control units / devices.
Subchannels
Channel Subsystem
Channels
170
8049ch05.fm
Subchannels
A subchannel provides the logical representation of a device to a program and contains the information required for sustaining a single I/O operation. A subchannel is assigned for each device defined to the logical partition. Multiple subchannel sets, described in 5.1.3, Multiple subchannel sets on page 171, are available to increase addressability. Three subchannel sets per CSS are supported on EC12. Subchannel set 0 can have up to 63.75 K subchannels, and subchannel sets 1 and 2 can have up to 64 K subchannels each.
Channel paths
Each CSS can have up to 256 channel paths. A channel path is a single interface between a server and one or more control units. Commands and data are sent across a channel path to perform I/O requests.
Control units
A control unit provides the logical capabilities necessary to operate and control an I/O device and adapts the characteristics of each device so that it can respond to the standard form of control provided by the CSS. A control unit can be housed separately, or it can be physically and logically integrated with the I/O device, the channel subsystem, or within the system itself.
I/O devices
An I/O device provides external storage, a means of communication between data-processing systems, or a means of communication between a system and its environment. In the simplest case, an I/O device is attached to one control unit and is accessible through one channel path.
171
8049ch05.fm
Subchannel numbers
Subchannel numbers (including their implied path information to a device) are limited to four hexadecimal digits by the architecture (0x0000 to 0xFFFF). Four hexadecimal digits provide 64 K addresses, known as a set. IBM has reserved 256 subchannels, leaving over 63 K subchannels for general use1. Again, addresses, device numbers, and subchannels are often used as synonyms, although this is not technically accurate. We might hear that there is a maximum of 63.75 K addresses or a
z196 EC12
Logical Channel Subsystem 0 Logical Channel Subsystem 1 Logical Channel Subsystem 2 Logical Channel Subsystem 3
Up to 15 Logical Partitions
Up to 15 Logical Partitions
Up to 15 Logical Partitions
Up to 15 Logical Partitions
Partitions
Partitions
Partitions
Partitions
CSS-0 Subchannels
CSS-1 Subchannels
CSS-2 Subchannels
CSS-3 Subchannels
SS-0
SS-1
SS-2
64K
SS-0
SS-1
SS-2
64K
SS-0
SS-1
SS-2
64K
SS-0
SS-1
SS-2
64K
Channels
Channels
Channels
Channels
The additional subchannel sets, in effect, add an extra high-order digit (either 0, 1 or 2) to existing device numbers. For example, we might think of an address as 08000 (subchannel set 0), or 18000 (subchannel set 1) or 28000 (subchannel set 2). Adding a digit is not done in system code or in messages because of the architectural requirement for four-digit addresses (device numbers or subchannels). However, certain messages do contain the subchannel set number, and you can mentally use that as a high-order digit for device numbers. Only a few
The number of reserved subchannel is 256. We abbreviate this to 63.75 K in this discussion to easily differentiate it from the 64 K subchannels available in subchannel sets 1 and 2.The informal name, 63.75 K subchannel, represents the following equation: (63 x 1024) + (0.75 x 1024) = 65280
172
8049ch05.fm
requirements refer to the subchannel sets 1 and 2, because they are only used for these special devices. JCL, messages, and programs rarely refer directly these special devices. Moving these special devices into an alternate subchannel set creates additional space for device number growth. The appropriate subchannel set number must be included in IOCP definitions or in the HCD definitions that produce the IOCDS. The subchannel set number defaults to zero.
173
8049ch05.fm
Summary of identifiers
It is good practice to establish a naming convention for the logical partition identifiers. As shown in Figure 5-4 on page 175, which summarizes the identifiers and how they are defined, you can use the CSS number concatenated to the MIF ID, which means that logical partition
174
8049ch05.fm
ID 3A is in CSS 3 with MIF ID A. This fits within the allowed range of logical partition IDs and conveys helpful information to the user.
CSS0
CSS1
CSS2
CSS3
Logical Partition ID
02 04 0A
Logical Partition ID
14 16 1D
Log Part ID
22
Logical Partition ID
35 3A
MIF ID 2
MIF ID 4
MIF ID A
MIF ID 4
MIF ID 6
MIF ID D
MIF ID 2
MIF ID 5
MIF ID A
175
8049ch05.fm
A channel is considered a spanned channel if the same CHPID number in different CSSs is assigned to the same PCHID in IOCP, or is defined as spanned in HCD. In the case of internal channels (for example, IC links and HiperSockets), the same applies, but with no PCHID association. They are defined with the same CHPID number in multiple CSSs. In Figure 5-5, CHPID 04 is spanned to CSS0 and CSS1. Because it is an internal channel link, no PCHID is assigned. CHPID 06 is an external spanned channel and has a PCHID assigned.
Partition Partition 2 1
...
...
Partition 60
CSS0
CSS1
MIF-1
MIF-2
...
MIF-F
MIF-1
MIF-2
MIF-3
...
MIF-F
CHPID 00
CHPID 01
...
CHPID 04 SPAN
CHPID 06 SPAN
...
CHPID 00
CHPID 01
...
PCHID 120
CHPIDs that span CSSs reduce the total number of channels available. The total is reduced, because no CSS can have more than 256 CHPIDs. For a zEC12 with two CSSs defined, a total of 512 CHPIDs is supported. If all CHPIDs are spanned across the two CSSs, then only 256 channels are supported. For a zEC12 with four CSSs defined, a total of 1024 CHPIDs is supported. If all CHPIDs are spanned across the four CSSs, then only 256 channels are supported. Channel spanning is supported for internal links (HiperSockets and Internal Coupling (IC) links) and for certain external links (FICON Express8S, and FICON Express8 channels, OSA-Express4S, OSA-Express3, and Coupling Links).
176
8049ch05.fm
zEC12
CSS LPAR name LPAR ID MIF ID MSS CHPID PCHID
80 140
CSS 0 LP1 01 1
SS 0 81 150
CSS 1 LP3 05 5
SS 2 90 1E0 91 1F0 80 141
LP2 03 3
SS 1
LP14 12 2
SS 0 81 151
LP15 13 3
SS 1 90 1E1
LP16 15 5
SS 2
91 1F1
Directors
61
62
Disk LCUs
Disk LCUs
In each CSS, the CHPIDs are shared across all logical partitions. The CHPIDs in each CSS can be mapped to their designated PCHIDs using the CHPID Mapping Tool (CMT) or manually using HCD or IOCP. The output of the CMT is used as input to HCD or the IOCP to establish the CHPID to PCHID assignments.
5.1.9 Adapter ID
When using HCD or IOCP to assign a CHPID to a Parallel Sysplex over an InfiniBand (IFB) coupling link port, an adapter ID (AID) number is required. The AID is bound to the serial number of the fanout. If the fanout is moved, the AID moves with it. No IOCDS update is required if adapters are moved to a new physical location.
177
8049ch05.fm
optimal performance. This enhancement is accomplished by exploiting the in-band I/O instrumentation and metrics of the System z FICON and zHPF protocols. This channel subsystem enhancement is exclusive to zEC12 and is supported on all FICON channels when configured as CHPID type FC. This enhancement is transparent to operating systems.
178
8049ch05.fm
Maximum number of CSSs Maximum number of CHPIDs Maximum number of LPARs supported per CSS Maximum number of LPARs supported per system Maximum number of HSA subchannels Maximum number of devices Maximum number of CHPIDs per CSS Maximum number of CHPIDs per logical partition Maximum number of subchannel per logical partition
4 1024 15 60 11505 K (191.75 K per partition x 60 partitions) 255 K (4 CSSs x 63.75 K devices) 256 256 191.75 K (63.75 K + 2 x 64 K)
All channel subsystem images (CSS images) are defined within a single I/O configuration data set (IOCDS). The IOCDS is loaded and initialized into the hardware system area (HSA) during system power-on reset.The HSA is pre-allocated in memory with a fixed size of 32 GB. This eliminates planning for HSA and pre-planning for HSA expansion, because HCD/IOCP always reserves the following items by the IOCDS process: Four CSSs 15 LPARs in each CSS Subchannel set 0 with 63.75 K devices in each CSS Subchannel set 1 with 64 K devices in each CSS Subchannel set 2 with 64 K devices in each CSS All these are designed to be activated and used with dynamic I/O changes. Figure 5-7 shows a logical view of the relationships. Note that each CSS supports up to 15 logical partitions. System-wide, a total of up to 60 logical partitions are supported.
179
8049ch05.fm
Driver
H89 HA1
1-101 PUs
H20
1-20 PUs 32-704 GB
H43
1-43 PUs
H66
1-66 PUs
Up to four books per server Available Processing Units (min 1, max 101) Memory (min 32 GB, max 3040 GB) Up to 15 Logical Partitions per LCSS Up to 60 Logical Partitions per CEC Up to 4 Logical Channel Subsystems Up to 3 Subchannel Sets per CSS Single HSA with one active IOCDS IFB-C, and IFB-O cables
Logical Partitions
CSS 0 up to 256 CHPIDs SS 0 SS 1 SS 2 CSS 1 CSS 2 up to 256 up to 256 CHPIDs CHPIDs SS 0 SS 1 SS 2 SS 0 SS 1 SS 2 CSS 3 up to 256 CHPIDs SS 0 SS 1 SS 2
Figure 5-7 Logical view of zEC12 models, CSSs, IOCDS, and HSA
Book repair: The HSA can be moved from one book to a different book in an enhanced availability configuration as part of a concurrent book repair action. The channel definitions of a CSS are not bound to a single book. A CSS can define resources that are physically connected to any InfiniBand cable of any book in a multibook CPC.
180
8049ch05.fm
181
8049ch05.fm
182
8049ch06.fm
Chapter 6.
Cryptography
This chapter describes the hardware cryptographic functions available on the IBM zEnterprise EC12. The CP Assist for Cryptographic Function (CPACF) along with the PCIe Cryptographic Coprocessors offer a balanced use of resources and unmatched scalability. The zEC12 include both standard cryptographic hardware and optional cryptographic features for flexibility and growth capability. IBM has a long history of providing hardware cryptographic solutions, from the development of Data Encryption Standard (DES) in the 1970s to have the Crypto Express tamper-sensing and tamper-responding programmable features designed to meet the U.S. Government's highest security rating FIPS 140-2 Level 41. The cryptographic functions include the full range of cryptographic operations necessary for e-business, e-commerce, and financial institution applications. User-Defined Extensions (UDX) permits to add custom cryptographic functions to the set of functions that the zEC12 offers. Secure Sockets Layer/Transport Layer Security (SSL/TLS) is a key technology for conducting secure e-commerce using Web servers, and it has being adopted by a rapidly increasing number of applications, demanding new levels of security, performance, and scalability. This chapter discusses the following topics: Cryptographic synchronous functions on page 184 Cryptographic asynchronous functions on page 184 CPACF protected key on page 187 PKCS #11 Overview on page 189 Cryptographic feature codes on page 193 CP Assist for Cryptographic Function on page 194 Crypto Express4S on page 194 Crypto Express3 on page 197 Tasks performed by PCIe Crypto Express on page 199 TKE workstation feature on page 203 Cryptographic functions comparison on page 208
Federal Information Processing Standards (FIPS)140-2 Security Requirements for Cryptographic Modules
183
8049ch06.fm
184
8049ch06.fm
Single-length key DES Double-length key DES Triple-length key DES (Triple-DES) DES key generation and distribution PIN generation, verification, and translation functions Random number generator PKCS #11 functions2 ICSF has implemented callable services in support of PKCS #11 standard. Secure IBM Enterprise PKCS #11 (EP11) coprocessor mode implements secure keys for PKCS #11 functions. Public key algorithm (PKA) functions Supported callable services intended for application programs that use PKA include: Importing RSA public-private key pairs in clear and encrypted forms Rivest-Shamir-Adelman (RSA), which can provide: Key generation, up to 4,096-bit Signature generation and verification, up to 4,096-bit Import and export of DES keys under an RSA key, up to 4,096-bit
Public key encryption (PKE) / Public key decryption (PKD): The PKE and PKD callable services are provided for assisting the SSL/TLS handshake. They are used to offload compute-intensive portions of the protocol onto the cryptographic adapters. Europay Mastercard VISA (EMV) standard: Applications can be written to comply with the EMV standard for financial transactions between heterogeneous hardware and software. EMV standards have been updated to exploit improved security properties of EMV contact and contactless cards. ICSF HRC770A improved support of EMV card applications that support American Express cards.
Chapter 6. Cryptography
185
8049ch06.fm
PKA Key Token Change (CSNDKTC), and Digital Signature Verify (CSFNDFV) callable services support the remote key loading process. Key exchange with non-CCA cryptographic systems: This function allows the exchange of operational keys between the Crypto Express4S or Crypto Express3 coprocessors and non-CCA systems, such as the Automated Teller Machines (ATM). IBM Common Cryptographic Architecture (CCA) employs control vectors to control the usage of cryptographic keys. Non-CCA systems use other mechanisms, or can use keys that have no associated control information. Enhancements to key exchange functions added to the CCA the ability to exchange keys between CCA systems and systems that do not use control vectors.It allows the CCA system owner to define permitted types of key to be imported and exported while preventing uncontrolled key exchange that can open the system to an increased threat of attack. Elliptic Curve Cryptography (ECC) Digital Signature Algorithm support: Elliptic Curve Cryptography is an emerging public-key algorithm intended to eventually replace RSA cryptography in many applications. ECC is capable of providing digital signature functions and key agreement functions. The CCA functions provide ECC key generation and key management and provide digital signature generation and verification functions compliant with the ECDSA method described in ANSI X9.62 Public Key Cryptography for the Financial Services Industry: The Elliptic Curve Digital Signature Algorithm (ECDSA). ECC uses keys that are shorter than RSA keys for equivalent strength-per-key-bit. So the strength-per-key-bit is substantially greater in an algorithm that uses elliptic curves. This Crypto function is supported by z/OS, z/VM and Linux on System z. Elliptic Curve Diffie-Hellman (ECDH) algorithm support: The Common Cryptographic Architecture has been extended to include the Elliptic Curve Diffie-Hellman (ECDH) algorithm. Elliptic Curve Diffie-Hellman (ECDH) is a key agreement protocol that allows two parties, each having an elliptic curve public-private key pair, to establish a shared secret over an insecure channel. This shared secret can be used directly as a key, or to derive another key which can then be used to encrypt subsequent communications using a symmetric key cipher such as AES KEK. Enhancements include: Key management function to support AES KEK Generating an ECC private key wrapped with an AES KEK Importing and exporting an ECC private key wrapped with an AES KEK Support for ECDH with a new service
User-Defined Extensions (UDX) support: UDX allows the user to add customized operations to a cryptographic coprocessor. User-Defined Extensions to the Common Cryptographic Architecture (CCA) support customized operations that execute within the Crypto Express features when defined as coprocessor. UDX is supported under a special contract through an IBM or approved third-party service offering. The CryptoCards website directs your request to an IBM Global Services location appropriate for your geographic location. A special contract is negotiated between you and IBM Global Services. The contract is for development of the UDX code by IBM Global Services according to your specifications and an agreed-upon level of the UDX. An UDX toolkit for System z is tied to specific versions of the CCA card code and the related host code. UDX is available for the Crypto Express4S (Secure IBM CCA coprocessor mode only) and Crypto Express3 feature. An UDX migration is no more disruptive than a normal MCL or ICSF release migration.
186
8049ch06.fm
In zEC12 it is allowed to import up to 4 UDX files and they can be imported only from a DVD-ROM. The UDX configuration panel was updated to include a Reset to IBM Default bottom. Note: CCA will have a new code level at zEC12 and the UDX customers will require a new UDX. For more information, see the IBM CryptoCards website: http://www.ibm.com/security/cryptocards
Chapter 6. Cryptography
187
8049ch06.fm
CPACF Wrapped key stored in ICSF address space in fetch protected storage
ICSF ICSF
CPACF Wrapped Key
CKDS
Secure Key
Cleartext Key
Wrapping Key
CPACF Wrapping key resides in protected HSA storage and is not visible for operating system
If a CEX4C or a CEX3C is available, a protected key can begin its life as a secure key. Otherwise, an application is responsible for creating or loading a clear key value and then using the PCKMO instruction to wrap the key. ICSF is not called by application if the Crypto Express4S or the Crypto Express3 is not available. A new segment in the profiles at the CSFKEYS class in RACF restricts which secure keys can be used as protected keys. By default, all secure keys are considered not eligible to be used as protected keys. The process described in Figure 6-1 considers a secure key as being the source of a protected key. The source key in this case was already stored in CKDS as a secure key (encrypted under the master key). This secure key is sent to Crypto Express4S or to the Crypto Express3 to be deciphered and sent to CPACF in clear text. At CPACF, the key is wrapped under the LPAR wrapping key and then it is returned to ICSF. After the key is wrapped, ICSF can keep the protected value in memory, passing it to the CPACF, where the key will be unwrapped for each encryption/decryption operation. The protected key is designed to provide substantial throughput improvements for a large volume of data encryption as well as low latency for encryption of small blocks of data. A high performance secure key solution, also known as a protected key solution, requires HCR7770 as a minimum release.
188
8049ch06.fm
189
8049ch06.fm
TOKEN
Public Objects
SLOT
Processor Private Objects
The PKCS #11 general model components are represented in Figure 6-2: Token - Logical view of a cryptographic device (e.g. smart card, HSM) Slot - Logical view of a smart card reader Objects - Items stored in a token (e.g. digital certificate, cryptographic key) User - The owner of the private data on the token by knowing the access Personal Identification Number (PIN) Security Officer - Person who initializes the token and the User PIN.
190
8049ch06.fm
A C language application program interface (API) that implements a subset of the PKCS #11 specification. Token management ICSF callable services, which are also used by the C API. The ICSF ISPF panel, called Token Browser, which provides the capability to see a formatted view of TKDS objects and make minor, limited updates to them. The RACF RACDCERT command supports the certificate, public key, and private key objects and provides subfunctions to manage these objects together with tokens. The gskkyman command supports management of certificates and keys similar to the way RACFDCERT does. ICSF supports PKCS #11 session objects and token objects. ICSF supports PKCS#11 session objects and token objects. Session objects exist in memory only. They are not maintained on the direct access storage device (DASD).I An application has only one session objects database, even if the application spawns multiple PKCS #11 sessions. Token objects are stored in the TKDS with one record per object. They are visible to all applications having sufficient permission to the token. The objects are persistent and remain associated with the token even after a session is closed. The PKCS #11 standard was designed for systems that grant access to token information based on a PIN. z/OS does not use PINs; instead, profiles in the SAF CRYPTOZ class control access to tokens. Each token has two resources in the CRYPTOZ class: The resource USER.token-name controls the access of the user role to the token. The resource SO.token-name controls the access of the SO role to the token. A users access level to each of these resources (read, update, and control) determines the users access level to the token.
APPLICATION
RACF
Tokens In TKDS
PKCS#11 API (CRYPTOKI)
ICSF
TKDS
Figure 6-3 Mapping the PKCS #11 model to the z/OS PKCS #11 implementation
Chapter 6. Cryptography
191
8049ch06.fm
.In Figure 6-3 on page 191 we map the concepts introduced by PKCS #11 model to the z/OS PKCS #11 implementation.
Tokens
The PKCS #11 tokens on z/OS are virtual, conceptually similar to RACF (SAF) keyrings. An application can have one or more z/OS PKCS #11 tokens, depending on its need. z/OS PKCS #11 tokens are created using system software such as RACF, the gskkyman utility or by applications using the C API. Also ICSF panels can be used for token management with limited usability. Each token has a unique token name or label that is specified by the user or application when the token is created. Similar to the way that z/OS PKCS #11 supports token creation, the PKCS #11 tokens can be deleted using the same system software tools used when created.
192
8049ch06.fm
Feature code
3863
Description
CP Assist for Cryptographic Function (CPACF) enablement: This feature is a prerequisite to use CPACF (except for SHA-1, SHA-224, SHA-256, SHA-384, and SHA-512) and Crypto Express features. Crypto Express3 feature: A maximum of eight features can be carry forwarded. It is an optional feature and each feature contains two PCI Express cryptographic adapters (adjunct processors). This feature is not supported as a new build, it is available only in a carry forward basis when upgrading from earlier generations to zEC12. Crypto Express4S feature: A maximum of sixteen features can be ordered. It is an optional feature and each feature contains one PCI Express cryptographic adapters (adjunct processor). Trusted Key Entry (TKE) workstation: This feature is optional. TKE provides a basic key management (key identification, exchange, separation, update, backup), as well as security administration. The TKE workstation has one Ethernet port and supports connectivity to an Ethernet Local Area Network (LAN) operating at 10, 100, or 1000 Mbps. Up to ten (10) features per zEC12 can be installed TKE 7.2 Licensed Internal Code (TKE 7.2 LIC): The 7.2 LIC requires Trusted Key Entry workstation feature code 0841. It is required to support CEX4P. The 7.2 LIC can also be used to control z196, z114, z10 EC, z10 BC, z9 EC, z9 BC, z990, and z890 servers. TKE Smart Card Reader: Access to information in the smart card is protected by a personal identification number (PIN). One (1) feature code includes two Smart Card Readers, two cables to connect to the TKE workstation, and 20 smart cards. Smart card part 74Y0551 is required to support CEX4P. TKE additional smart cards: When one feature code is ordered a quantity of 10 smart cards are shipped. Order increment is one up to 99 (990 blank smart cards). Smart card part 74Y0551 is required to support CEX4P.
0864
0865
0841
0850
0885
0884
TKE includes support for the AES encryption algorithm with 256-bit master keys and key management functions to load or generate master keys to the cryptographic coprocessor. If the TKE workstation is chosen to operate the Crypto Express features in a zEC12, a TKE workstation with the TKE 7.2 LIC or later is required. See 6.10, TKE workstation feature on page 203 for a more detailed description. Important: Products that include any of the cryptographic feature codes contain cryptographic functions that are subject to special export licensing requirements by the United States Department of Commerce. It is the customers responsibility to understand and adhere to these regulations when moving, selling, or transferring these products.
Chapter 6. Cryptography
193
8049ch06.fm
These functions are provided as problem-state z/Architecture instructions, directly available to application programs. These instructions are known as Message-Security Assist (MSA). When enabled, the CPACF runs at processor speed for every CP, IFL, zIIP, and zAAP. Details on MSA instructions can be found in the z/Architecture Principles of Operation, SA22-7832. The functions of the CPACF must be explicitly enabled using FC 3863 by the manufacturing process or at the customer site as a MES installation, except for SHA-1, and SHA-2 support for SHA-224, SHA-256, SHA-384, and SHA-512, which are always enabled.
Secure IBM CCA coprocessor (CEX4C) for Federal Information Processing Standard
(FIPS) 140-2 Level 4 certification. This includes secure key functions and it is optionally
194
8049ch06.fm
programmable to deploy additional functions and algorithms using User Defined Extension (UDX).
Secure IBM Enterprise PKCS #11 (EP11) coprocessor (CEX4P) implements industry
standardized set of services that adhere to the PKCS #11 specification v2.20 and more recent amendments. It was designed for extended FIPS and Common Criteria evaluations to meet public sector requirements. This new cryptographic coprocessor mode introduced the PKCS #11 secure key function. Trusted Key Entry (TKE) workstation is required to support the administration of the Crypto Express4S when configured as EP11 mode.
Accelerator (CEX4A) for acceleration of public key and private key cryptographic
operations used with Secure Sockets Layer / Transport Layer Security (SSL/TLS) processing. These modes can be configured via the Support Element, and the PCIe adapter must be configured offline to change the mode. Note: Switching between configuration modes will erase all card secrets. Exception occurs when switching from Secure CCA to accelerator and vice versa. The Crypto Express4S uses the 4765 PCIe Coprocessor. It holds a secured subsystem module. The card layout can be seen at Figure 6-4.
The Crypto Express4S feature does not have external ports and does not use fiber optic or other cables. It does not use CHPIDs, but requires one slot in PCIe I/O drawer. and one PCHID for each PCIe cryptographic adapter. Removal of the feature or card zeroizes its content. The zEC12 supports a maximum of sixteen Crypto Express4S features. Access to the PCIe cryptographic adapter is controlled through the setup in the image profiles on the SE. Adapter: Though PCIe cryptographic adapters have no CHPID type and are not identified as external channels, all logical partitions in all channel subsystems have access to the adapter (up to 16 logical partitions per adapter). Having access to the adapter requires setup in the image profile for each partition. The adapter must be in the candidate list.
Chapter 6. Cryptography
195
8049ch06.fm
Minimum number of orderable features for each servera Order increment above two features Maximum number of features for each server Number of PCIe cryptographic adapters for each feature (coprocessor or accelerator) Maximum number of PCIe adapters for each server Number of cryptographic domains for each PCIe adapter
b
2 1 16 1 16 16
a. The minimum initial order of Crypto Express4S features is two. After the initial order, additional Crypto Express4S can be ordered one feature at a time up to a maximum of sixteen. b. More than one partition, defined to the same CSS or to different CSSs, can use the same domain number when assigned to different PCIe cryptographic adapters.
The concept of dedicated processor does not apply to the PCIe cryptographic adapter. Whether configured as coprocessor or accelerator, the PCIe cryptographic adapter is made available to a logical partition as directed by the domain assignment and the candidate list in the logical partition image profile, regardless of the shared or dedicated status given to the CPs in the partition. When installed non-concurrently, Crypto Express4S features are assigned PCIe cryptographic adapter numbers sequentially during the power-on reset following the installation. When a Crypto Express4S feature is installed concurrently, the installation can select an out-of-sequence number from the unused range. When a Crypto Express4S feature is removed concurrently, the PCIe adapter numbers are automatically freed. The definition of domain indexes and PCIe cryptographic adapter numbers in the candidate list for each logical partition needs to be planned ahead to allow for nondisruptive changes: Operational changes can be made by using the Change LPAR Cryptographic Controls task from the Support Element, which reflects the cryptographic definitions in the image profile for the partition. With this function, adding and removing the cryptographic feature without stopping a running operating system can be done dynamically. The same usage domain index can be defined more than once across multiple logical partitions. However, the PCIe cryptographic adapter number coupled with the usage domain index specified must be unique across all active logical partitions. The same PCIe cryptographic adapter number and usage domain index combination can be defined for more than one logical partition, for example to define a configuration for backup situations. Note that only one of the logical partitions can be active at any one time. The zEC12 allows for up to 60 logical partitions to be active concurrently. Each PCI Express supports 16 domains, whether it is configured as a Crypto Express4S accelerator or a Crypto Express4S coprocessor. The server configuration must include at least four Crypto Express4S features (four PCIe adapters and 16 domains per PCIe adapter) when all 60 logical partitions require concurrent access to cryptographic functions. More Crypto Express4S features might be needed to satisfy application performance and availability requirements.
196
8049ch06.fm
Secure coprocessor (CEX3C) for Federal Information Processing Standard (FIPS) 140-2
Level 4 certification. This includes secure key functions and it is optionally programmable to deploy additional functions and algorithms using User Defined Extension (UDX).
Accelerator (CEX3A) for acceleration of public key and private key cryptographic
operations used with Secure Sockets Layer / Transport Layer Security (SSL/TLS) processing. These modes can be configured via the Support Element, and the PCIe adapter must be configured offline to change the mode. The Crypto Express3 feature like its predecessors is designed to complement the functions of CPACF. This feature is tamper-sensing and tamper-responding. Unauthorized removal of the adapter or feature zeroizes its content. It provides dual processors operating in parallel supporting cryptographic operations with high reliability. The CEX3 uses the 4765 PCIe Coprocessor. It holds a secured subsystem module, batteries for backup power and a full-speed USB 2.0 host port available through a mini-A connector. On System z these USB ports are not used. The securely encapsulated subsystem contains two 32-bit PowerPC 405D5 RISC processors running in lock-step with cross-checking to detect malfunctions. The subsystem also includes a separate service processor used to manage self-test and firmware updates, RAM, flash memory, and battery-powered memory, cryptographic-quality random number generator, AES, DES, TDES, SHA-1, SHA-224, SHA-256, SHA-384, SHA-512 and modular-exponentiation (for example, RSA, DSA) hardware, and full-duplex DMA communications. Figure 6-5 brings the Crypto Express3 feature physical layout.
Chapter 6. Cryptography
197
8049ch06.fm
2 boards
SP
+RSA
Core +SHA
New Interface
The Crypto Express3 feature does not have external ports and does not use fiber optic or other cables. It does not use CHPIDs, but requires one slot in the I/O cage and one PCHID for each PCIe cryptographic adapter. Removal of the feature or card zeroizes the content. The zEC12 supports a maximum of eight Crypto Express3 features in a carry forward basis, offering a combination of up to 16 coprocessor and accelerators. Access to the PCIe cryptographic adapter is controlled through the setup in the image profiles on the SE. Adapter: Though PCIe cryptographic adapters have no CHPID type and are not identified as external channels, all logical partitions in all channel subsystems have access to the adapter (up to 16 logical partitions per adapter). Having access to the adapter requires setup in the image profile for each partition. The adapter must be in the candidate list.
Minimum number of carry forward features for each server Maximum number of features for each server Number of PCIe cryptographic adapters for each feature (coprocessor or accelerator) Maximum number of PCIe adapters for each server Number of cryptographic domains for each PCIe adaptera
2 8 2 16 16
a. More than one partition, defined to the same CSS or to different CSSs, can use the same domain number when assigned to different PCIe cryptographic adapters.
198
8049ch06.fm
The concept of dedicated processor does not apply to the PCIe cryptographic adapter. Whether configured as coprocessor or accelerator, the PCIe cryptographic adapter is made available to a logical partition as directed by the domain assignment and the candidate list in the logical partition image profile, regardless of the shared or dedicated status given to the CPs in the partition. When installed non-concurrently, Crypto Express3 features are assigned PCIe cryptographic adapter numbers sequentially during the power-on reset following the installation. When a Crypto Express3 feature is installed concurrently, the installation can select an out-of-sequence number from the unused range. When a Crypto Express3 feature is removed concurrently, the PCIe adapter numbers are automatically freed. The definition of domain indexes and PCIe cryptographic adapter numbers in the candidate list for each logical partition needs to be planned ahead to allow for nondisruptive changes: Operational changes can be made by using the Change LPAR Cryptographic Controls task from the Support Element, which reflects the cryptographic definitions in the image profile for the partition. With this function, adding and removing the cryptographic feature without stopping a running operating system can be done dynamically. The same usage domain index can be defined more than once across multiple logical partitions. However, the PCIe cryptographic adapter number coupled with the usage domain index specified must be unique across all active logical partitions. The same PCIe cryptographic adapter number and usage domain index combination can be defined for more than one logical partition, for example to define a configuration for backup situations. Note that only one of the logical partitions can be active at any one time. The zEC12 allows for up to 60 logical partitions to be active concurrently. Each PCI Express supports 16 domains, whether it is configured as a Crypto Express3 accelerator or a Crypto Express3 coprocessor. The server configuration must include at least two Crypto Express3 features (four PCIe adapters and 16 domains per PCIe adapter) when all 60 logical partitions require concurrent access to cryptographic functions. More Crypto Express3 features might be needed to satisfy application performance and availability requirements.
Chapter 6. Cryptography
199
8049ch06.fm
ANSI TR-31 defines a method of cryptographically protecting Triple Data Encryption Standard (TDES) cryptographic keys and their associated usage attributes. The TR-31 method complies with the security requirements of the ANSI X9.24 Part 1 standard, although use of TR-31 is not required in order to comply with that standard. CCA has added functions that can be used to import and export CCA TDES keys in TR-31 formats. These functions are designed primarily as a secure method of wrapping TDES keys for improved and more secure key interchange between CCA and non-CCA devices and systems. PIN block decimalization table protection: To help avoid a decimalization table attack to learn a personal identification number (PIN), a solution is now available in the CCA to thwart this attack by protecting the decimalization table from manipulation. PINs are most often used for automated teller machines (ATMs) but are increasingly used at point-of sale, for debit and credit cards. ANSI X9.8 PIN security: This function facilitates compliance with the processing requirements defined in the new version of the ANSI X9.8 and ISO 9564 PIN Security Standards and provides added security for transactions that require Personal Identification Numbers (PINs). Enhanced CCA key wrapping to comply with ANSI X9.24-1 key bundling requirements: This support permits that Common Cryptographic Architecture (CCA) key token wrapping method uses Cipher Block Chaining (CBC) mode in combination with other techniques to satisfy the key bundle compliance requirements in standards including ANSI X9.24-1 and the recently published Payment Card Industry Hardware Security Module (PCI HSM) standard. Secure key HMAC (Keyed-Hash Message Authentication Code): HMAC is a method for computing a message authentication code using a secret key and a secure hash function. It is defined in the standard FIPS (Federal Information Processing Standard) 198, The Keyed-Hash Message Authentication Code (HMAC). The CCA function supports HMAC using SHA-1, SHA-224, SHA-256, SHA-384, and SHA-512 hash algorithms. The HMAC keys are variable-length and are securely encrypted so that their values are protected. This Crypto function is supported by z/OS, z/VM and Linux on System z.
200
8049ch06.fm
Three methods of master key entry are provided by Integrated Cryptographic Service Facility (ICSF) for the Crypto Express feature coprocessors: A pass-phrase initialization method, which generates and enters all master keys that are necessary to fully enable the cryptographic system in a minimal number of steps A simplified master key entry procedure provided through a series of Clear Master Key Entry panels from a TSO terminal A Trusted Key Entry (TKE) workstation, which is available as an optional feature in enterprises that require enhanced key-entry security Linux on System z also permits the master key entry through panels or through the TKE workstation. The security-relevant portion of the cryptographic functions is performed inside the secure physical boundary of a tamper-resistant card. Master keys and other security-relevant information are also maintained inside this secure boundary. The Processor Resource/Systems Manager (PR/SM) fully supports the Crypto Express coprocessor features to establish a logically partitioned environment on which multiple logical partitions can use the cryptographic functions. A 128-bit data-protection symmetric master key, a 256-bit AES master key, a 256-bit ECC master key, and one 192-bit public key algorithm (PKA) master key are provided for each of 16 cryptographic domains that a cryptographic adapter can serve. Use the dynamic addition or deletion of a logical partition name to rename a logical partition. Its name can be changed from NAME1 to * (single asterisk) and then changed again from * to NAME2. The logical partition number and MIF ID are retained across the logical partition name change. The master keys in the Crypto Express feature coprocessor that were associated with the old logical partition NAME1 are retained. No explicit action is taken against a cryptographic component for this dynamic change. Coprocessors: Cryptographic coprocessors are not tied to logical partition numbers or MIF IDs. They are set up with PCIe adapter numbers and domain indices that are defined in the partition image profile. The customer can dynamically configure them to a partition and change or clear them when needed.
201
8049ch06.fm
The function extension capability through UDX is not available to the EP11. When Crypto Express4S is defined in EP11 mode, the Trusted Key Entry (TKE) workstation is required to manage the Crypto Express4S feature.
202
8049ch06.fm
Derived Unique Key Per Transaction (DUKPT) is defined in the ANSI X9.24 Part 1 standard. It provides a method in which a separate key is used for each transaction or other message sent from a device, so that an attacker who is able to discover the value of a key would only be able to gain information about a single transaction and not about any that preceded it or which follow it. The keys are derived from a base key that is initially loaded into the device, but then erased as soon as the first keys are derived from it. Those keys, in turn, are erased as subsequent keys are derived. The original definition of DUKPT only allowed derivation of keys to be used in encryption of personal identification number (PIN) blocks. The purpose was to protect PINs that were entered at a point of sale (POS) device and then sent to a host system for verification. Recent versions of X9.24 Part 1 expanded this so that DUKPT can also be used to derive keys for MAC generation and verification, and for data encryption and decryption. Three separate variations of the DUKPT key derivation process are used so that there is key separation between the keys derived for PIN, MAC, and encryption purposes. Secure Cipher Text Translate2 (CTT2) CTT2 is a new data encryption service which will take as input data that is encrypted with one key and return the same data encrypted under a different key. This verb has the advantage that it provides the ability to securely change the encryption key for cipher text without exposing the intermediate plain text. The decryption of data and reencryption of data happens entirely inside the secure module on the Crypto Express feature. Compliance with new random number generation standards The standards defining acceptable methods for generating random numbers have been enhanced to include improved security properties. The Crypto Express coprocessor function has been updated to support methods compliant with these new standards. In this release, the random number generation in the Crypto Express feature when defined as a coprocessor conforms to the Deterministic Random Bit Generator (DRBG) requirements defined in NIST Special Publication 800-90/90A, using the SHA-256 based DBRG mechanism. The methods in these NIST standards supersede those previously defined in NIST FIPS 186-2, ANSI X9.31 and ANSI X9.62. With these improvements, customer applications can help to meet the timeline outlined in Chapter 4 of NIST SP800-131 for switching to the new methods and ceasing use of the old methods. EMV enhancements for applications supporting American Express cards Two changes have been made to the CCA application programming interfaces (APIs) to help improve support of EMV card applications that support American Express cards. The Transaction_Validation verb is used to generate and verify American Express card security codes (CSCs). This release adds support for the American Express CSC version 2.0 algorithm used by contact and contactless cards. The PIN_Change/Unblock verb is used for PIN maintenance. It prepares an encrypted message portion for communicating an original or replacement PIN for an EMV smart card. The verb embeds the PINs in an encrypted PIN block from information that is supplied. With this CCA enhancement, PIN_Change/Unblock adds support for the message format used to change or unblock the PIN on American Express EMV cards.
Chapter 6. Cryptography
203
8049ch06.fm
workstation feature code 0841 was the first to have the 4765 crypto card installed. TKE LIC V7.2 requires CEX4 or CEX3 and TKE workstation feature code 0841. Adapters: The TKE workstation supports Ethernet adapters only to connect to a LAN. A TKE workstation is part of a customized solution for using the Integrated Cryptographic Service Facility for z/OS program product (ICSF for z/OS) or the Linux for System z that provide a basic key management system for cryptographic keys of a zEC12 that has Crypto Express features installed and that is configured for using DES, AES, ECC and PKA cryptographic keys. The TKE provides a secure, remote and flexible method of providing Master Key Part Entry, and to remotely manage PCIe Cryptographic Coprocessor(s). The cryptographic functions on the TKE are performed by one PCIe Cryptographic Coprocessor. The TKE workstation communicates with the System z server using a TCP/IP connection. The TKE workstation is available with Ethernet LAN connectivity only. Up to ten TKE workstations can be ordered. TKE feature number 0841 can be used to control the zEC12 and it can also be used to control z196, z114, z10 EC, z10 BC, z9 EC, z9 BC, z990, and z890 servers.
204
8049ch06.fm
Contain at least 2 numeric and 2 non-numeric characters Not contain the user ID These rules are enforced when you define a new user profile for passphrase logon, or when you change the passphrase for an existing profile. Your current passphrases will continue to work. Simplified TKE usability with Crypto Express migration wizard: A wizard is now available to allow users to collect data, including key material, from a Crypto Express coprocessor and migrate the data to a different Crypto Express coprocessor. The target Crypto Express coprocessor must have the same or greater capabilities. This wizard is an aid to help facilitate migration from Crypto Express2 to Crypto Express3. Crypto Express2 is not supported on zEC12 neither on z196 and z114. The following benefits are obtained when using this wizard: Reduces migration steps, thereby minimizing user errors Minimizes the number of user clicks Significantly reduces migration task duration
Chapter 6. Cryptography
205
8049ch06.fm
CCA V4.2 for the Crypto Express feature includes support for 100 decimalization tables for each domain on a Crypto Express feature. From the TKE, users can manage the decimalization tables on the Crypto Express feature from the TKE workstation crypto module notebook. Users can manage the tables for a specific domain or manage the tables of a set of domains if they are using the TKE workstation Domain Grouping function. Host cryptographic module status support: From the TKE workstation crypto module notebook, users will be able to display the current status of the host cryptographic module that is being managed. If they view the Crypto Express feature module information from a crypto module group or a domain group, they will see only the status of the group's master module. Display of active IDs on the TKE console: A user can be logged onto the TKE workstation in privileged access mode. In addition, the user can be signed onto the TKE workstation's local cryptographic adapter. If a user is signed on in privileged access mode, that ID is shown on the TKE console. With this new support, both the privileged access mode ID and the TKE local cryptographic adapter ID will be displayed on the TKE console. Increased number of key parts on smart card: If a TKE smart card is initialized on a TKE workstation with a 7.1 level of LIC, it will be able to hold up to 50 key parts. Previously, TKE smart cards could hold only 10 key parts. Use of ECDH to derive shared secret: When the TKE workstation with a 7.1 level of LIC exchanges encrypted material with a Crypto Express card at CCA level V4.2, Elliptic Curve Diffie-Hellman (ECDH) is used to derive the shared secret. This increases the strength of the transport key used to encrypt the material.
206
8049ch06.fm
CCA supports both 16-byte and 24-byte DES master keys. The DES master key length for a domain is determined by a new domain control bit that can be managed using the TKE. Two Access Control Points (ACP) allow the user to choose between warning or prohibiting the loading of a weak Master Key. The latest CCA version is required. Protect generated RSA keys with AES importer keys TKE generated RSA keys are encrypted by AES keys before being sent to System z. It allows the generation for 2046-bit and 4096-bit RSA keys for target crypto card use. New Data Encryption Standard (DES) operational keys: Four new DES operational keys can be managed from the TKE workstation (#0841). The key types are: CIPHERXI CIPHERXL CIPHERXO DUKPT-KEYGENKY
The new keys are managed the same way as any other DES operational key. New Advanced Encryption Standard (AES) CIPHER key attribute: A new attribute, key can be used for data translate only can now be specified when creating an AES CIPHER operational key part. Allow creation of corresponding keys: There are some cases where operational keys need to be loaded to different host systems to serve an opposite purpose. For example, one host system needs an exporter key encrypting key; another system needs a corresponding importer key encrypting key with the same value. The TKE workstation now allows nine types of key material to be used for creating a corresponding key. Support for four smart card readers: The TKE workstation supports two, three, or four smart card readers when smart cards are being used. The additional readers were added to help reduce the number of smart card swaps needed while managing EP11 configured coprocessors. EP11 can be managed with only two smart card readers. CCA configured coprocessors can be managed with three or four smart card readers.
207
8049ch06.fm
workstation, and 20 smart cards. The reader supports the use of smart cards that contain an embedded microprocessor and associated memory for data storage that can contain the keys to be loaded into the Crypto Express features. Access to and use of confidential data on the smart card is protected by a user-defined personal identification number (PIN). Up to 990 additional smart cards can be ordered for backup. The additional smart card feature code is FC 0884, and when one feature code is ordered, a quantity of ten smart cards are shipped. The order increment is one up to 99 (990 blank smart cards).
CPACF
a
CEX4Ca X X X X X
CEX4Pa X X X
CEX4Aa X X X X X -
CEX3Cab X X X X X
CEX3Aab X X X X X -
Supports z/OS applications using ICSF Supports Linux on System z CCA applications Encryption and decryption using secret-key algorithm Provides highest SSL/TLS handshake performance Supports SSL/TLS functions Provides highest symmetric (clear key) encryption performance Provides highest asymmetric (clear key) encryption performance Provides highest asymmetric (encrypted key) encryption performance Disruptive process to enable Requires IOCDS definition Uses CHPID numbers Uses PCHIDs Requires CPACF enablement (FC 3863) Requires ICSF to be active Offers user programming function (UDX) Usable for data privacy: encryption and decryption processing
X X X X -
Note c Xd
Note c Xd Xe X X
Note c Xd Xe X -
Note c Xd Xe X X X
Note c Xd Xe X -
Xe X
Xe X X X
208
8049ch06.fm
Functions or attributes
CPACF
a
CEX4Ca X X X X X X X X X X X
CEX4Pa X X X X X X X X X X
CEX4Aa X Xf X -
CEX3Cab X X X X X X X X X X X
CEX3Aab X Xf X X -
Usable for data integrity: hashing and message authentication Usable for financial processes and key management operations Crypto performance RMF monitoring Requires system master keys to be loaded System (master) key storage Retained key storage Tamper-resistant hardware packaging Designed for FIPS 140-2 Level 4 certification Supports Linux applications doing SSL handshakes RSA functions High performance SHA-1 and SHA2 Clear key DES or triple DES Advanced Encryption Standard (AES) for 128-bit, 192-bit, and 256-bit keys Pseudorandom number generator (PRNG) Clear key RSA Europay Mastercard VISA (EMV) support Public Key Decrypt (PKD) support for Zero-Pad option for clear RSA private keys Public Key Encrypt (PKE) support for MRP function Remote loading of initial keys in ATM Improved key exchange with non CCA system ISO 16609 CBC mode triple DES MAC support
X X X X
X -
X X X
X -
X X
X X X
X X
X X X X
X -
X X X X
X -
a. This requires CPACF enablement feature code 3863. b. Available only in a carry forward basis when upgrading from earlier generations to zEC12. c. To make the addition of the Crypto Express features nondisruptive, the logical partition must be predefined with the appropriate PCI Express cryptographic adapter number selected in its candidate list in the partition image profile. d. One PCHID is required for each PCIe cryptographic adapter.
Chapter 6. Cryptography
209
8049ch06.fm
e. This is not required for Linux if only RSA clear key operations are used. DES or triple DES encryption requires CPACF to be enabled. f. This is physically present but is not used when configured as an accelerator (clear key only).
210
8049ch07.fm
Chapter 7.
211
8049ch07.fm
The number of chassis and blades varies depending on the type of the blades configured within zBX. See 7.2.4, zBX blades on page 218 for more information.
212
8049ch07.fm
You can read more about zBX reliability, availability, and serviceability (RAS) in 10.6, RAS capability for zBX on page 370. The zBX can be ordered with a new zEC12 or as an MES to an existing zEC12. If a z196 controlling a zBX is upgraded to a zEC12, the controlled zBX Model 002 must be upgraded to a Model 003. Either way, the zBX is treated as an extension to a zEC12 and cannot be ordered as a standalone feature. Figure 7-1 shows a zEC12 with a maximum zBX configuration. The first rack (Rack B) in the zBX is the primary rack where one or two BladeCenter chassis and four top of rack (TOR) switches reside. The other three racks (C, D, and E) are expansion racks with one or two BladeCenter chassis each.
zEnterprise EC12
14 blade server slots per chassis 1 Gbps Ethernet switch modules (ESM) 10 Gbps High speed switch Ethernet (HSS) modules 8 Gbps Fibre Channel switches for connectivity to the SAN environment2 Blower modules
Client supplied FC switches are required and must support N_Port ID Virtualization (NPIV). Some FC switch vendors also require interop mode. Check the interoperability matrix for the latest details, at this website: http://www-03.ibm.com/systems/support/storage/ssic/interoperability.wss
213
8049ch07.fm
TOR Intranode Management Network switches TOR Intraensemble Data Netw ork switches
A zBX rack supports a maximum of two BladeCenter chassis. Each rack is designed for enhanced air flow and is shipped loaded with the initial configuration. It can be upgraded on-site. The zBX racks are shipped with lockable standard non-acoustic doors and side panels. The following optional features are also available: IBM rear door heat eXchanger (FC 0540) reduces the heat load of the zBX emitted into ambient air. The rear door heat eXchanger is an air-to-water heat exchanger that diverts heat of the zBX to chilled water (customer supplied data center infrastructure). The rear door heat eXchanger requires external conditioning units for its use. IBM acoustic door (FC 0543) can be used to reduce the noise from the zBX. Height reduction (FC 0570) reduces the rack height to 36U high and accommodates doorway openings as low as 1832 mm (72.1 inches). Order this choice if you have doorways with openings less than 1941 mm (76.4 inches) high.
214
8049ch07.fm
A zBX Model 003 can only be managed by one zEC12 through the INMN connections. Each VLAN-capable 1000BASE-T switch has 48 ports. The switch ports are reserved as follows: One port for each of the two bulk power hubs (BPH) on the controlling zEC12 One port for each of the advanced management modules (AMM) and Ethernet switch modules (ESM), in each zBX BladeCenter chassis One port for each of the two IEDN 10 GbE TOR switches Two ports each for interconnecting the two switches Both switches have the same connections to the corresponding redundant components (BPH, AMM, ESM, and IEDN TOR switches) to avoid any single point of failure. Table 7-5 on page 228 shows port assignments for the 1000BASE-T TOR switches. Important: Although IBM provides a 26m cable for the INMN connection, it is best to have the zBX installed next to or near the controlling zEC12, for easy access to the zBX for service related activities or tasks. Each VLAN-capable 10 GbE TOR switch has 40 ports dedicated to the IEDN. The switch ports have the following connections: Up to 16 ports are used for connections to an HSS module (SM07 and SM09) of each BladeCenter chassis in the same zBX (as part of IEDN), to provide data paths to blades. Up to eight ports are used for OSA-Express4S 10 GbE or OSA-Express3 10 GbE (LR or SR) connections to the ensemble CPCs (as part of IEDN), to provide data paths between the ensemble CPCs and the blades in a zBX. Up to seven ports are used for zBX to zBX connections within a same ensemble (as part of the IEDN). Up to seven ports are used for the customer managed data network. Customer network connections are not part of IEDN, and cannot be managed or provisioned by the Unified Resource Manager. The Unified Resource Manager will recognize them as migration connections and provide access control from the customer network to the 10 GbE TOR switches. The management port is connected to the INMN 1000BASE-T TOR switch. Two ports are used for interconnections between two switches (as a failover path), using two Direct Attached Cables (DAC) to interconnect both switches. Figure 7-3 shows the connections of TOR switches and only the first BladeCenter chassis in frame B. For more information about the connectivity options for the INMN and the IEDN, as well as the connectivity rules, see 7.4, zBX connectivity on page 226.
215
8049ch07.fm
P06 P07
OSM
B P H A
47 01
P03 P04
45 44 46
00
<- Port# B36P INMN <- Port# B35P INMN <- Port# B32P INMN <- Port# B30P INMN
03
47
01 45 44
46
00
P04 P07
OSM
B P H B
P03 P06
08 16 41
38 39
08 16 41 38 39
01 02 S M 0 1 SM07 01 10 09 M M 0 1
S M 01 0 2
10 09 SM09
01
02
M M 0 2
03
216
8049ch07.fm
2
SM01 (Bay 1)
SM07 (Bay 7)
1
SM03 (Bay 3) MM01
Blower Module 1
The rear of a zBX BladeCenter chassis has following components: Advanced management module: The advanced management module (AMM) provides systems-management functions and keyboard/video/mouse (KVM) multiplexing for all of the blade servers in the BladeCenter unit that support keyboard/video/mouse (KVM). It controls the external keyboard, mouse, and video connections, for use by a local console, and a 10/100 Mbps Ethernet remote management connection. Use of KVM is not supported on zBX. All the required management function are available on the controlling zEC12 support element (SE) or the HMC. The management module communicates with all components in the BladeCenter unit, detecting their presence, reporting their status, and sending alerts for error conditions when required. The service processor in the management module communicates with the service processor (iMM) in each blade server to support features such as blade server power-on requests, error and event reporting, keyboard/video/mouse (KVM) requests, and requests to use the BladeCenter shared media tray. The AMMs are connected to the INMN through the 1000BASE-T TOR switches. Thus, firmware and configuration for the AMM is controlled by the SE of the controlling zEC12, together with all service management and reporting function of AMMs. Two AMMs (MM01 and MM02) are installed in the zBX BladeCenter chassis. Only one AMM has primary control of the chassis (it is active); the second module is in passive (standby) mode. If the active module fails, the second module is automatically enabled with all of the configuration settings of the primary module. Ethernet switch module: Two 1000BASE-T (1 Gbps) Ethernet switch modules (SM01 and SM02) are installed in switch bays 1 and 2 in the chassis. Each ESM has 14 internal full-duplex Gigabit ports,
SM02 (Bay 2)
SM04 (Bay 4)
MM02
Blower Module 2
SM09 (Bay 9)
217
8049ch07.fm
one connected to each of the blade servers in the BladeCenter chassis, two internal full-duplex 10/100 Mbps ports connected to the AMM modules, and six 1000BASE-T copper RJ-45 connections for INMN connections to the TOR 1000BASE-T switches. The ESM port 01 is connected to one of the 1000BASE-T TOR switches. As part of the INMN, configuration and firmware of ESM is controlled by the controlling zEC12 Support Element (SE). High speed switch module: Two high-speed switch modules (SM07 and SM09) are installed to the switch bay 7 and 9. The HSS modules provide 10 GbE uplinks to the 10 GbE TOR switches and 10 GbE downlinks to the blades in the chassis. Port 01 and 02 are connected to one of the 10 GbE TOR switches. Port 09 and 10 are used to interconnect HSS in bay 7 and 9 as a redundant failover path. 8 Gbps Fibre Channel switch module: Two 8 Gbps Fibre Channel (FC) switches (SM03 and SM04) are installed in switch bays 3 and 4. Each switch has 14 internal ports reserved for the blade servers in the chassis, and six external Fibre Channel ports to provide connectivity to SAN environment. Blower module: There are two hot swap blower modules installed. The blower speeds vary depending on the ambient air temperature at the front of the BladeCenter unit and the temperature of internal BladeCenter components. If a blower fails, the remaining blowers run full speed. BladeCenter mid-plane fabric connections: The BladeCenter mid-plane provides redundant power, control, and data connections to a blade server by internally routed chassis components (power modules, AMMs, switch modules, media tray) to connectors in a blade server slot. There are six connectors in a blade server slot on the mid-plane, from top to bottom: Top 1X fabric connects blade to MM01, SM01, and SM03 Power connector from power module 1 (blade server slot 1 to 7) or power module 3 (blade server slot 8 to 14) Top 4X fabric connects blade to SM07 Bottom 4X fabric connects blade to SM09 Bottom 1X fabric connects blade to MM02, SM02 and SM04 Power connector from power module 2 (blade server slot 1 to 7) or power module 4 (blade server slot 8 to 14) Thus, each blade server has redundant power, data, and control links from separate components.
218
8049ch07.fm
Up to 56 IBM System x HX5 blades are supported. All zBX blades are connected to AMMs and ESMs through the chassis mid-plane. The AMMs are connected to the INMN.
POWER7 blade
The POWER7 blade (Table 7-1) is a single width blade, which includes a POWER7 processor, up to 16 DIMMs, and an HDD (Hard Disk Drive). The POWER7 blade supports 10 GbE connections to IEDN, and 8 Gbps FC connections to client provided Fibre Channel storage through the Fibre Channel (FC) switches (SM03 and SM04) in the chassis. The POWER7 blade is loosely integrated to a zBX, so that you can acquire supported blades through existing channels from IBM. The primary HMC and SE of the controlling zEC12 perform entitlement management for installed POWER7 blades on a one-blade basis. PowerVM Enterprise Edition must be ordered with each POWER blade. AIX 5.3, AIX 6.1 and AIX 7.1 and subsequent releases are supported.
Table 7-1 Supported configuration of POWER7 blades Feature FC Config 1 quantity Config 2 quantity Config 3 quantity
Processor (3.0GHz@150W) 8 GB Memory 16 GB Memory Internal HDD (300GB) CFFh 10 GbE expansion CIOv 8 Gb FC expansion
8 4 0 1 1 1
8 8 0 1 1 1
8 0 8 1 1 1
219
8049ch07.fm
Enables SOA and XML applications with System z web services for seamless integration of distributed and System z platforms. It can help to simplify, govern, and enhance the network security for XML and web services. Provides drop-in integration for heterogeneous environments by enabling core Enterprise Service Bus (ESB) functionality, including routing, bridging, transformation, and event handling. Offers standards-based, centralized System z governance, and extreme reliability through integrated operational controls, call home, and integration with RACF security through a secured private network. The zBX provides additional benefits to the DataPower appliance environment in these areas: Blade hardware management: Improved cooling and power management controls, includes cooling of the frame and energy monitoring and management of the DataPower blades Virtual network provisioning Call-home for current and expected problems. Hardware Management Console integration: Single view showing the System z environment together with the DataPower blades in an overall hardware operational perspective Group GUI operations for functions supported on HMC such as activate or deactivate blades Improved availability: Guided placement of blades to optimize built-in redundancy in all components at the rack, BladeCenter, and HMC levels, including top of rack switch, ESM switches, and physical network. Detection and reporting by the HMC/SE on appliance failures. The HMC/SE can also be used to re-cycle the DataPower appliance. Networking: Virtual network provisioning Enforced isolation of network traffic by VLAN support 10 Gbps end-to-end network infrastructure Built-in network redundancy Network protection vie IEDN, possibly obviating any perceived need for encryption of flows between DataPower and the target back-end System z server
Monitoring and reporting: Monitoring and reporting of DataPower hardware health and degraded operation by HMC Monitoring of all hardware, call-home, and auto-parts replacement Consolidation and integration of DataPower hardware problem reporting with other problems reported in zBX System z value: Simplified ordering of the DataPower appliance by System z allows the proper blade infrastructure to be transparently ordered. Simplified upgrades keep MES history so the upgrades flow based on what is installed. System z service on the zBX and DataPower blade with a single point of service. The DataPower appliance becomes part of the data center and comes under data center control.
220
8049ch07.fm
In addition, although not specific to the zBX environment, dynamic load balancing to DataPower appliances is available using the z/OS Communications Server Sysplex Distributor.
Configuration
The DataPower XI50z is a double-wide IBM HS22 blade. Each one takes two BladeCenter slots, so the maximum number of DataPower blades per BladeCenter is seven, and the maximum number of DataPower blades per zBX is 28. It can coexist with POWER7 blades and with IBM BladeCenter HX5 blades in the same zBX BladeCenter. DataPower XI50z blades are configured and ordered as zBX (machine type 2458-003) features, but have their own machine type (2462-4BX). The DataPower XI50z with DataPower expansion unit has the following specification: 2.13GHz 2x quad core processors 8M cache 3 x 4Gb DIMMs (12Gb Memory) 4Gb USB Flash Key that contains the DataPower XI50z firmware load 2 x 300GB HDDs used by the customer for logging, storing style sheets, and storing XML files. The hard disk array consists of two hard disk drives in a RAID-1 (mirrored) configuration. Broadcom BCM5709S x2 with TOE (integrated on planar) BPE4 Expansion Unit. It is sealed FRU with one-way tamper-proof screws (contains the crypto for secure SOA applications). XG5 accelerator PCIe card CN1620 Cavium crypto PCIe card Dual 10Gb Ethernet card
221
8049ch07.fm
0001 License with 1 year SWMA 0002 Option for TIBCO 0003 Option for Application Optimization 0004 Option for Database Connectivity 0005 Option for Tivoli Access Manager Every IBM 2462 Model 4BX includes feature codes 0001, 0003, and 0005 (they are optional on DataPower XI50B). Optional Software feature codes 0002 and 0004 are required if FC 0652 TIBCO or FC 0653 Database Connectivity are ordered. The TIBCO option (FC 0002) lets you extend the DataPower XI50z so you can send and receive messages from TIBCO Enterprise Message Service (EMS). Option for Database Connectivity (FC 0004) lets you extend the DataPower XI50z to read and write data from relational databases such as IBM DB2, Oracle, Sybase, and Microsoft SQL Server. For software PID number 5765-G85 (registration and renewal), every IBM 2462 Model 4BX includes feature code 0001. Feature code 0003 is available at the end of the first year to renew SW maintenance for one more year. For software PID number 5765-G86 (maintenance reinstatement 12 months), feature code 0001 is available if SW PID 5765-G85 feature code 0003 was not ordered before the one year expired. For software PID number 5765-G87 (3-year registration) feature code 0001 can be ordered instead of SW PID 5765-G85 feature code 0003 to make the initial period three years, instead of one year. For software PID number 5765-G88 (3-year renewal), feature code 0001 can be used as alternative SW PID 5765-G85 feature code 0003 if a three year renewal is desired. The maximum duration is five years. For software PID number 5765-G89 (3-year after license), feature code 0001 is available if SW PID 5765-G85 feature code 0003 was not ordered before the one year expired if a three year renewal is desired.
8049ch07.fm
Support of select IBM System x blades in the zBX allows the zEnterprise to access a whole new application portfolio. Front-end applications that need access to centralized data serving would be a good fit for running on the blades, as well as applications that are a front end to core CICS or IMS transaction processing such as IBM WebSphere. BladeCenter HX5 blades are customer acquired through existing channels or through IBM. POWER7, DataPower XI50z, and System x blades can be in the same BladeCenter chassis. Supported configuration options are listed in Table 7-2 on page 223. IBM BladeCenter HX5 7873 is a dual-socket 16-core blade with the following features: Intel 8-core processor Two processor sockets 2.13 GHz 105W processor Max 14 A16Ms per BC-H Memory: up to 16 DIMM DDR-3 with 6.4 GTs 100 GB SSD Internal Disk
Table 7-2 Supported configuration of System x blades System x blades Part number Feature code Config 0 Config 1 Config 2 Config 3
Blades base - HX5 Processor 2.13 GHz 105W Intel Processors Blade width Total Cores Memory kits 8 GB 1333 MHz 16 GB 1333 MHz GB/core Speed Burst SSD Exp Card 50GB SSD MLC No Internal Raid CFFh 10GbE CIOv 8Gb FC
1 1 1 2 Single 16
1 1 1 2 Single 16 16 0 8 1 1 2 1 1 1
1 1 1 2 Single 16 8 8 12 1 1 2 1 1 1
1 1 1 2 Single 16 0 16 16 1 1 2 1 1 1
46C0558 49Y1527
A17Q 2422
8 0 4
1 1 2 1 1 1
223
8049ch07.fm
flag, and only a blade quantity equal to or less than installed in the zBX can communicate with the CPC. Also, Unified Resource Manager has two management suites, Manage suite (FC 0019) and Automate/Advanced Management Suite (FC 0020). If the controlling zEC12 has Manage suite (FC 0019), then the same quantity entered for any blade enablement feature code (FC 0611, FC 0612 or FC 0613) will be used for Manage Firmware (FC 0047, FC 0048, or FC 0049) of the corresponding blades. If the controlling zEC12 has Automate/Advanced Management Suite (FC 0020), then the same quantity entered for Blade Enablement feature codes (FC 0611, FC 0612, or FC 0613) will be used for the Manage Firmware (FC 0047, FC 0048, FC 0049) and Automate/Advanced Management Firmware (FC 0050, FC 0051, FC 0053) of the corresponding blades. Table 7-3 lists these features, while Table 7-4 shows maximum quantities for these feature codes. Minimum quantity to order depends on the number of corresponding blades configured and located in the zBX Model 003.
Table 7-3 Feature codes for blade enablements and Unified Resource Manager suites Blade Enablement Manage - per connection Automate/Advanced Management - per connection
z/OS only IFL DataPower XI50z POWER7 Blade IBM System x HX5 Blade
Blades: If any attempt is made to install additional blades that exceed the FC 0611, FC 0612, or FC 0613 count, those blades will be not be powered on by the system. The blades will also be checked for minimum hardware requirements.
Table 7-4 Maximum quantities for Unified Resource Manager feature codes Feature code Maximum quantity Feature Description
Manage firmware Data Power Automate/Advanced Management firmware Data Power Manage firmware POWER blade Automate/Advanced Management firmware POWER blade Manage firmware System x blade Advanced Management firmware System x blade Automate/Advanced Management firmware IFL
Note that FC 0047, FC 0048, FC 0049, FC 0050, FC 0051, FC 0053, and FC 0054 are priced features. In order to get ensemble member management and cables, FC 0025 should also be ordered on the zEC12. 224
IBM zEnterprise EC12 Technical Guide
8049ch07.fm
Feature codes are available to detach a zBX from an existing CPC and attach a zBX to another CPC. FC 0030 indicates that the zBX will be detached, while FC 0031 indicates that the detached zBX is going to be attached to a CPC. Only zBX Model 003 (2458-003) is supported with zEC12. When upgrading a z196 with zBX Model 002 attachment to zEC12, the zBX Model 002 (2458-002) must be upgraded to a zBX Model 003 as well. Upgrades from zBX Model 002 to zBX Model 003 are disruptive. A zBX Model 003 has the following improvements compared the zBX Model 002: New AMM in BladeCenter chassis with enhanced functions Additional ethernet connectivity in IEDN network for redundancy and higher bandwidth between TOR switches and BladeCenter switch modules. New Firmware level with improved functionality
225
8049ch07.fm
zEnterprise node
T OR Sw itch es T OR Switches
OSM
OSX
zBX
Figure 7-5 INMN, IEDN, customer managed local area network
The IEDN provides private and secure 10 GbE high speed data paths between all elements of a zEnterprise ensemble (up to eight zEC12 with optional zBXs). The zBX is managed by the HMC through the physically isolated INMN, which interconnects all resources of the zEC12 and zBX components.
Carry forward only for zEC12 when upgrading from earlier generations.
226
8049ch07.fm
INMN communication
Communication across the INMN is exclusively for the purpose of enabling the Unified Resource Manager of the HMC to perform its various management disciplines (virtual server, performance, network virtualization, energy, storage management, and so on) for the node. The zEC12 connection to the INMN is achieved through the definition of a CHPID type OSM, which can be defined over an OSA1000BASE-T Ethernet feature. There is also a 1 GbE infrastructure within the zBX.
INMN configuration
The key points to consider for an INMN are as follows: Each zEC12 must have two OSA 1000BASE-T ports connected to the Bulk Power Hub in the same zEC12: The two ports provide a redundant configuration for failover purposes in case one link fails. For availability, each connection must be from two different OSA 1000BASE-T features within the same zEC12. Figure 7-6 shows both the OSA 1000BASE-T features and required cable type.
1 CHPID p er feature
2 C HPIDs p er feature
X
Port 0 Port 1 is not used with CHPID typ e=OSM Port 0 Port 1 is not used with CHPID typ e=OSM
X
OSA-Express4S 1000BASE-T Ethernet (F C 0408)
OSA-Express3 1000BASE-T Ethernet (FC 3367) Category 6 Ethernet RJ45 3.2m cab le (FC 0025)
Figure 7-6 OSA-Express4S 1000BASE-T and OSA-Express3 1000BASE-T features and cable type
OSA 1000BASE-T ports can be defined in the IOCDS as SPANNED, SHARED, or DEDICATED:
227
8049ch07.fm
DEDICATED restricts the OSA 1000BASE-T port to a single LPAR. SHARED allows the OSA 1000BASE-T port to be used by all or selected LPARs in the same zEC12. SPANNED allows the OSA 1000BASE-T port to be used by all or selected LPARs across multiple Channel Subsystems in the same zEC12. SPANNED and SHARED ports can be restricted by the PARTITION keyword in the CHPID statement to allow only a subset of LPARs in the zEC12 to use the OSA 1000BASE-T port. SPANNED, SHARED, and DEDICATED link pairs can be defined within the maximum of 16 links supported by the zBX. z/OS Communication server TCPIP stack must be enabled for IPv6. The CHPID type OSM related definitions will be dynamically created. No IPv4 address is needed, a IPv6 link local address will be dynamically applied z/VM virtual switch types provide INMN access: Up-link can be virtual machine NIC (Network Interface Card). Ensemble membership conveys UUID (Universally Unique IDentifier) and MAC (Media Access Control) prefix. Two 1000BASE-T top of rack switches in the zBX (Rack B) are used for the INMN; no additional 1000BASE-T Ethernet switches are required. Figure 7-7 shows the 1000BASE-T TOR switches.
J00 J02 J04 J06 J08 J10 J12 J14 J16 J18 J20 J22 J24 J26 J 28 J30 J32 J34 J36 J38 J40 J42 J44 J46 J01 J03 J05 J07 J09 J11 J13 J15 J17 J19 J21 J23 J25 J27 J 29 J31 J33 J35 J37 J39 J41 J43 J45 J47
J00 J02 J04 J06 J08 J10 J12 J14 J16 J18 J20 J22 J24 J26 J 28 J30 J32 J34 J36 J38 J40 J42 J44 J46 J01 J03 J05 J07 J09 J11 J13 J15 J17 J19 J21 J23 J25 J27 J 29 J31 J33 J35 J37 J39 J41 J43 J45 J47
The port assignments for both 1000BASE-T TOR switches are listed in Table 7-5.
Table 7-5 Port assignments for the 1000BASE-T TOR switches Ports Description
Management for BladeCenters located in zBX Rack-B Management for BladeCenters located in zBX Rack-C Management for BladeCenters located in zBX Rack-D Management for BladeCenters located in zBX Rack-E Not used INMN switch B36P(Top) to INMN switch B35P(Bottom) INMN-A to IEDN-A port J41 / INMN-B to IEDN-B port J41 INMN-A to zEC12 BPH-A port J06 / INMN-B to zEC12 BPH-B port J06
228
8049ch07.fm
3.2 meter Category 6 Ethernet cables are shipped with the zEC12 ensemble management flag feature (FC 0025). Those cables connect the OSA 1000BASE-T (OSM) ports to the Bulk Power Hubs (port 7). 26 meter Category 5 Ethernet cables are shipped with the zBX. Those cables are used to connect the zEC12 Bulk Power Hubs (port 6) and the zBX top of rack switches (port J47).
Customer Network 1
J01 J05
SE
SE
229
8049ch07.fm
Table 7-6 shows the port assignments for both bulk power hubs (BPHs).
Table 7-6 Port assignments for the BPHs BPH A Port # Connects to: Port # BPH B Connects to:
HMC to SE Customer Network2 (VLAN 0.40) HMC to SE Customer Network1 (VLAN 0.30) BPH B J03 BPH B J04 SE A-Side (Top SE) zBX TOR Switch B36P, Port 47 (INMN-A) OSA 1000BASE-T (CHPID type OSM) Not used Used for internal zEC12 components
HMC to SE Customer Network2 (VLAN 0.40) HMC to SE Customer Network1 (VLAN 0.30) BPH A J03 BPH A J04 SE B-Side (Bottom SE) zBX TOR Switch B35P, Port 47 (INMN-B) OSA 1000BASE-T (CHPID type OSM) Not used Used for internal zEC12 components
For more information, see Chapter 12, Hardware Management Console and Support Element on page 397.
IEDN configuration
The IEDN connections can be configured in a number of ways. The key points to consider for an IEDN are as follows: Each zEC12 must have a minimum of two OSA 10 GbE ports connected to the zBX through the IEDN: The two ports provide a redundant configuration for failover purposes in case one link fails. For availability, each connection must be from a different OSA 10 GbE feature within the same zEC12. The zBX can have a maximum of 16 IEDN connections (8 pairs of OSA 10 GbE ports). 230
IBM zEnterprise EC12 Technical Guide
8049ch07.fm
Four connections between IEDN TOR switches and high speed switch modules in each BladeCenter chassis (2 pairs of 10GbE ports) For redundancy two connections between both high speed switch modules in each BladeCenter. Figure 7-9 shows the OSA 10 GbE feature (long reach or short reach) and the required fiber optic cable types.
Figure 7-9 OSA-Express4S 10 GbE and OSA-Express3 10 GbE features and cables
OSA 10 GbE ports can be defined in the IOCDS as SPANNED, SHARED, or DEDICATED: DEDICATED restricts the OSA 10 GbE port to a single LPAR. SHARED allows the OSA 10 GbE port to be used by all or selected LPARs in the same zEC12. SPANNED allows the OSA 10 GbE port to be used by all or selected LPARs across multiple Channel Subsystems (CSSs) in the same zEC12. SHARED and SPANNED ports can be restricted by the PARTITION keyword in the CHPID statement to allow only a subset of LPARs on the zEC12 to use OSA 10 GbE port. SPANNED, SHARED, and DEDICATED link pairs can be defined within the maximum of 16 links supported by the zBX z/OS Communication Server requires minimal configuration: IPv4 or IPv6 addresses are used. VLAN must be configured to match HMC (Unified Resource Manager) configuration. z/VM virtual switch types provide IEDN access: Up-link can be a virtual machine NIC. Ensemble membership conveys Ensemble UUID and MAC prefix. IEDN network definition are completed from the primary HMC Manage Virtual Network task.
231
8049ch07.fm
Two 10 GbE top of rack switches in the zBX (Rack B) are used for the IEDN, no additional Ethernet switches are required. Figure 7-10 on page 232 shows the 10 GbE TOR switches.
The IBM zEnterprise EC12 provides the capability to integrate HiperSockets connectivity to the intraensemble data network (IEDN). This extends the reach of the HiperSockets network outside the CPC to the entire ensemble, appearing as a single, Layer 2. Because HiperSockets and IEDN are both internal System z networks, the combination allows System z virtual servers to use the optimal path for communications. The support of HiperSockets integration with the IEDN function is available on z/OS Communication Server V1R13 and z/VM V6R2.
Port assignments
The port assignments for both 10 GbE TOR switches are listed in Table 7-7.
Table 7-7 Port assignments for the 10 GbE TOR switches Ports Description
J00 - J07 J08 - J23 J24 - J30 J31 - J37 J38 - J39 J40 J41
SFP and reserved for zEC12 (OSX) IEDN connections DAC reserved for BladeCenter SM07/SM09 IEDN connections SFP reserved for zBX-to-zBX IEDN connections SFP reserved for client IEDN connections (OSD) DAC for TOR switch-to-TOR switch IEDN communication RJ-45 (not used) RJ-45 IEDN Switch Management Port to INMN TOR switch port 46
All IEDN connections must be point-to-point to the 10 GbE switch: The IEDN connection uses MAC address, not IP address (Layer 2 connection). No additional switches or routers are needed. This limits the distances CPCs the 10 GbE TOR switches in an ensemble. The 10 GbE TOR switches utilize small form factor plugable (SFP) optics for the external connections and direct attach cables (DAC) for connections, as follows: Ports J00-J07 are reserved for the zEC12 OSX IEDN connections. These ports utilize SFPs (Small Form Factor Pluggable Modules) plugged according to the zBX order: FC 0632 LR SFP to FC 0406 OSA-Express4S 10 GbE LR
232
8049ch07.fm
FC 0633 SR SFP to FC 0407 OSA-Express4S 10 GbE SR FC 0632 LR SFP to FC 3370 OSA-Express3 10 GbE LR FC 0633 SR SFP to FC 3371 OSA-Express3 10 GbE SR Ports J08-J23 are reserved for IEDN to BladeCenter attachment. The cables used are Direct Attached Cables (DAC) and are included with the zBX. These are hard wired 10 GbE SFP cables. The Feature codes indicate the length of the cable: FC 0626 - one meter for Rack B BladeCenters and IEDN to IEDN FC 0627 - five meters for Rack C BladeCenter FC 0628 - seven meters for Racks D and E BladeCenters Ports J31-J37 are reserved for the zEC12 OSD IEDN connections. These ports utilize SFPs (Small Form Factor Pluggable Modules) plugged according to the zBX order.10 GbE fiber optic cable types and maximum distance: Client provides all IEDN cables (except for zBX internal connections). Multimode fiber: 50 micron fiber at 2000 MHz-km: 300 meters 50 micron fiber at 500 MHz-km: 82 meters 62.5 micron fiber at 200 MHz-km: 33 meters Single mode fiber: 10 km
233
8049ch07.fm
The IEDN is built on a flat network design (same IPv4 or IPv6 network) and each server accessing the IEDN must be an authorized Virtual Server and must belong to an authorized Virtual LAN within the physical IEDN. VLAN enforcement sits within the hypervisor functions of the ensemble; controls reside in the OSA (CHPID type OSX), in the z/VM VSWITCH, and in the VSWITCH hypervisor function of the blades on the zBX. The VLAN IDs and the Virtual MACs that are assigned to the connections from the Virtual Servers are tightly controlled through the Unified Resource Manager and thus there is no chance of either MAC or VLAN spoofing for any of the servers on the IEDN. If a you decide to attach your network to the TOR switches of the zBX in order to communicate with the Virtual Servers on the zBX blades, access must be authorized in the TOR switches (MACor VLAN-based). Although the TOR switches will enforce the VMACs and VLAN IDs here, you must take the usual network security measures to ensure that the devices in the Customer Managed Data Network are not subject to MAC or VLAN spoofing; the Unified Resource Manager functions cannot control the assignment of VLAN IDs and VMACs in those devices. In other words, whenever you decide to interconnect the external network to the secured IEDN, the security of that external network must involve all the usual layers of the IBM Security Framework: physical security, platform security, application and process security, data and information security, and so on. The INMN and the IEDN are both subject to Network Access Controls as implemented in z/OS and z/VM; so that not just any virtual server on the zEC12 that can utilize these networks. INMN is not accessible at all from within the virtual servers. Although we deem it unnecessary to implement firewalls, IP filtering, or encryption for data flowing over the IEDN, if company policy or security require such measures to be taken, then these are supported. You can implement any of the security technologies available: for example, SSL/TLS or IP filtering. The centralized and internal network design of both the INMN and the IEDN limit the vulnerability to security breaches. Both networks reduce the amount of network equipment and administration tasks, as well as routing hops that are under the control of multiple individuals and subject to security threats. Both use IBM-only equipment (switches, blades) which have been tested previously and in certain cases pre-installed. In summary, many more technologies than in the past have been architected in a more robust, secure fashion to integrate into the client network. This is achieved with the help of either the Unified Resource Manager, or additional SAF controls specific to zEnterprise System and the ensemble, such as these: MAC filtering VLAN enforcement ACCESS control Role-based security The following standard security implementations are still available for use in the IEDN: Authentication Authorization and access control (including Multilevel Security (MLS); also firewall IP filtering. Only stateless firewalls or IP filtering implementations can be installed in a Virtual Server in the ensemble) Confidentiality Data integrity Non-repudiation
234
8049ch07.fm
BladeCenter
Switch Module 2 SM04
Figure 7-11 8 Gb FC switch external ports
Client provided multi-mode LC duplex cables are used for the FC switch connections to support speeds of 8 Gbps, 4 Gbps, or 2 Gbps. (1 Gbps is not supported.) Maximum distance depends on the speed and fiber type. Cabling specifications are defined by the Fibre Channel - Physical Interface - 4 (FC-PI-4) standard. Table 7-8 on page 236 identifies cabling types and link data rates that are supported in the zBX SAN environment, including their allowable maximum distances and link loss budget.
235
8049ch07.fm
The link loss budget is derived from the channel insertion loss budget defined by the FC-PI-4 standard (Revision 8.00).
Table 7-8 Fiber optic cabling for zBX FC switch - maximum distances and link loss budget FC-PI-4 2 Gbps 4 Gbps 8 Gbps
Fiber core (light source) 50 m MMa (SX laser) 50 m MMb (SX laser) 62.5 m MMc (SX laser)
a. OM3: 50/125 m laser optimized multimode fiber with a minimum overfilled launch bandwidth of 1500 MHz-km at 850nm as well as an effective laser launch bandwidth of 2000 MHz-km at 850 nm in accordance with IEC 60793-2-10 Type A1a.2 fiber b. OM2: 50/125 m multimode fiber with a bandwidth of 500 MHzkm at 850 nm and 500 MHz-km at 1300 nm in accordance with IEC 60793-2-10 Type A1a.1 fiber. c. OM1: 62.5/125 m multimode fiber with a minimum overfilled launch bandwidth of 200 MHzkm at 850 nm and 500 MHz-km at 1300 nm in accordance with IEC 60793-2-10 Type A1b fiber.
Cabling: IBM does not support a mix of 50 m and 62.5 m fiber optic cabling in the same physical link.
SM03
Up to six 8 Gbps FC links
SM04
236
8049ch07.fm
Up to six external ports of each BladeCenter switch module can be used to connect to the SAN. All fiber links of a switch module have to be attached to the same SAN switch as shown in Figure 7-12. SAN switches must support NPIV to allow virtualization. The client provides all cables, FC disk storage, and FC switches. It is also the clients responsibility to configure and cable the FC switches that connect to the zBX.
237
8049ch07.fm
Primary HM C 1
Alternate HMC
O S M
4 FC Switch
2458-003 zBX
The diagram shows the following components: 1. Client provided management network: IBM supplies a 15 meter Ethernet RJ-45 cable with the 1000BASE-T (1GbE) switch (FC 0070). The 1000BASE-T switch (FC 0070) connects to the reserved client network ports of the Bulk Power Hubs in zEC12 - Z29BPS11 (on A side) and Z29BPS31 (on B side) port J02. A second switch connects to Z29BPS11 and Z29BPS31 on port J01. 2. Intranode management network: Two CHPIDs from two different OSA1000BASE-T features are configured as CHPID type OSM. IBM supplies two 3.2 meter Ethernet Category 6 cables from the OSM CHPIDs (ports) to both Z29BPS11 and Z29BPS31 on port J07. (This is a zEC12 internal connection supplied with feature code 0025.) 3. Intranode management network - extension:
238
8049ch07.fm
IBM supplies two 26 meter Category 5 Ethernet cables (chrome gray plenum rated cables) from zBX Rack B INMN-A/B switches port J47 to Z29BPS11 and Z29BPS31 on port J06. 4. Intraensemble data network: Two ports from two different OSA10 GbE (SR Short Reach or LR Long Reach) features are configured as CHPID type OSX. The client supplies the fiber optic cables (single mode or multimode). 5. 8 Gbps Fibre Channel switch: The client supplies all Fibre Channel cables (multimode) from the zBX to the attached FC switch. The client is responsible for the configuration and management of the FC switch.
1
O S M
CPC2
O S M O S X
2
O S X
2458-003 zBX
FC S witc h
IBM B lades
The diagram shows the following components: 1. Client provided management network: The client supplies an Ethernet RJ-45 cable.
239
8049ch07.fm
The 1000BASE-T switch (FC 0070) connects to the reserved client network ports of Z29BPS11 and Z29BPS31 - J02. A second switch connects to Z29BPS11 and Z29BPS31 on port J01. 2. Intranode management network: Two ports from two different OSA1000BASE-T features are configured as CHPID type OSM IBM supplies two 3.2 meter Ethernet Category 6 cables from the OSM CHPIDs (ports) to both Z29BPS11 and Z29BPS31 on port J07. (this is a zEC12 internal connection supplied with feature code 0025). 3. Intraensemble data network: Two ports from two different OSA10 GbE (SR Short Reach or LR Long Reach) features are configured as CHPID type OSX. The client supplies the fiber optic cables (single mode or multimode).
O S M
O S M
O S X
FC Switch
FC Switch
2 458 -003 zB X
1
IEDN
2 458 -003 zB X
The diagram shows the following components: Intraensemble data network: Two 10 GbE ports in the TORs are used to connect the two zBXs (10 GbE TOR switch to 10 GbE TOR switch).
240
8049ch07.fm
Up to eight CPCs can be connected to a zBX using the IEDN. Additional CPCs are added and connected to the zBX through the OSA 10 GbE (SR Short Reach or LR Long Reach) features configured as CHPID type OSX.
7.6 References
Installation details can be found in IBM zEnterprise BladeCenter Extension Model 003 Installation Manual for Physical Planning, GC27-2611 and IBM zEnterprise BladeCenter Extension Model 003 Installation Manual, GC27-2610. For details about the BladeCenter components, see IBM BladeCenter Products and Technology, SG24-7523. For more information about DataPower XI50z blades; visit this web site: http://www-01.ibm.com/software/integration/datapower/xi50z Additional documentation is available on the IBM Resource Link at this web site: http://www.ibm.com/servers/resourcelink
241
8049ch07.fm
242
8049ch08.fm
Chapter 8.
Software support
This chapter lists the minimum operating system requirements and support considerations for the zEC12 and its features. It discusses z/OS, z/VM, z/VSE, z/TPF, and Linux on System z. Because this information is subject to change, see the Preventive Service Planning (PSP) bucket for 2827DEVICE for the most current information. Also discussed is generic software support for zEnterprise BladeCenter Extension (zBX) Model 003. Support of IBM zEnterprise EC12 functions is dependent on the operating system, its version, and release. This chapter discusses the following topics: Operating systems summary on page 244 Support by operating system on page 244 Support by function on page 256 Cryptographic support on page 288 z/OS migration considerations on page 295 Coupling facility and CFCC considerations on page 297 MIDAW facility on page 299 IOCP on page 302 ICKDSF on page 303 zEnterprise BladeCenter Extension (zBX) Model 003 software support on page 304 Software licensing considerations on page 305 References on page 309
243
8049ch08.fm
No No No Yes Nod
Yes Yesc Yes Yes See Table 8-2 on page 246 Service is required. See the following box, entitled Features
a. Regular service support for z/OS V1R10 and V1R11 ended in September 2011 and September 2012, respectively. However, by ordering the IBM Lifecycle Extension for z/OS V1R10 and V1R11 product, fee-based corrective service can be obtained for up to September 2013 and September 2014, respectively. b. z/VM V5R4, V6R1 and V6R2 provide compatibility support only. z/VM 6.1 and 6.2 require an ALS at z10 or higher c. VM supports both 31-bit and 64-bit mode guests d. 64-bit distributions included 31-bit emulation layer to run 31-bit software products.
Features: Exploitation of certain features depends on a particular operating system. In all cases, PTFs might be required with the operating system level indicated. Check the z/OS, z/VM, z/VSE, and z/TPF subsets of the 2827DEVICE Preventive Service Planning (PSP) buckets. The PSP buckets are continuously updated and contain the latest information about maintenance. Hardware and software buckets contain installation information, hardware and software service levels, service guidelines, and cross-product dependencies. For Linux on System z distributions, consult the distributors support information.
8049ch08.fm
8.2.1 z/OS
z/OS Version 1 Release 12 is the earliest in-service release supporting the EC12. After September 2014, a fee-based Extended Service for defect support (for up to three years) can be obtained for z/OS V1.12. Although service support for z/OS Version 1 Release 11 ended in September of 2012, a fee-based extension for defect support (for up to two years) can be obtained by ordering the IBM Lifecycle Extension for z/OS V1.11. Similarly, IBM Lifecycle Extension for z/OS V1.10 provides fee-based support for z/OS Version 1 Release 10 until September 2013. Also note that z/OS.e is not supported on EC12 and that z/OS.e Version 1 Release 8 was the last release of z/OS.e. See Table 8-3 on page 247 for a list of supported functions and their minimum required support levels.
8.2.2 z/VM
At general availability, z/VM V5R4, V6R1 and V6R2 provide compatibility support with limited exploitation of new zEC12 functions. See Table 8-3 on page 247 for a list of supported functions and their minimum required support levels. Capacity: For the capacity of any z/VM logical partitions, and any z/VM guests, in terms of the number of IFLs and CPs, real or virtual, it is desirable that these be adjusted to accommodate the PU capacity of the EC12.
8.2.3 z/VSE
Support is provided by z/VSE V4R3 and above. Note the following considerations: z/VSE executes in z/Architecture mode only. z/VSE exploits 64-bit real memory addressing. Support for 64-bit virtual addressing is provided by z/VSE V5R1. z/VSE V5R1 requires an architectural level set specific to the IBM System z9. See Table 8-4 on page 252 for a list of supported functions and their minimum required support levels.
8.2.4 z/TPF
See Table 8-4 on page 252 for a list of supported functions and their minimum required support levels.
SLES is SUSE Linux Enterprise Server RHEL is Red Hat Enterprise Linux
245
8049ch08.fm
toleration support. Table 8-2 on page 246 shows the service levels of SUSE and Red Hat releases supported at the time of writing.
Table 8-2 Current Linux on System z distributions Linux on System z distribution z/Architecture (64-bit mode)
SUSE SLES 10 SP4 SUSE SLES 11 SP2 Red Hat RHEL 5.8 Red Hat RHEL 6.2
IBM is working with its Linux distribution partners to provide further exploitation of selected zEC12 functions in future Linux on System z distribution releases. Consider the following guidelines: Use SUSE SLES 11 or Red Hat RHEL 6 in any new projects for the zEC12. Update any Linux distributions to their latest service level before migration to zEC12. Adjust the capacity of any z/VM and Linux on System z logical partitions guests, as well as z/VM guests, in terms of the number of IFLs and CPs, real or virtual, according to the PU capacity of the zEC12.
246
8049ch08.fm
Table 8-3 EC12 functions minimum support requirements summary, part 1 Function z/OS V1 R13 z/OS V1 R12 z/OS V1 R11 z/OS V1 R10 z/VM V6 R2 z/VM V6 R1 z/VM V5 R4
zEC12 Support of Unified Resource Manager Greater than 64 PUs single system image Greater than 54 PUs single system image Support of IBM zAware zIIP zAAP zAAP on zIIP Java Exploitation of Transactional Execution Large memory (> 128 GB) Large page support Out-of-order execution Guest support for execute-extensions facility Hardware decimal floating point Zero address detection 60 logical partitions LPAR group capacity limit CPU measurement facility Separate LPAR management of PUs Dynamic add and delete logical partition name Capacity provisioning Enhanced flexibility for CoD HiperDispatch 63.75 K subchannels Four logical channel subsystems (LCSS) Dynamic I/O support for multiple LCSS Third subchannel set Multiple subchannel sets IPL from alternate subchannel set MIDAW facility
Ym Y Y Y Ym Y Y Y Ym Y Ye f Y Y
h
Ym Ym Y Y N Y Y Y N Y Y Y Yh Y Y Y Ym Y Y Y Yh Y Y Y Y Ym Y Ym Y
Ym Ym Y Y N Y Y Y N Y Y Y Yh N Y Y Ym Y Y Y Yh Y Y Y Y Ym Y Ym Y
Ym Ym Ym Y N Y Y Ym N Y Y Y Yh N Y Y Ym Y Y Y Yh Y Y Y Y Ym Y N Y
Ym Y N Na Yb Yb Yc N Yd Ng Y Y Yb N Y Yb Y Y Ng Yh Ng Y Y Y Ng Ng Ng Yb
Ym Y N Na Yb Yb Yc N Yd Ng Y Y Yb N Y Ybm Y Y Ng Yh Ng Y Y Y Ng Ng Ng Yb
Ym N N Na Yb Yb Yc N Yd Ng Y Y Yb N Y Ybm Y Y Ng Yh Ng Y Y Y Ng Ng Ng Yb
Y Y Y Ym Y Y Y Y Y Y Y Y Y Y Ym Y
Cryptography
CPACF
Yb
Yb
Yb
247
8049ch08.fm
Function
z/OS V1 R13
z/OS V1 R12
z/OS V1 R11
z/OS V1 R10
z/VM V6 R2
z/VM V6 R1
z/VM V5 R4
CPACF AES-128, AES-192 and AES-256 CPACF SHA-1, SHA-224, SHA-256, SHA-384, SHA-512 CPACF protected key Crypto Express4S Secure IBM Enterprise PKCS #11 (EP11) coprocessor mode Crypto Express3 Elliptic Curve Cryptography (ECC)
Y Y Y Yjk Yjk Y Y
HiperSockets
Y Y Y Yjk Yjk Y Yi
Y Y Yi Yj Yj Yi Yi
Y Y Yi Yj Yj Yi Yi
Yb Yb Yb Ybm Ybm Yb Yb
32 Hipersockets HiperSockets Completion Queue HiperSockets integration with IEDN HiperSockets Virtual Switch Bridge HiperSockets Network Traffic Analyzer HiperSockets Multiple Write Facility HiperSockets support of IPV6 HiperSockets Layer 2 support HiperSockets
Y Ym Ym N Y Y Y Y
Ym N N N Y Y Y Y
Ym N N N Y Y N Y
Ym N N N Y Y N Y
Y N Ym Ym Yb Ng Y Yb Y
Ym N N N Ybm Ng Y Yb Y
Ym N N N Ybm Ng Y Yb Y
Flash Express
Yilm
z/OS Discovery and auto configuration (zDAC) FICON Express8S support of zHPF enhanced multitrack CHPID type FC FICON Express8 support of zHPF enhanced multitrack CHPID type FC High Performance FICON for System z (zHPF) FCP - increased performance for small block sizes Request node identification data FICON link incident reporting N_Port ID Virtualization for FICON (NPIV) CHPID type FCP
Y Y
Ym Y
N Ym
N Ym
N Ym
N N
N N
Ym
Ym
Ng
Ng
Ng
Y N Y Y N
Y N Y Y N
Y N Y Y N
Ym N Y Y N
Ng Y N N Y
Ng Y N N Y
Ng Y N N Y
248
8049ch08.fm
Function
z/OS V1 R13
z/OS V1 R12
z/OS V1 R11
z/OS V1 R10
z/VM V6 R2
z/VM V6 R1
z/VM V5 R4
FCP point-to-point attachments FICON SAN platform & name server registration FCP SAN management SCSI IPL for FCP Cascaded FICON Directors CHPID type FC Cascaded FICON Directors CHPID type FCP FICON Express8S support of hardware data router CHPID type FCP FICON Express8S and FICON Express8 support of T10-DIF CHPID type FCP FICON Express8S, FICON Express8, FICON Express4 10KM LX, and FICON Express4 SX support of SCSI disks CHPID type FCP FICON Express8S CHPID type FC FICON Express8 CHPID type FC FICON Express4 10KM LX and SXo CHPID type FC
N Y N N Y N N
N Y N N Y N N
N Y N N Y N N
N Y N N Y N N
Y Y N Y Y Y N
Y Y N Y Y Y N
Y Y N Y Y Y N
Ybm
Ym
Y Y Y
Y Y Y
Ym Yn Y
Ym Yn Y
Y Yn Y
Y Yn Y
Y Yn Y
VLAN management VLAN (IEE 802.1q) support QDIO data connection isolation for z/VM virtualized environments OSA Layer 3 Virtual MAC OSA Dynamic LAN idle OSA/SF enhancements for IP, MAC addressing (CHPID type OSD) QDIO diagnostic synchronization Network Traffic Analyzer Large send for IPv6 packet Broadcast for IPv4 packets Checksum offload for IPv4 packets
Y Y Y Y Y Y Y Ym Y Y
Y Y Y Y Y Y Y N Y Y
Y Y Y Y Y Y Y N Y Y
Y Y Y Y Y Y Y N Y Y
Y Y Y Yb Yb Y Yb Yb Yb Y Yp
Y Y Y Yb Yb Y Yb Yb Yb Y Yp
Y Y Ym Yb Yb Y Yb Yb Yb Y Yp
249
8049ch08.fm
Function
z/OS V1 R13
z/OS V1 R12
z/OS V1 R11
z/OS V1 R10
z/VM V6 R2
z/VM V6 R1
z/VM V5 R4
OSA-Express4S and OSA-Express3 inbound workload queueing for Enterprise Extender OSA-Express4S 10 Gigabit Ethernet LR and SR CHPID type OSD OSA-Express4S 10 Gigabit Ethernet LR and SR CHPID type OSX OSA-Express4S Gigabit Ethernet LX and SX CHPID type OSD (using two ports per CHPID) OSA-Express4S Gigabit Ethernet LX and SX CHPID type OSD (using one port per CHPID) OSA-Express4S 1000BASE-T CHPID type OSC (using one or two ports per CHPID) OSA-Express4S 1000BASE-T CHPID type OSD (using two ports per CHPID) OSA-Express4S 1000BASE-T CHPID type OSD (using one port per CHPID) OSA-Express4S 1000BASE-T CHPID type OSE (using one or two ports per CHPID) OSA-Express4S 1000BASE-T CHPID type OSMr (using one port per CHPID) OSA-Express4S 1000BASE-T CHPID type OSNq OSA-Express3 10 Gigabit Ethernet LR and SR CHPID type OSD OSA-Express3 10 Gigabit Ethernet LR and SR CHPID type OSX OSA-Express3 Gigabit Ethernet LX and SX CHPID types OSD, OSN q (using two ports per CHPID) OSA-Express3 Gigabit Ethernet LX and SX CHPID types OSD, OSN q (using one port per CHPID) OSA-Express3 1000BASE-T CHPID type OSC (using two ports per CHPID) OSA-Express3 1000BASE-T CHPID type OSD (using two ports per CHPID) OSA-Express3 1000BASE-T CHPID types OSC and OSD (using one port per CHPID)
Y Y Y Y Y Y
N Y Ym Y Y Y
N Ym Ym Ym Ym Ym
N Ym Ym Ym Ym Ym
Yb Y Y Y Y Y
Yb Y
Yb Y
Ym Y Y Y
Ym Ym Y Y
Y Y Y
Y Y Y
Ym Ym Ym
Ym Ym Ym
Y Y Y
Y Y Y
Ym Y Y
Y Y Y Y Y
Ym Y Y Ym Y
Ym Ym Y Ym Y
Ym Ym Y Ym Y
Y Y Y Y Y
Ym Y Y Ym Y
Ym Y Y Yms Ym
Y Y Y
Y Y Y
Y Y Y
Y Y Y
Y Y Y
Y Y Y
Y Ym Y
250
8049ch08.fm
Function
z/OS V1 R13
z/OS V1 R12
z/OS V1 R11
z/OS V1 R10
z/VM V6 R2
z/VM V6 R1
z/VM V5 R4
OSA-Express3 1000BASE-T CHPID type OSE (using one or two ports per CHPID) OSA-Express3 1000BASE-T CHPID type OSM r (using one port per CHPID) OSA-Express3 1000BASE-T CHPID type OSN q
Y Y
Ym Y
Ym Y
Ym Y
Y Y
Ym Y
Yms Y
z/VM integrated systems management System-initiated CHPID reconfiguration Program-directed re-IPL Multipath IPL STP enhancements Server Time Protocol Coupling over InfiniBand CHPID type CIB InfiniBand coupling links 12x at a distance of 150 m InfiniBand coupling links 1x at an unrepeated distance of 10 km Dynamic I/O support for InfiniBand CHPIDs CFCC Level 18 CFCC Level 17
Y Y Y Y Y Y Y Ym Y
Y Y Y Y Y Y Y Ym Ym
Y Y Y Y Y Ym Ym N Ym
Y Y Y Y Y Ym Ym N Ym
Y Y N Y
s
Y Y N Ys Ys Ys Ys Yb Yb
m
Y Y N Ys Ys Ys Ys Yb Yb
m
Ys Ys Ys Yb Yb
m
a. A maximum of 32 PUs per system image is supported. Guests can be defined with up to 64 virtual PUs. z/VM V5R4 and above support up to 32 real PUs. b. Support is for guest use only. c. Available for z/OS on virtual machines without virtual zAAPs defined when the z/VM LPAR does not have zAAPs defined. d. 256 GB of central memory are supported by z/VM V5R4 and later. z/VM V5R4 and later are designed to support more than 1 TB of virtual memory in use for guests. e. Web Deliverable required for Pageable 1M Large Page Support. f. 2G Large Page Support is planned as Web Deliverable in first quarter 2013. All statements regarding IBM's plans, directions, and intent are subject to change or withdrawal without notice. g. Not available to guests. h. Support varies by operating system and by version and release. i. FMIDs are shipped in a Web deliverable. j. Crypto Express4S Toleration requires Web deliverable and PTFs. k. Crypto Express4S Exploitation requires Web deliverable. l. Dynamic Reconfiguration Support for Flash Express is planned as Web Deliverable in first quarter 2013. All statements regarding IBM's plans, directions, and intent are subject to change or withdrawal without notice. m. Service is required. n. Support varies with operating system and level. For details see 8.3.37, FCP provides increased performance on page 274.
251
8049ch08.fm
o. FICON Express4 features are withdrawn from marketing. p. Supported for dedicated devices only. q. CHPID type OSN does not use ports. All communication is LPAR to LPAR. r. One port is configured for OSM. The other port in the pair is unavailable. s. Support is for dynamic I/O configuration only.
Table 8-4 EC12 functions minimum support requirements summary, part 2 Function z/VSE V5R1a z/VSE V4R3b z/TPF V1R1 Linux on System z
zEC12 Support of Unified Resource Manager Greater than 64 PUs single system image Greater than 54 PUs single system image Support of IBM zAware zIIP zAAP zAAP on zIIP Java Exploitation of Transactional Execution Large memory (> 128 GB) Large page support Out-of-order execution Guest support for Execute-extensions facility Hardware decimal floating point c Zero address detection 60 logical partitions CPU measurement facility LPAR group capacity limit Separate LPAR management of PUs Dynamic add/delete logical partition name Capacity provisioning Enhanced flexibility for CoD HiperDispatch 63.75 K subchannels Four logical channel subsystems Dynamic I/O support for multiple LCSS Third subchannel set Multiple subchannel sets IPL from alternate subchannel set
Yg N N N N N Y Y N N Y N Y N N N Y N N N N
Yg N N N N N Y Y N N Y N Y N N N Y N N N N
Y N Y Y N Y N Y N N Y N Y N N N N N N N N N N
Y N N Y N Y Y Y Y N Y N Y Y N Y Y Y N Y N
252
8049ch08.fm
Function
z/VSE V5R1a
z/VSE V4R3b
z/TPF V1R1
Linux on System z
MIDAW facility
Cryptography
CPACF CPACF AES-128, AES-192 and AES-256 CPACF SHA-1, SHA-224, SHA-256, SHA-384, SHA-512 CPACF protected key Crypto Express4S Toleration Secure IBM Enterprise PKCS #11 (EP11) coprocessor mode Crypto Express3 Elliptic Curve Cryptography (ECC)
HiperSockets
Y Y Y N Yf N Y N
Y Y Y N N N Y N
Y Yd Ye N Ygh N Ygh N
Y Y Y N Yj N Y Nj
32 Hipersockets HiperSockets Completion Queue HiperSockets integration with IEDN HiperSockets Virtual Switch Bridge HiperSockets Network Traffic Analyzer HiperSockets Multiple Write Facility HiperSockets support of IPV6 HiperSockets Layer 2 support HiperSockets
Flash Express Storage
Y Yg N N N Y N Y
Y N N N N Y N Y
Y N N N N N N N
Y Y N Yi Yj N Y Y Y
Flash Express
Nj
z/OS Discovery and auto configuration (zDAC) FICON Express8S support of zHPF enhanced multitrack CHPID type FC FICON Express8 support of zHPF enhanced multitrack CHPID type FC High Performance FICON for System z (zHPF) FCP - increased performance for small block sizes Request node identification data FICON link incident reporting N_Port ID Virtualization for FICON (NPIV) CHPID type FCP
N N N N Y N Y
N N N N Y N Y
N N N N N N N
N Y N Yk Y N Y
253
8049ch08.fm
Function
z/VSE V5R1a
z/VSE V4R3b
z/TPF V1R1
Linux on System z
FCP point-to-point attachments FICON SAN platform and name registration FCP SAN management SCSI IPL for FCP Cascaded FICON Directors CHPID type FC Cascaded FICON Directors CHPID type FCP FICON Express8S support of hardware data router CHPID type FCP FICON Express8S and FICON Express8 support of T10-DIF CHPID type FCP FICON Express8S, FICON Express8, FICON Express4 10KM LX, and FICON Express4 SX support of SCSI disks CHPID type FCP FICON Express8S c CHPID type FC FICON Express8 c CHPID type FC FICON Express4 10KM LX and SX c CHPID type FC
m
Y Y N Y Y Y N N Y
Y Y N Y Y Y N N Y
N Y N N Y N N N N
Y Y Y Y Y Y Nj Yk Y
Y Y Y
Y Yl Y
Y Yl Y
Y Yl Y
VLAN management VLAN (IEE 802.1q) support QDIO data connection isolation for z/VM virtualized environments OSA Layer 3 Virtual MAC OSA Dynamic LAN idle OSA/SF enhancements for IP, MAC addressing (CHPID=OSD) QDIO Diagnostic Synchronization Network Traffic Analyzer Large send for IPv6 packets Broadcast for IPv4 packets Checksum offload for IPv4 packets OSA-Express4S and OSA-Express3 inbound workload queueing for Enterprise Extender OSA-Express4S 10 Gigabit Ethernet LR and SR CHPID type OSD
N Y N N N N N N N N Y
N N N N N N N N N N Y
N N N N N N N N N N Y
N Y N N N N N Y Y N Y
254
8049ch08.fm
Function
z/VSE V5R1a
z/VSE V4R3b
z/TPF V1R1
Linux on System z
OSA-Express4S 10 Gigabit Ethernet LR and SR CHPID type OSX OSA-Express4S Gigabit Ethernet LX and SX CHPID type OSD (using two ports per CHPID) OSA-Express4S Gigabit Ethernet LX and SX CHPID type OSD (using one port per CHPID) OSA-Express4S 1000BASE-T CHPID type OSC (using one or two ports per CHPID) OSA-Express4S 1000BASE-T CHPID type OSD (using two ports per CHPID) OSA-Express4S 1000BASE-T CHPID type OSD (using one port per CHPID) OSA-Express4S 1000BASE-T CHPID type OSE (using one or two ports per CHPID) OSA-Express4S 1000BASE-T CHPID type OSMp OSA-Express4S 1000BASE-T CHPID type OSNo (using one or two ports per CHPID) OSA-Express3 10 Gigabit Ethernet LR and SR CHPID type OSD OSA-Express3 10 Gigabit Ethernet LR and SR CHPID type OSX OSA-Express3 Gigabit Ethernet LX and SX CHPID types OSD, OSN o (using two ports per CHPID) OSA-Express3 Gigabit Ethernet LX and SX CHPID types OSD, OSN o (using one port per CHPID) OSA-Express3 1000BASE-T CHPID type OSC (using four ports) OSA-Express3 1000BASE-T (using two ports per CHPID) CHPID type OSD OSA-Express3 1000BASE-T (using one port per CHPID) CHPID type OSD OSA-Express3 1000BASE-T (using one or two ports per CHPID) CHPID type OSE OSA-Express3 1000BASE-T Ethernet CHPID type OSN o OSA-Express3 1000BASE-T CHPID type OSM p(using two ports)
Parallel Sysplex and other
Y Y Y Y Y Y Y N Y Y Y Y Y Y Y Y Y Y N
N Y Y Y Y Y Y N Y Y N Y Y Y Y Y Y Y N
Yn Yn Y N Yg Y N N Yg Y N Yn Y Y Yn Y N Y N
Y Y Y Y Y N Y Y Y Yj Y Y Y Y N Y N
255
8049ch08.fm
Function
z/VSE V5R1a
z/VSE V4R3b
z/TPF V1R1
Linux on System z
z/VM integrated systems management System-initiated CHPID reconfiguration Program-directed re-IPL q Multipath IPL STP enhancements Server Time Protocol Coupling over InfiniBand CHPID type CIB InfiniBand coupling links 12x at a distance of 150 m InfiniBand coupling links 1x at unrepeated distance of 10 km Dynamic I/O support for InfiniBand CHPIDs CFCC Level 18 CFCC Level 17
Y -
Y -
Y Y Y
Y Y -
a. z/VSE V5R1 is designed to exploit z/Architecture, specifically 64-bit real and virtual-memory addressing. z/VSE V5R1 requires an architectural level set available with IBM System z9 or later. b. z/VSE V4 is designed to exploit z/Architecture, specifically 64-bit real-memory addressing, but does not support 64-bit virtual-memory addressing. c. Support varies with operating system and level. d. z/TPF supports only AES-128 and AES-256. e. z/TPF supports only SHA-1 and SHA-256 f. Crypto Express4S Exploitation requires PTFs g. Service is required. h. Supported only running in accelerator mode i. Applicable to Guest Operating Systems. j. IBM is working with its Linux distribution partners to include support in future Linux on System z distribution releases. k. Supported by Novell SUSE SLES 11. l. For details see 8.3.37, FCP provides increased performance on page 274. m. FICON Express4 features are withdrawn from marketing. n. Requires PUT4 with PTFs. o. CHPID type OSN does not use ports. All communication is LPAR to LPAR. p. One port is configured for OSM. The other port is unavailable. q. This is for FCP-SCSI disks.
256
8049ch08.fm
z/OS V1R13 z/OS V1R12 z/OS V1R11 z/OS V1R10 z/VM V6R2 z/VM V6R1 z/VM V5R4 z/VSE V4R3 and above z/TPF V1R1 CFCC Level 18 zAware Linux on System zd
99a 99a 99a 64ab 32c 32 32 z/VSE Turbo Dispatcher can exploit up to 4 CPs and tolerates up to 10-way LPARs 86 CPs 16 CPs or ICFs; CPs and ICFs can not be mixed 80 SUSE SLES 10: 64 CPs or IFLs SUSE SLES 11: 64 CPs or IFLs Red Hat RHEL 5: 80 CPs or IFLs Red Hat RHEL 6: 80 CPs or IFLs
a. The number of purchased zAAPs and the number of purchased zIIPs each cannot exceed the number of purchased CPs. A logical partition can be defined with any number of the available zAAPs and zIIPs. The total refers to the sum of these PU characterizations. b. Service is required. c. When running on a VM-mode LPAR z/VM can manage CPs, IFLs, zAAPs and zIIPS. Otherwise only CPs or IFLs (but not both simultaneously) are supported. d. Values are for z196 support. IBM is working with its Linux distribution partners to provide exploitation of this function in future Linux on System z distribution releases.
257
8049ch08.fm
258
8049ch08.fm
259
8049ch08.fm
Support is available on z/OS V1R11 and above. This capability is enabled by default (ZAAPZIIP=YES). To disable it, specify NO for the ZAAPZIIP parameter in the IEASYSxx PARMLIB member. On z/OS V1R10 support is provided by PTF for APAR OA27495 and the default setting in the IEASYSxx PARMLIB member is ZAAPZIIP=NO. Enabling or disabling this capability is disruptive. After changing the parameter, z/OS must be re-IPLed for the new setting to take effect.
z/OS V1R10 and above support 4 TB and up to 3 TB per servera z/VM V5R4 and above support 256 GB z/VSE V4R3 and above support 32 GB z/TPF supports 4 TBa Level 18 supports up to 3 TB per servera Supports up to 3 TB per servera SUSE SLES 11 supports 4 TBa SUSE SLES 10 supports 4 TBa Red Hat RHEL 5 supports 3 TBa Red Hat RHEL 6 supports 3 TBa
The z/VM system administrator can use the SET CPUAFFINITY command to influence the dispatching of virtual specialty engines on CPs or real specialty engines.
260
8049ch08.fm
z/OS
z/OS V1R13a
Dynamic reconfiguration support for Storage Class Memory (SCM) is planned as Web Deliverable in first quarter 2013. All statements regarding IBM's plans, directions, and intent are subject to change or withdrawal without notice.
261
8049ch08.fm
z/OS V1R10 z/OS V1R13a for Pageable 1MB large page Not supported; not available to guests z/VSE V4R3; supported for data spaces SUSE SLES 10 SP2 Red Hat RHEL 5.2
Table 8-9 lists the support requirements for 2GB large pages.
Table 8-9 Minimum support requirements for 2GB large page Operating system Support requirements
z/OS
z/OS V1R13a
a. IBM plans to support 2GB Large Page in first quarter 2013. All statements regarding IBM's plans, directions, and intent are subject to change or withdrawal without notice.
z/VM
262
8049ch08.fm
Table 8-11 Minimum support requirements for decimal floating point Operating system Support requirements
z/OS V1R10: Support includes XL, C/C++, HLASM, Language Environment, DBX, and CDA RTLE. z/VM V5R4: Support is for guest use. SUSE SLES 11 SP1 Red Hat RHEL 6
z/OS V1R10 z/VM V5R4 z/VSE V4R3 z/TPF V1R1 SUSE SLES 10 Red Hat RHEL 5
Note: The IBM zAware virtual appliance runs in a dedicated LPAR so, when activated, it will reduce by one the maximum number of logical partitions available.
z/OS V1R10 z/VM V5R4 z/VSE V4R3 z/TPF V1R1 SUSE SLES 10 Red Hat RHEL 5
263
8049ch08.fm
z/OS z/VM
264
8049ch08.fm
8.3.16 HiperDispatch
HiperDispatch, which is exclusive to zEC12, z196 and System z10, represents a cooperative effort between the z/OS operating system and the zEC12 hardware. It improves efficiencies in both the hardware and the software in the following ways: Work can be dispatched across fewer logical processors, therefore reducing the multiprocessor (MP) effects and lowering the interference across multiple partitions. Specific z/OS tasks can be dispatched to a small subset of logical processors that Processor Resource/Systems Manager (PR/SM) then ties to the same physical processors, thus improving the hardware cache reuse and locality of reference characteristics, and reducing the rate of cross-book communications. For more information, see 3.6, Logical partitioning on page 107. Table 8-15 lists HiperDispatch support requirements.
Table 8-15 Minimum support requirements for HiperDispatch Operating system Support requirements
z/OS z/VM
z/OS V1R10 and later with PTFs Not supported; not available to guests
265
8049ch08.fm
Table 8-17 lists the minimum operating systems level required on the zEC12.
Table 8-17 Minimum software requirement for MSS Operating system Support requirements
z/OS
z/OS z/VM
8049ch08.fm
and asynchronously if necessary, thus combining ultra-low latency with more tolerance for traffic peaks. This could be especially helpful in burst situations. HiperSockets Completion Queue is planned to be supported in the z/VM environments. Table 8-20 lists the minimum support requirements for HiperSockets Completion Queue.
Table 8-20 Minimum support requirements for HiperSockets Completion Queue Operating system Support requirements
z/OS V1R13a z/VSE V5R1a SLES 11 SP2 Red Hat RHEL 6.2
z/VM V6R2a SLES 10 SP4 update (kernel 2.6.16.60-0.95.1) Red Hat RHEL 5.8 (GA-level)
267
8049ch08.fm
z/OS
z/OS V1R10
z/OS V1R10 z/VM V5R4 SUSE SLES 10 SP2 Red Hat RHEL 5.2
268
8049ch08.fm
Table 8-25 Minimum support requirements for HiperSockets Layer 2 Operating system Support requirements
z/VM V5R4 for guest exploitation SUSE SLES 10 SP2 Red Hat RHEL 5.2
Native FICON and Channel-to-Channel (CTC) CHPID type FC zHPF single track operations CHPID type FC zHPF multitrack operations CHPID type FC
V1R10a
V5R4
V4R3
V1R1
SUSE SLES 10 Red Hat RHEL 5 SUSE SLES 11 SP1 Red Hat RHEL 6 SUSE SLES 11 SP2 Red Hat RHEL 6.1 SUSE SLES 10 Red Hat RHEL 5
V1R10b
V6R2b
N/A
N/A
V1R10b
V6R2b
N/A
N/A
N/A
V5R4b
V4R3
N/A
All FICON Express4, FICON Express2 and FICON features are withdrawn from marketing.
269
8049ch08.fm
Operating system
z/OS
z/VM
z/VSE
z/TPF
Linux on System z
Support of hardware data router CHPID type FCP Support of T10-DIF CHPID type FCP
N/A
N/A
N/A
N/A
N/Ac
N/A
V5R4b
N/A
N/A
a. PTFs required to support GRS Ficon CTC toleration. b. PTFs required c. IBM is working with its Linux distribution partners to provide exploitation of this function in future Linux on System z distribution releases.
Native FICON and Channel-to-Channel (CTC) CHPID type FC zHPF single track operations CHPID type FC zHPF multitrack operations CHPID type FC Support of SCSI devices CHPID type FCP Support of T10-DIF CHPID type FCP
V1R10
V5R4
V4R3
V1R1
V1R10a
N/A
N/A
N/A
a. PTFs required b. IBM is working with its Linux distribution partners to provide exploitation of this function in future Linux on System z distribution releases.
270
8049ch08.fm
including parallel access volume (PAV) definitions, control unit numbers, and device number ranges. Then, when new controllers are added to an I/O configuration or changes are made to existing controllers, the system is designed to discover them and propose configuration changes based on that policy. zDAC provides real-time discovery for the FICON fabric, subsystem and I/O device resource changes from z/OS. By exploring the discovered control units for defined logical control units (LCU) and devices, zDAC compares the discovered controller information with the current system configuration to determine delta changes to the configuration for a proposed configuration. All new added or changed logical control units and devices will be added into the proposed configuration, with proposed control unit and device numbers, and channel paths based on the defined policy. zDAC uses channel path chosen algorithms to minimize single points of failure. The zDAC proposed configurations are created as work I/O definition files (IODF), that can be converted to production IODF and activated. zDAC is designed to perform discovery for all systems in a sysplex that support the function. Thus, zDAC helps simplifying I/O configuration on zEC12 systems running z/OS and reduces complexity and setup time. zDAC applies to all FICON features supported on zEC12 when configured as CHPID type FC. Table 8-28 lists the minimum support requirements for zDAC.
Table 8-28 Minimum support requirements for zDAC Operating system Support requirements
z/OS V1R12a
271
8049ch08.fm
allows the channel to fully exploit the bandwidth of FICON channels, resulting in higher throughputs and lower response times. On the zEC12 and z196 the multitrack operations extension applies exclusively to the FICON Express8S, FICON Express8, and FICON Express47, when configured as CHPID type FC, and connecting to z/OS. zHPF requires matching support by the DS8000 series, otherwise the extended multitrack support is transparent to the control unit. From the z/OS point of view, the existing FICON architecture is called command mode and zHPF architecture is called transport mode. During link initialization, the channel node and the control unit node indicate whether they support zHPF. Attention: All FICON channel paths (CHPIDs) defined to the same Logical Control Unit (LCU) must support zHPF. The inclusion of any non-compliant zHPF features in the path group will cause the entire path group to support command mode only. The mode used for an I/O operation depends on the control unit supporting zHPF and settings in the z/OS operating system. For z/OS exploitation there is a parameter in the IECIOSxx member of SYS1.PARMLIB (ZHPF=YES or NO) and in the SETIOS system command to control whether zHPF is enabled or disabled. The default is ZHPF=NO. Support is also added for the D IOS,ZHPF system command to indicate whether zHPF is enabled, disabled, or not supported on the server. Similar to the existing FICON channel architecture, the application or access method provides the channel program (channel command words, CCWs). The way that zHPF (transport mode) manages channel program operations is significantly different from the CCW operation for the existing FICON architecture (command mode). While in command mode, each single CCW is sent to the control unit for execution. In transport mode, multiple channel commands are packaged together and sent over the link to the control unit in a single control block. Less overhead is generated compared to the existing FICON architecture. Certain complex CCW chains are not supported by zHPF. The zHPF is exclusive to zEC12, z196, z114 and System z10. The FICON Express8S, FICON Express8, and FICON Express47,8 (CHPID type FC) concurrently support both the existing FICON protocol and the zHPF protocol in the server Licensed Internal Code. Table 8-29 lists the minimum support requirements for zHPF
Table 8-29 Minimum support requirements for zHPF Operating system Support requirements
z/OS
Single track operations: z/OS V1R10 with PTFs Multitrack operations: z/OS V1R10 with PTFs 64K enhancement: z/OS V1R10 with PTFs Not supported; not available to guests SLES 11 SP1 supports zHPF. IBM continues to work with its Linux distribution partners on exploitation of appropriate zEC12 functions be provided in future Linux on System z distribution releases.
For more information about FICON channel performance, see the performance technical papers on the System z I/O connectivity web site:
7 8
FICON Express4 LX 4KM is not support on zEC12 All FICON Express4, FICON Express2 and FICON features are withdrawn from marketing.
272
8049ch08.fm
http://www-03.ibm.com/systems/z/hardware/connectivity/ficon_performance.html
z/OS
z/OS V1R10
273
8049ch08.fm
z/OS
z/OS V1R10
z/VM
z/VM V5R4 provides support for guest operating systems and VM users to obtain virtual port numbers. Installation from DVD to SCSI disks is supported when NPIV is enabled. z/VSE V4R3 SUSE SLES 10 SP3 Red Hat RHEL 5.4
8049ch08.fm
Non-QDIO mode, with CHPID type OSE Local 3270 emulation mode, including OSA-ICC, with CHPID type OSC Ensemble management, with CHPID type OSM. Operating system support is required in order to recognize and use the second port on the OSA-Express4S 1000BASE-T feature. Table 8-33 lists the minimum support requirements for OSA-Express4S 1000BASE-T.
Table 8-33 Minimum support requirements for OSA-Express4S 1000BASE-T Ethernett Operating system Support requirements exploiting two ports per CHPID Support requirements exploiting one port per CHPID
z/OS z/VM
OSC, OSD, OSE, OSNb : z/OS V1R10a OSC: z/VM V5R4 OSD: z/VM V5R4a OSE: z/VM V5R4 OSNb: z/VM V5R4 OSC, OSD, OSE, OSNb : z/VSE V4R3 OSD: z/TPF V1R1 PUT 4a OSNb : z/TPF V1R1 PUT 4a OSD OSD: SUSE SLES 10 SP2 Red Hat RHEL 5.2 OSNb : SUSE SLES 10 Red Hat RHEL 5
OSC, OSD, OSE, OSM, OSNb : z/OS V1R10a OSC: z/VM V5R4 OSD: z/VM V5R4 OSE: z/VM V5R4 OSM: z/VM V5R4a OSNb : z/VM V5R4 OSC, OSD, OSE, OSNb : z/VSE V4R3 OSD: z/TPF V1R1 OSNb : z/TPF V1R1 PUT4a OSD OSD: SUSE SLES 10 Red Hat RHEL 5 OSM: SUSE SLES 10 SP4 Red Hat RHEL 5.6 OSNb : SUSE SLES 10 Red Hat RHEL 5
a. PTFs required b. Although CHPID type OSN does not use any ports (because all communication is LPAR to LPAR), it is listed here for completeness.
275
8049ch08.fm
Table 8-34 Minimum support requirements for OSA-Express4S 10 Gigabit Ethernet LR and SR Operating system Support requirements
z/OS z/VM z/VSE z/TPF IBM zAware Linux on System z a. PTFs required
OSD: z/OS V1R10a OSX: z/OS V1R10a OSD: z/VM V5R4 OSX: z/VM V5R4a for dynamic I/O only OSD: z/VSE V4R3 OSX: z/VSE V5R1 OSD: z/TPF V1R1 OSX: z/TPF V1R1 PUT4a OSD OSX OSD: SUSE SLES 10, Red Hat RHEL 5 OSX: SUSE SLES 10 SP4, Red Hat RHEL 5.6
z/OS z/VM z/VSE z/TPF IBM zAware Linux on System z a. PTFs required
OSD: z/OS V1R10a OSD: z/VM V5R4a OSD: z/VSE V4R3 OSD: z/TPF V1R1 PUT 4a OSD OSD: SUSE SLES 10 SP2 Red Hat RHEL 5.2
OSD: z/OS V1R10a OSD: z/VM V5R4 OSD: z/VSE V4R3 OSD: z/TPF V1R1
8049ch08.fm
Table 8-36 lists the minimum support requirements for OSA-Express3 10 Gigabit Ethernet LR and SR features.
Table 8-36 Minimum support requirements for OSA-Express3 10 Gigabit Ethernet LR and SR Operating system Support requirements
OSD: z/OS V1R10 OSX: z/OS V1R12, z/OS V1R10 and z/OS V1R11, with service OSD: z/VM V5R4 OSX: z/VM V5R4 for dynamic I/O only; z/VM V6R1 with service OSD: z/VSE V4R3 OSD: z/TPF V1R1 OSD OSD: SUSE SLES 10 OSD: Red Hat RHEL 5
z/OS V1R10; service required z/VM V5R4; service required z/VSE V4R3 z/TPF V1R1; service required
Table 8-38 Minimum support requirements for OSA-Express3 Gigabit Ethernet LX and SX, two ports Operating system Support requirements
277
8049ch08.fm
Operating system
Support requirements
z/OS
OSD: z/OS V1R10; service required OSE: z/OS V1R10 OSM: z/OS V1R12, z/OS V1R10 and z/OS V1R11, with service OSNb : z/OS V1R10 OSD: z/VM V5R4; service required OSE: z/VM V5R4 OSM: z/VM service required; V5R4 for dynamic I/O only OSNb : z/VM V5R4 OSD: z/VSE V4R3; service required OSE: z/VSE V4R3 OSNb : z/VSE V4R3 OSD and OSNb : z/TPF V1R1; service required OSD OSD: SUSE SLES 10 SP2 Red Hat RHEL 5.2 OSNb : SUSE SLES 10 SP2 Red Hat RHEL 5.2
z/VM
z/VSE
a. Applies to CHPID types OSC, OSD, OSE, OSM and OSN. For support, see Table 8-40 on page 279. b. Although CHPID type OSN does not use any ports (because all communication is LPAR to LPAR), it is listed here for completeness.
278
8049ch08.fm
Table 8-40 lists the minimum support requirements for OSA-Express3 1000BASE-T Ethernet (two ports).
Table 8-40 Minimum support requirements for OSA-Express3 1000BASE-T Ethernet, two ports Operating system Support requirements
OSD, OSE, OSM and OSN: V1R10 OSD, OSE, OSM and OSN: V5R4 V4R3 OSD, OSN, and OSC: V1R1 OSD OSD: SUSE SLES 10 Red Hat RHEL 5 OSN: SUSE SLES 10 SP3 Red Hat RHEL 5.4
279
8049ch08.fm
280
8049ch08.fm
Channel Data Link Control (CDLC), when used with the Communication Controller for Linux, emulates selected functions of IBM 3745/NCP operations. The port used with the OSN support appears as an ESCON channel to the operating system. This support can be used with OSA-Express4S 1000BASE-T features. Table 8-41 lists the minimum support requirements for OSN.
Table 8-41 Minimum support requirements for OSA-Express4S OSN Operating system Support requirements
z/OS V1R10a z/VM V5R4 z/VSE V4R3 z/TPF V1R1 PUT 4a SUSE SLES 10 Red Hat RHEL 5
z/OS z/VM
z/OS V1R10 z/VM V5R4. Support of guests is transparent to z/VM if the device is directly connected to the guest (pass through).
281
8049ch08.fm
to register or unregister its VLAN IDs with a GVRP-capable switch and dynamically update its table as the VLANs change. This simplifies the network administration and management of VLANs as manually entering VLAN IDs at the switch is no longer necessary. Minimum support requirements are listed in Table 8-43.
Table 8-43 Minimum support requirements for GVRP Operating system Support requirements
z/OS z/VM
z/OS z/VM
z/OS V1R12 z/VM V5R4 for guest exploitation only; service required
282
8049ch08.fm
z/OS z/VM
z/OS V1R13 z/VM V5R4 for guest exploitation only; service required
283
8049ch08.fm
VSWITCH isolation support is provided by APAR VM64281. z/VM 5R4 and later support is provided by CP APAR VM64463 and TCP/IP APAR PK67610. QDIO data connection isolation is supported by all OSA-Express4S and OSA-Express3 features on zEC12.
284
8049ch08.fm
Table 8-47 Minimum support requirements for OSA-Express4S checksum offload Traffic Support requirements
LPAR to LPAR IPv6 LPAR to LPAR traffic for IPv4 and IPv6
z/OS V1R12 a z/VM V5R4 for guest exploitationb z/OS V1R13 z/VM V5R4 for guest exploitationb z/OS V1R13 z/VM V5R4 for guest exploitationb
a. PTFs are required b. Device is directly attached to guest, PTFs are required.
285
8049ch08.fm
8049ch08.fm
z/VM V5R4 SUSE SLES 10 SP3 Red Hat RHEL 5.4 V4R3 on SCSI disks
z/OS V1R10 z/VM V5R4 (dynamic I/O support for InfiniBand CHPIDs only; coupling over InfiniBand is not supported for guest use) z/TPF V1R1
z/OS z/VM
z/OS V1R10; service required z/VM V5R4 (dynamic I/O support for InfiniBand CHPIDs only; coupling over InfiniBand is not supported for guest use)
287
8049ch08.fm
when entering and leaving configuration mode is also supported. z/VM does not use InfiniBand and does not support the use of InfiniBand coupling links by guests. Table 8-51 lists the minimum support requirements of dynamic I/O support for InfiniBand CHPIDs.
Table 8-51 Minimum support requirements for dynamic I/O support for InfiniBand CHPIDs Operating system Support requirements
z/VM
z/VM V5R4
z/OS V1R10 and later with the Cryptographic Support for z/OS V1R10-V1R12 Web deliverable. z/VM V5R4 with PTFs and higher: Supported for guest use. z/VSE V4R2 and later: Supports the CPACF features with the functionality supported on IBM System z10. z/TPF V1R1 SUSE SLES 11 SP1 Red Hat RHEL 6.1 For Message-Security-Assist-Extension 4 exploitation, IBM is working with its Linux distribution partners to include support in future Linux on System z distribution releases.
a. CPACF is also exploited by several IBM Software product offerings for z/OS, such as IBM WebSphere Application Server for z/OS.
288
8049ch08.fm
z/OS
z/OS V1R13, or z/OS V1R12 with the Cryptographic Support for z/OS V1R12-V1R13 Web deliverable. z/OS V1R10, or z/OS V1R11 with toleration maintenance. z/VM V6R2, or z/VM V6R1, or z/VM V5R4 with maintenance. z/VSE V5R1 with PTFs. Service required (accelerator mode only). IBM is working with its Linux distribution partners to include support in future Linux on System z distribution releases
z/OS
z/OS V1R12 (ICSF FMID HCR7770) and above z/OS V1R10, or z/OS V1R11 with the Cryptographic Support for z/OS V1R9-V1R11 Web deliverable. Z/OS V1R8 with toleration maintenance: Crypto Express3 features handled as Crypto Express2 features. z/VM V5R4: Service required; supported for guest use only. z/VSE V4R2 Service required (accelerator mode only). For toleration: SUSE SLES10 SP3 and SLES 11. Red Hat RHEL 5.4 and RHEL 6.0. For exploitation: SUSE SLES11 SP1. Red Hat RHEL 6.1.
289
8049ch08.fm
For Linux on System z, support is delivered through IBM and distribution partners. For more information see Linux on System z on the developerWorks website: http://www.ibm.com/developerworks/linux/linux390/
290
8049ch08.fm
Table 8-55 z/OS ICSF FMIDs z/OS ICSF FMID Web deliverable name Supported function
V1R10
HCR7750
CPACF AES-192 and AES-256 CPACF SHA-224, SHA-384 and SHA-512 4096-bit RSA keys ISO-3 PIN block format IBM System z10 BC support Secure key AES Key store policy PKDS Sysplex-wide consistency In-storage copy of the PKDS 13-digit through 19-digit PANs Crypto Query service Enhanced SAF checking Crypto Express3 and Crypto Express3-1P support PKA Key Management Extensions CPACF Protected Key Extended PKCS #11 ICSF Restructure (Performance, RAS, ICSF-CICS Attach Facility) IBM zEnterprise 196 support Elliptic Curve Cryptography Message-Security-Assist-4 HMAC Support ANSI X9.8 Pin ANSI X9.24 (CBC Key Wrapping) CKDS constraint relief PCI Audit All callable services AMODE(64) PKA RSA OAEP with SHA-256 algorithmb
HCR7751
HCR7770
HCR7780
291
8049ch08.fm
z/OS
ICSF FMID
Supported function
V1R11
HCR7751
IBM System z10 BC support Secure key AES Key store policy PKDS Sysplex-wide consistency In-storage copy of the PKDS 13-digit through 19-digit PANs Crypto Query service Enhanced SAF checking Crypto Express3 and Crypto Express3-1P support PKA Key Management Extensions CPACF Protected Key Extended PKCS #11 ICSF Restructure (Performance, RAS, ICSF-CICS Attach Facility) IBM zEnterprise 196 support Elliptic Curve Cryptography Message-Security-Assist-4 HMAC Support ANSI X9.8 Pin ANSI X9.24 (CBC Key Wrapping) CKDS constraint relief PCI Audit All callable services AMODE(64) PKA RSA OAEP with SHA-256 algorithmb Expanded key support for AES algorithm Enhanced ANSI TR-31 PIN block decimalization table protection Elliptic Curve Diffie-Hellman (ECDH) algorithm RSA in the Modulus Exponent (ME) and Chinese Remainder Theorem (CRT) formats.
HCR7770
HCR7780
HCR7790
292
8049ch08.fm
z/OS
ICSF FMID
Supported function
V1R12
HCR7770
Crypto Express3 and Crypto Express3-1P support PKA Key Management Extensions CPACF Protected Key Extended PKCS #11 ICSF Restructure (Performance, RAS, ICSF-CICS Attach Facility) IBM zEnterprise 196 support Elliptic Curve Cryptography Message-Security-Assist-4 HMAC Support ANSI X9.8 Pin ANSI X9.24 (CBC Key Wrapping) CKDS constraint relief PCI Audit All callable services AMODE(64) PKA RSA OAEP with SHA-256 algorithmb Expanded key support for AES algorithm Enhanced ANSI TR-31 PIN block decimalization table protection Elliptic Curve Diffie-Hellman (ECDH) algorithm RSA in the ME and CRT formats. Support for the Crypto Express4S feature when configured as an EP11 coprocessor Support for the Crypto Express4S feature when configured as a CCA coprocessor Support for 24-byte Data Encryption Standard (DES) master keys Improved wrapping key strength DUKPT for Message Authentication Code (MAC) and encryption keys Secure Cipher Text Translate2 Compliance with new random number generation standards EMV enhancements for applications supporting American Express cards
HCR7780
HCR7790
HCR77A0
293
8049ch08.fm
z/OS
ICSF FMID
Supported function
V1R13
HCR7780
IBM zEnterprise 196 support Elliptic Curve Cryptography Message-Security-Assist-4 HMAC Support ANSI X9.8 Pin ANSI X9.24 (CBC Key Wrapping) CKDS constraint relief PCI Audit All callable services AMODE(64) PKA RSA OAEP with SHA-256 algorithmb Expanded key support for AES algorithm Enhanced ANSI TR-31 PIN block decimalization table protection Elliptic Curve Diffie-Hellman (ECDH) algorithm RSA in the ME and CRT formats. Support for the Crypto Express4S feature when configured as an EP11 coprocessor Support for the Crypto Express4S feature when configured as a CCA coprocessor Support for 24-byte Data Encryption Standard (DES) master keys Improved wrapping key strength DUKPT for Message Authentication Code (MAC) and encryption keys Secure Cipher Text Translate2 Compliance with new random number generation standards EMV enhancements for applications supporting American Express cards
HCR7790
HCR77A0
294
8049ch08.fm
Permit the use of a PKDS with RSA private key tokens encrypted under the ECC Master Key. Support for installation options data sets which use the keyword BEGIN(FMID). New SMP/E Fix Category will be created for ICSF coexistence IBM.Coexistence.ICSF.z/OS_V1R12-V1R13-HCR77A0
8.5.2 HCD
On z/OS V1R10 and above, HCD or the Hardware Configuration Manager (HCM) assist in the defining a configuration for zEC12.
295
8049ch08.fm
The parameter indicates the amount of storage, in percentage, megabytes, or gigabytes. The value cannot be changed dynamically.
8.5.5 HiperDispatch
The HIPERDISPATCH=YES/NO parameter in the IEAOPTxx member of SYS1.PARMLIB and on the SET OPT=xx command can control whether HiperDispatch is enabled or disabled for a z/OS image. It can be changed dynamically, without an IPL or any outage. The default is that HiperDispatch is disabled on all releases, from z/OS V1R10 (requires PTFs for zIIP support) through z/OS V1R13. To effectively exploit HiperDispatch, the Workload Manager (WLM) goal adjustment might be required. Review the WLM policies and goals, and update them as necessary. You might want to run with the new policies and HiperDispatch on for a period, turn it off and use the older WLM policies while analyzing the results of using HiperDispatch, re-adjust the new policies and repeat the cycle, as needed. In order to change WLM policies, turning HiperDispatch off then on is not necessary. A health check is provided to verify whether HiperDispatch is enabled on a system image that is running on zEC12.
8049ch08.fm
ARCHITECTURE: this option selects the minimum level of machine architecture on which the program will run. Note that certain features provided by the compiler require a minimum architecture level. ARCH(10) exploit instructions available on the zEC12. TUNE: this option allows optimization of the application for a specific machine architecture, within the constraints imposed by the ARCHITECTURE option. The TUNE level must not be lower than the setting in the ARCHITECTURE option. For more information about the ARCHITECTURE and TUNE compiler options, see the z/OS V1R13.0 XL C/C++ Users Guide, SC09-4767. Attention: The previous System z ARCHITECTURE or TUNE options for C/C++ programs should be used if the same applications should run on both the zEC12 as well as on previous System z servers. However, if C/C++ applications will only run on zEC12 servers the latest ARCHITECTURE and TUNE options should be used assuring that the best performance possible is delivered through the latest instruction set additions.
297
8049ch08.fm
The initial support of the CFCC on the zEC12 is level 18. CFCC level 18 offers the following enhancements: Coupling channel reporting enhancements Enables RMF to differentiate various IFB link types and detect if CIB link running degraded Serviceability enhancements Additional structure control info in CF dumps Enhanced CFCC tracing support Enhanced Triggers for CF non-disruptive dumping. Performance enhancements Dynamic structure size alter improvement DB2 GBP cache bypass Cache structure management Attention: Having more than 1024 structures requires a new version of the CFRM CDS. In addition, all systems in the sysplex need to be at z/OS V1R12 (or above) or have the coexistence/preconditioning PTFs installed. Falling back to a previous level, without the coexistence PTF installed, is not supported without at sysplex IPL. zEC12 systems with CFCC level 18 require z/OS V1R10 or later, and z/VM V5R4 or later for guest virtual coupling. The current CFCC level for zEC12 servers is CFCC level 18, see Table 8-56. To support migration from one CFCC level to the next, different levels of CFCC can be run concurrently while the coupling facility logical partitions are running on different servers (CF logical partitions running on the same server share the same CFCC level).
Table 8-56 System z CFCC code level considerations
CFCC level 18 or later CFCC level 17 or later CFCC level 15 or later CFCC level 14 or later CFCC level 13 or later
CF structure sizing changes are expected when going from CFCC Level 17 (or below) to CFCC Level 18. We suggest reviewing the CF LPAR size by using the CFSizer tool available at the following web page: http://www.ibm.com/systems/z/cfsizer Previous to migration, installation of compatibility/coexistence PTFs is highly desirable. A planned outage is required when migrating the CF or the CF LPAR to CFCC level 18. For additional details about CFCC code levels, see the Parallel Sysplex web site: http://www.ibm.com/systems/z/pso/cftable.html
298
8049ch08.fm
There are exceptions to this statement and we skip a number of details in the following description. We assume that the reader can merge this brief description with an existing understanding of I/O operations in a virtual memory environment.
299
8049ch08.fm
The number of IDAWs required for a CCW is determined by the IDAW format as specified in the operation request block (ORB), by the count field of the CCW, and by the data address in the initial IDAW. For example, three IDAWS are required when the following three events occur: 1. The ORB specifies format-2 IDAWs with 4 KB blocks. 2. The CCW count field specifies 8 KB. 3. The first IDAW designates a location in the middle of a 4 KB block. CCWs with data chaining can be used to process I/O data blocks that have a more complex internal structure, in which portions of the data block are directed into separate buffer areas (this is sometimes known as scatter-read or scatter-write). However, as technology evolves and link speed increases, data chaining techniques are becoming less efficient in modern I/O environments for reasons involving switch fabrics, control unit processing and exchanges, and others. The MIDAW facility is a method of gathering and scattering data from and into discontinuous storage locations during an I/O operation. The modified IDAW (MIDAW) format is shown in Figure 8-2 on page 300. It is 16 bytes long and is aligned on a quadword.
300
8049ch08.fm
The use of MIDAWs is indicated by the MIDAW bit in the CCW. If this bit is set, then the skip flag cannot be set in the CCW. The skip flag in the MIDAW can be used instead. The data count in the CCW should equal the sum of the data counts in the MIDAWs. The CCW operation ends when the CCW count goes to zero or the last MIDAW (with the last flag) ends. The combination of the address and count in a MIDAW cannot cross a page boundary. This means that the largest possible count is 4 K. The maximum data count of all the MIDAWs in a list cannot exceed 64 K, which is the maximum count of the associated CCW. The scatter-read or scatter-write effect of the MIDAWs makes it possible to efficiently send small control blocks embedded in a disk record to separate buffers from those used for larger data areas within the record. MIDAW operations are on a single I/O block, in the manner of data chaining. Do not confuse this operation with CCW command chaining.
301
8049ch08.fm
8.8 IOCP
All System z servers require a description of their I/O configuration. This description is stored in input/output configuration data set (IOCDS) files. The input/output configuration program (IOCP) allows creation of the IOCDS file from a source file known as the input/output configuration source (IOCS). The IOCS file contains detailed information for each channel and path assignment, each control unit, and each device in the configuration. The required level of IOCP for the zEC12 is V2 R1 L0 (IOCP 2.1.0) or later with PTF. See the Input/Output Configuration Program Users Guide, SB10-7037, for details.
302
8049ch08.fm
8.10 ICKDSF
Device Support Facilities, ICKDSF, Release 17 is required on all systems that share disk subsystems with a zEC12 processor. ICKDSF supports a modified format of the CPU information field, which contains a two-digit logical partition identifier. ICKDSF uses the CPU information field instead of CCW reserve/release for concurrent media maintenance. It prevents multiple systems from running ICKDSF on the same volume, and at the same time allows user applications to run while ICKDSF is processing. To prevent any possible data corruption, ICKDSF must be able to determine all sharing systems that can potentially run ICKDSF. Therefore, this support is required for zEC12. Important: The need for ICKDSF Release 17 applies even to systems that are not part of the same sysplex, or that are running an operating system other than z/OS, such as z/VM.
303
8049ch08.fm
Red Hat RHEL 5.5 and up, 6.0 and up SUSE SLES 10 (SP4)and up, SLES 11 (SP1)a and up Microsoft Windows Server 2008 R2b Microsoft Windows Server 2008 (SP2)b (Datacenter Edition recommended)
z/OS
z/OS V1R11 for IPv4 z/OS V1R12 for IPv4 and IPv6
304
8049ch08.fm
11
305
8049ch08.fm
Charge models
There is a variety of Workload Licence Charges (WLC) pricing structures that support two charge models: Variable charges (several pricing metrics): Variable charges apply to products such as z/OS, z/VSE, z/TPF, DB2, IMS, CICS, MQSeries, and Lotus Domino. There are several pricing metrics employing the following charge types: Full-capacity: The CPCs total number of MSUs is used for charging. Full-capacity is applicable when the clients CPC is not eligible for sub-capacity. Sub-capacity: Software charges are based on the utilization of the logical partitions where the product is running. Flat charges: Software products licensed under flat charges are not eligible for sub-capacity pricing. There is a single charge per CPC on the zEC12.
Sub-capacity
For eligible programs, sub-capacity allows software charges based on utilization of logical partitions instead of the CPCs total number of MSUs. Sub-capacity removes the dependency between software charges and CPC (hardware) installed capacity. The sub-capacity licensed products are charged monthly based on the highest observed 4-hour rolling average utilization of the logical partitions in which the product runs (with the exception of products licensed using the SALC pricing metric). This requires measuring the utilization and reporting it to IBM. The logical partitions 4-hour rolling average utilization can be limited by a defined capacity value on the partitions image profile. This activates the soft capping function of PR/SM, limiting the 4-hour rolling average partition utilization to the defined capacity value. Soft capping controls the maximum 4-hour rolling average usage (the last 4-hour average value at every 5-minute interval), but does not control the maximum instantaneous partition use. Also available is an LPAR group capacity limit, which allows you to set soft capping by PR/SM for a group of logical partitions running z/OS. Even using the soft capping option, the partitions use can reach up to its maximum share based on the number of logical processors and weights in the image profile. Only the 4-hour rolling average utilization is tracked, allowing utilization peaks above the defined capacity value. Some pricing metrics apply to stand alone System z servers, others apply to the aggregation of multiple zEC12 and System z servers workloads within the same Parallel Sysplex.
306
8049ch08.fm
For further information about WLC and details about how to combine logical partitions utilization, see the publication z/OS Planning for Workload License Charges, SA22-7506, available from the following web page: http://www-03.ibm.com/systems/z/os/zos/bkserv/find_books.html Metrics applicable to a stand-alone zEC12 are: Advanced Workload License Charges (AWLC) System z New Application License Charges (zNALC) Parallel Sysplex License Charges (PSLC) Metrics applicable to a zEC12 in an actively coupled Parallel Sysplex are: Advanced Workload License Charges (AWLC), when all CPCs are zEC12, z196 or z114 Variable Workload License Charges (VWLC) are only allowed under the AWLC Transition Charges for Sysplexes when not all CPCs are zEC12, z196 or z114. System z New Application License Charges (zNALC) Parallel Sysplex License Charges (PSLC)
307
8049ch08.fm
308
8049ch08.fm
Aggregation allows charging a product based on the total MSU value of the machines where the product executes (as opposed to all the machines in the cluster). In an uncoupled environment software charges are based on the MSU capacity of the machine. For additional information, see the PSLC web page: http://www.ibm.com/systems/z/resources/swprice/mlc/pslc.html
8.13 References
For the most current planning information, see the support web site for each of the following operating systems: z/OS: http://www.ibm.com/systems/support/z/zos/ z/VM: http://www.ibm.com/systems/support/z/zvm/ z/VSE: http://www.ibm.com/servers/eserver/zseries/zvse/support/preventive.html
309
8049ch08.fm
310
8049ch09.fm
Chapter 9.
System upgrades
This chapter provides an overview of IBM zEnterprise EC12 upgrade capabilities and procedures, with an emphasis on Capacity on Demand offerings. The upgrade offerings to the zEC12 systems have been developed from previous IBM System z systems. In response to customer demands and changes in market requirements, a number of features have been added. The provisioning environment gives the customer an unprecedented flexibility and a finer control over cost and value. For detailed tutorials on all aspects of system upgrades, see Resource Link1 - Customer Initiated Upgrade Information, then select Education. A list of available systems will help you select your particular product: https://www-304.ibm.com/servers/resourcelink/hom03010.nsf/pages/CIUInformation?Ope nDocument Given today's business environment, the growth capabilities provided by the zEC12 are plentiful, including the following benefits: Enabling exploitation of new business opportunities Supporting the growth of dynamic, smart environments Managing the risk of volatile, high-growth, and high-volume applications Supporting 24x365 application availability Enabling capacity growth during lock down periods Enabling planned-downtime changes without availability impacts This chapter discusses the following topics: Upgrade types on page 312 Concurrent upgrades on page 316 MES upgrades on page 322 Permanent upgrade through the CIU facility on page 328 On/Off Capacity on Demand on page 332 Capacity for Planned Event on page 342 Capacity Backup on page 344 Nondisruptive upgrades on page 347 Summary of Capacity on Demand offerings on page 352
1
311
8049ch09.fm
9.1.1 Overview
Upgrades can be categorized as described in the following discussion.
312
8049ch09.fm
Disruptive An upgrade is considered disruptive, when resources modified or added to an operating system image, require that the operating system be recycled to configure the newly added resources. Nondisruptive Nondisruptive upgrades do not require the running software or operating system to be restarted for the upgrade to take effect. Thus, even concurrent upgrades can be disruptive to those operating systems or programs that do not support the upgrades while at the same time being nondisruptive to others. For details, see 9.8, Nondisruptive upgrades on page 347.
Activated capacity Billable capacity Capacity Capacity Backup (CBU) Capacity for planned event CPE) Capacity levels
Capacity that is purchased and activated. Purchased capacity can be greater than the activated capacity. Capacity that helps handle workload peaks, either expected or unexpected. The one billable offering available is On/Off Capacity on Demand. Hardware resources (processor and memory) able to process workload can be added to the system through various capacity offerings. Capacity Backup allows you to replace model capacity or specialty engines to a backup system, in the event of an unforeseen loss of system capacity because of an emergency. Used when temporary replacement capacity is needed for a short term event. CPE activate processor capacity temporarily to facilitate moving machines between data centers, upgrades, and other routine management tasks. CPE is an offering of Capacity on Demand. Can be full capacity or subcapacity. For the zEC12 system, capacity levels for the CP engine are 7, 6, 5, and 4: 199, A0, A1 for capacity level 7nn 120 for capacity levels 6yy, 5yy 020 for capacity levels 4xx. An all IFL or an all ICF system has a capacity level of 400. Derived from the capacity level and the number of processors. For the zEC12 system, the capacity levels are 7nn, 6yy, 5yy, 4xx, where xx, yy or nn indicates the number of active CPs. The number of processors can have a range of: 199, A0, A1 for capacity level 7nn 120 for capacity levels 6yy, 5yy 020 for capacity levels 4xx. An all IFL or an all ICF system has a capacity level of 400. A Web-based facility where you can request processor and memory upgrades by using the IBM Resource Link and the system's remote support facility (RSF) connection. The ability of a computing system to increase or decrease its performance capacity as needed to meet fluctuations in demand. As a component of z/OS Capacity Provisioning, CPM monitors business-critical workloads that are running on z/OS systems on zEC12 systems.
Capacity setting
Customer Initiated Upgrade (CIU) Capacity on Demand (CoD) Capacity Provisioning Manager (CPM)
313
8049ch09.fm
Term
Description
Customer profile Full capacity CP feature High water mark Installed record Model capacity identifier (MCI)
This information resides on Resource Link and contains customer and machine information. A customer profile can contain information about more than one machine. For zEC12 feature (CP7), provides full capacity. Capacity settings 7nn are full capacity settings. Capacity purchased and owned by the customer. The LICCC record has been downloaded, staged to the SE, and is now installed on the CPC. A maximum of eight different records can be concurrently installed and active. Shows the current active capacity on the system, including all replacement and billable capacity. For the zEC12, the model capacity identifier is in the form of 7nn, 6yy, 5yy, or 4xx, where xx, yy or nn indicates the number of active CPs. nn can have a range of 01 - 99, A0, A1 yy can have a range of 01 - 20 xx can have a range of 00 - 20. An all IFL or an all ICF system has a capacity level of 400. Keeps information about capacity settings active before any temporary capacity was activated.
Model Permanent Capacity Identifier (MPCI) Model Temporary Capacity Identifier (MTCI) On/Off Capacity on Demand (CoD) Permanent capacity Permanent upgrade Purchased capacity Permanent / Temporary entitlement record Replacement capacity Resource Link
Reflects the permanent capacity with billable capacity only, without replacement capacity. If no billable temporary capacity is active, Model Temporary Capacity Identifier equals Model Permanent Capacity Identifier. Represents a function that allows a spare capacity in a CPC to be made available to increase the total capacity of a CPC. For example, On/Off CoD can be used to acquire additional capacity for the purpose of handling a workload peak. The capacity that a customer purchases and activates. This amount might be less capacity than the total capacity purchased. LIC licensed by IBM to enable the activation of applicable computing resources, such as processors or memory, for a specific CIU-eligible machine on a permanent basis. Capacity delivered to and owned by the customer. It can be higher than permanent capacity. The internal representation of a temporary (TER) or permanent (PER) capacity upgrade processed by the CIU facility. An entitlement record contains the encrypted representation of the upgrade configuration with the associated time limit conditions. A temporary capacity used for situations in which processing capacity in other parts of the enterprise is lost during either a planned event or an unexpected disaster. The two replacement offerings available are, Capacity for Planned Events and Capacity Backup. The IBM Resource Link is a technical support website providing comprehensive set of tools and resources available from the IBM Systems technical support site: http://www.ibm.com/servers/resourcelink/ An option, selected by the customer, that a second approver control each Capacity on Demand order. When a secondary approval is required, the request is sent for approval or cancellation to the Resource Link secondary user ID. The point when a record representing a capacity upgrade, either temporary or permanent, has been retrieved and loaded on the Support Element (SE) disk. For the zEC12, CP features (CP4, CP5, and CP6) provide reduced capacity relative to the full capacity CP feature (CP7). An optional capacity that is added to the current system capacity for a limited amount of time. It can be capacity that is owned or not owned by the customer.
Secondary approval
314
8049ch09.fm
Term
Description
Information that uniquely defines system, hardware, software, and microcode elements of a processing system.
315
8049ch09.fm
Billable capacity
To handle a peak workload, you can activate up to double the purchased capacity of any PU type temporarily and be charge based on a daily basis. The one billable capacity offering is On/Off Capacity on Demand (On/Off CoD).
Replacement capacity
When a processing capacity is lost in another part of an enterprise, replacement capacity can be activated. It allows you to activate any PU type up to authorized limit. The two replacement capacity offerings are: Capacity Backup Capacity for Planned Event
316
8049ch09.fm
This capability is based on the flexibility of the design and structure, which allows concurrent hardware installation and Licensed Internal Code (LIC) control over the configuration. The subcapacity models allow additional configuration granularity within the family. The added granularity is available for models configured with up to 20 CPs and provides 60 additional capacity settings. Subcapacity models provide for CP capacity increase in two dimensions that can be used together to deliver configuration granularity. The first dimension is by adding CPs to the configuration, the second is by changing the capacity setting of the CPs currently installed to a higher model capacity identifier. The zEC12 allows the concurrent and non-disruptive addition of processors to a running logical partition. As a result, you can have a flexible infrastructure, in which you can add capacity without pre-planning. This function is supported by z/OS, z/VM and z/VSE. There are two ways to accomplish this: With planning ahead for the future need of extra processors. In the logical partitions profile reserved processors can be specified. When the extra processor(s) are actually installed, the number of active processors for that LPAR can be increased without the need for an IPL. Another (easier) way is to enable the dynamic addition of CPUs through the z/OS LOADxx member. Parameter DYNCPADD in member LOADxx should be set to ENABLE. The zEC12 supports dynamic CPU addition just as the z196 and z10 do. The operating system has to be z/OS V1R10 or higher. Another function concerns the system assist processor (SAP). When additional SAPs are concurrently added to the configuration, the SAP-to-channel affinity is dynamically re-mapped on all SAPs on the system to rebalance the I/O configuration.
zEC12 zero CP Model capacity identifier is 400. This applies to an all IFL or an all ICF systems. I/O cage and the 8-slot I/O drawer cannot be ordered as a MES on zEC12. They are available on carry forward only.
317
8049ch09.fm
Hardware installation upgrade: Can change the system model 2827-Hvv, if additional books are included Can change the model capacity identifier, the capacity setting, or both The system model and the model capacity identifier can be concurrently changed. Concurrent upgrades can be accomplished for both permanent and temporary upgrades. Tip: A model upgrade can be concurrent by using concurrent book add (CBA), except for upgrades to Model HA1.
318
8049ch09.fm
Important: Customer planning and operator action are required to exploit concurrent PU conversion. Consider the following information about PU conversion: It is disruptive if all current PUs are converted to different types. It might require individual logical partition outage if dedicated PUs are converted. Unassigned CP capacity is recorded by a model capacity identifier. CP feature conversions change (increase or decrease) the model capacity identifier.
CIU prerequisites
The CIU facility supports LICCC upgrades only. It does not support I/O upgrades. All additional capacity required for an upgrade must be previously installed. Additional books or I/O cards cannot be installed as part of an order placed through the CIU facility. The sum of CPs, unassigned CPs, ICFs, zAAPs, zIIPs, IFLs, and unassigned IFLs cannot exceed the customer characterizable PU count of the installed books. The total number of zAAPs or zIIPs cannot each exceed the number of purchased CPs.
319
8049ch09.fm
Permanent upgrades
Permanent upgrades can be ordered by using the CIU facility. Through the CIU facility, you can generate online permanent upgrade orders to concurrently add processors (CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs) and memory, or change the model capacity identifier, up to the limits of the installed books on an existing system.
Temporary upgrades
The base model zEC12 describes permanent and dormant capacity (Figure 9-1) using the capacity marker and the number of PU features installed on the system. Up to eight temporary offerings can be present. Each offering has its own policies and controls and each can be activated or deactivated independently in any sequence and combination. Although multiple offerings can be active at any time, if enough resources are available to fulfill the offering specifications, only one On/Off CoD offering can be active at any time.
Customer defined policy or user commands Manual operations
Management Application
API
Activation
Enforce terms and conditions and physical model limitations Up to 8 records installed and active
Authorization Layer R1 R2 R3 R4 R5 R6 R7 R8
Dormant capacity Base model Purchased capacity Change permanent capacity through CIU or MES order
Temporary upgrades are represented in the system by a record. All temporary upgrade records, downloaded from the remote support facility (RSF) or installed from portable media, are resident on the Support Element (SE) hard drive. At the time of activation, the customer can control everything locally. Figure 9-1 shows a representation of the provisioning architecture. The authorization layer enables administrative control over the temporary offerings. The activation and deactivation can be driven either manually or under control of an application through a documented application program interface (API). By using the API approach, you can customize, at activation time, the resources necessary to respond to the current situation, up to the maximum specified in the order record. If the situation changes, you can add or remove resources without having to go back to the base
320
8049ch09.fm
configuration. This eliminates the need for temporary upgrade specification for all possible scenarios. However, for CPE the ordered configuration is the only possible activation. In addition, this approach enables you to update and replenish temporary upgrades, even in situations where the upgrades are already active. Likewise, depending on the configuration, permanent upgrades can be performed while temporary upgrades are active. Figure 9-2 shows examples of the activation sequence of multiple temporary upgrades.
R1
R2
R3
R4
CBU
R3
CBU
CBU
CPE
R1 R3 R1 R3
CBU CPE
R4 R2
CBU
R4 R2
R3 R2
CBU
CBU
OOCoD
R2
OOCoD
OOCoD
OOCoD
OOCoD
R2
Time
In the case of the R2, R3, and R1 being active at the same time, only parts of R1 can be activated, because not enough resources are available to fulfill all of R1. When R2 is then deactivated, the remaining parts of R1 can be activated as shown. Temporary capacity can be billable as On/Off Capacity on Demand (On/Off CoD), or replacement capacity as Capacity Backup (CBU) or CPE: On/Off CoD is a function that enables concurrent and temporary capacity growth of the system. On/Off CoD can be used for customer peak workload requirements, for any length of time, and has a daily hardware and maintenance charge. The software charges can vary according to the license agreement for the individual products. See your IBM Software Group representative for exact details. On/Off CoD can concurrently add processors (CPs, ICFs, zAAPs. zIIPs, IFLs, and SAPs), increase the model capacity identifier, or both, up to the limit of the installed books of an existing system, and is restricted to twice the currently installed capacity. On/Off CoD requires a contractual agreement between the customer and IBM. You decide whether to either pre-pay or post-pay On/Off CoD. Capacity tokens inside the records are used to control activation time and resources. CBU is a concurrent and temporary activation of additional CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs, an increase of the model capacity identifier, or both.
321
8049ch09.fm
CBU cannot be used for peak workload management in any form. As stated, On/Off CoD is the right way to do that. A CBU activation can last up to 90 days when a disaster or recovery situation occurs. CBU features are optional and require unused capacity to be available on installed books of the backup system, either as unused PUs or as a possibility to increase the model capacity identifier, or both. A CBU contract must be in place before the special code that enables this capability, can be loaded on the system. The standard CBU contract provides for five 10-day tests4 (the so called CBU test activation) and one 90-day activation over a five-year period. Contact your IBM Representative for details. You can run production workload on a CBU upgrade during a CBU test, provided that at least an equivalent amount of production capacity is shut down for the duration of the CBU test. If you already have existing CBU contracts, you will also need to sign an Amendment (US form #Z125-8145) with IBM. CPE is a concurrent and temporary activation of additional CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs or an increase of the model capacity identifier, or both. The CPE offering is used to replace temporary lost capacity within a customers enterprise for planned downtime events, for example, with data center changes. CPE cannot be used for peak load management of customer workload or for a disaster situation. The CPE feature requires unused capacity to be available on installed books of the backup system, either as unused PUs or as a possibility to increase the model capacity identifier on a subcapacity system, or both. A CPE contract must be in place before the special code that enables this capability can be loaded on the system. The standard CPE contract provides for one three-day planned activation at a specific date. Contact your IBM representative for details.
Permanent
CPs, ICFs, zAAPs, zIIPs, IFLs, SAPs, book, memory, and I/Os CPs, ICFs, zAAPs, zIIPs, IFLs, SAPs, and memory CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs
Installed by IBM service personnel Performed through the CIU facility Performed through the OOCoD facility Performed through the CBU facility Performed through the CPE facility
Temporary
zEC12 provides additional improvements in the CBU activation panels. These have been improved to prevent inadvertent CBU activation.
322
8049ch09.fm
level. The MES upgrade can be done using Licensed Internal Code Configuration Control (LICCC) only, by installing additional books, adding I/O cards, or a combination: MES upgrades for processors are done by any of the following methods: LICCC assigning and activating unassigned PUs up to the limit of the installed books LICCC to adjust the number and types of PUs or to change the capacity setting, or both Installing additional books and LICCC assigning and activating unassigned PUs on installed books MES upgrades for memory are done by either of the following methods: Using LICCC to activate additional memory capacity up to the limit of the memory cards on the currently installed books. Plan-ahead and flexible memory features enable you to have better control over future memory upgrades. For details about the memory features, see: 2.5.7, Preplanned Memory on page 54 2.5.6, Flexible Memory Option on page 53 Installing additional books and using LICCC to activate additional memory capacity on installed books Using the enhanced book availability (EBA), where possible, on multibook systems to add or change the memory cards MES upgrades for I/O are done by either of the following methods Installing additional I/O cards and supporting infrastructure if required on PCIe drawers that are already installed, or installing additional PCIe drawers to hold the new cards. MES upgrades for the zEnterprise BladeCenter Extension can only be performed through your IBM customer representative. An MES upgrade requires IBM service personnel for the installation. In most cases, the time required for installing the LICCC and completing the upgrade is short. To better exploit the MES upgrade function, it is highly desirable to carefully plan the initial configuration to allow a concurrent upgrade to a target configuration. The availability of PCIe I/O drawers has improved the flexibility to do unplanned I/O configuration changes concurrently. The store system information (STSI) instruction gives more useful and detailed information about the base configuration and about temporary upgrades. This enables you to more easily resolve billing situations where Independent Software Vendor (ISV) products are in use. The model and model capacity identifier returned by the STSI instruction are updated to coincide with the upgrade. See Store System Information (STSI) instruction on page 349 for more details. Upgrades: The MES provides the physical upgrade, resulting in more enabled processors, different capacity settings for the CPs, additional memory, I/O ports and I/O drawers. Additional planning tasks are required for non-disruptive logical upgrades (see Guidelines to avoid disruptive upgrades on page 352).
323
8049ch09.fm
subcapacity models, additional capacity can be provided by adding CPs, by changing the capacity identifier on the current CPs, or by doing both. Limits: The sum of CPs, inactive CPs, ICFs, zAAPs, zIIPs, IFLs, unassigned IFLs, and SAPs cannot exceed the maximum limit of PUs available for customer use. The number of zAAPs and the number of zIIPs cannot exceed each the number of purchased CPs.
Example of MES upgrade: Figure 9-3 is an example of an MES upgrade for processors,
showing two upgrade steps.
2827-H20
MCI 708
Book 0
8 CPs
CP0
CP1
CP2
CP3
CP4
CP5
CP6
CP7
Spare
Spare
Spare
Spare
Book 0 20 CPs
Book 1
6 CPs
CP0
CP1
CP2
CP3
CP4
CP5
CP6
CP7
CP8
CP9
CP20
CP21
CP22
CP23
CP24
CP25
Spare
CP10
CP11
CP12
CP13
CP14
CP15 CP16
CP17
CP18
CP19
Spare
Spare
Book 1
20 CPs
Book 1
7 CPs 2 IFLs
CP0
CP1
CP2
CP3
CP4
CP5
CP6
CP7
CP8
CP9
CP20
CP21
CP22
CP23
CP24
CP25
CP26 Spare
Spare
Spare
CP10 CP11
CP12
CP13
CP14
CP15 CP16
CP17
CP18
CP19
IFL1
IFL0
A model H20 (one book), model capacity identifier 708 (eight CPs), is concurrently upgraded to a model H43 (two books), with model capacity identifier (MCI) 726 (which is 26 CPs). The model upgrade requires adding a book and assigning and activating eighteen PUs as CPs. Then, model H43, MCI 726, is concurrently upgraded to a capacity identifier 727 (which is 27 CPs) with two IFLs by assigning and activating three more unassigned PUs (one as CP and two as IFLs). If needed, additional logical partitions can be created concurrently to use the newly added processors. Important: Up to 101 logical processors, including reserved processors, can be defined to a logical partition. However, do not define more processors to a logical partition than the target operating system supports. The following table describes the number of processors supported by various z/OS and z/VM releases.
324
8049ch09.fm
Table 9-3 Number of processors supported by operating system Operating System Number of processors supported
z/OS V1R10 w/ PTF z/OS V1R11 w/ PTF z/OS V1R12 w/ PTF z/OS V1R13 w/ PTF z/VM V5R4 - z/VM V6R2 z/VSE z/TPF Linux on System z
64 99 99 99 32 z/VSE Turbo Dispatcher can exploit up to 4 CPs and tolerates up to 10-way LPARs 86 CPs SUSE SLES 10: 64 CPs or IFLs SUSE SLES 11: 64 CPs or IFLs Red Hat RHEL 5: 80 CPs or IFLs Red Hat RHEL 6: 80 CPs or IFLs
Software charges, based on the total capacity of the system on which the software is installed, are adjusted to the new capacity after the MES upgrade. Software products that use Workload License Charge (WLC) might not be affected by the system upgrade, because their charges are based on partition utilization and not based on the system total capacity. For more information about WLC, see 8.12, Software licensing considerations on page 305.
325
8049ch09.fm
The one-book model H20 has a minimum of 80 GB physical installed memory. The customer addressable storage in this case is 32 GB. If you require more than that, an additional memory upgrade can install up to 704 GB of memory for customer use, by changing the existing DIMM sizes and adding additional DIMMs in all available slots in the book. Another possibility is to add memory by concurrently adding a second book with sufficient memory into the configuration and then using LICCC to enable that memory. A logical partition can dynamically take advantage of a memory upgrade if reserved storage has been defined to that logical partition. The reserved storage is defined to the logical partition as part of the image profile. Reserved memory can be configured online to the logical partition by using the LPAR dynamic storage reconfiguration (DSR) function. DSR allows a z/OS operating system image, and z/VM partitions, to add reserved storage to their configuration if any unused storage exists. The nondisruptive addition of storage to a z/OS and z/VM partition necessitates that pertinent operating system parameters have been prepared. If reserved storage has not been defined to the logical partition, the logical partition must be deactivated, the image profile changed, and the logical partition reactivated to allow the additional storage resources to be available to the operating system image.
0 0 0-5
0 0 0-5
The number of cards that can be in a carry forward, is limited. The following table describes this in more detail.
Table 9-5 Number of I/O cards and I/O Cage / Drawers Number of cards in carry forward Number of I/O drawers Number of Cargo I/O cages
1-8
326
8049ch09.fm
9 - 16 17 - 28 29 - 36 37 - 44
2 0 1 2
0 1 1 1
Important: The maximum number of legacy I/O cards on a carry forward is 44. Depending on the amount of I/O features carried forward on an upgrade, the configurator will determine the number and mix of I/O cages, I/O drawers, and PCIe I/O drawers. To better exploit the MES for I/O capability, an initial configuration should be carefully planned to allow concurrent upgrades up to the target configuration. If legacy I/O features will be removed from the I/O cage/drawer, the configurator will not physically remove the I/O cage/drawer unless the I/O frame slot(s) are required to install a new PCIe I/O drawer. If an PCIe I/O drawer will be added to an existing zEC12 and some legacy features will need to be physically moved from one I/O cage/drawer to another I/O cage/drawer to completely empty the I/O cage/drawer for removal, legacy card moves are disruptive. I/O cage removal is disruptive. z/VSE, z/TPF,Linux on System z, and CFCC do not provide dynamic I/O configuration support. The installation of the new hardware is performed concurrently, but defining the new hardware to these operating systems requires an IPL.
Note: The zEC12 has a Hardware System Area (HSA) of 32 GB whereas the z196 has 16 GB HSA. It is not part of the customer purchased memory.
327
8049ch09.fm
Enables potential of 20 Gb Ethernet bandwidth via link aggregation. Doubled 10 GbE cables between BladeCenter 10 GbE switch and 10 GbE TOR (Top of Rack) switch. Doubled 10 GbE cables between the BladeCenter 10 GbE switches. New version of the Advanced Management Module (AMM) in the BladeCenter chassis. Upgrade hypervisors and other firmware changes.
8049ch09.fm
The capability to add permanent upgrades to a given system through the CIU facility requires that the permanent upgrade enablement feature (FC 9898) be installed on the system. A permanent upgrade might change the system model capacity identifier 4xx, 5yy, 6yy, or 7nn if additional CPs are requested or the capacity identifier is changed as part of the permanent upgrade, but it cannot change the system model. If necessary, additional logical partitions can be created concurrently to use the newly added processors. Attention: A permanent upgrade of processors can provide a physical concurrent upgrade, resulting in more enabled processors available to a system configuration. Thus, additional planning and tasks are required for nondisruptive logical upgrades. See Guidelines to avoid disruptive upgrades on page 352 for more information. Maintenance charges are automatically adjusted as a result of a permanent upgrade. Software charges based on the total capacity of the system on which the software is installed are adjusted to the new capacity in place after the permanent upgrade is installed. Software products that use Workload License Charge (WLC) might not be affected by the system upgrade, because their charges are based on a logical partition utilization and not based on the system total capacity. See 8.12.2, Advanced Workload License Charges (AWLC) on page 307, for more information about WLC. Figure 9-4 illustrates the CIU facility process on the IBM Resource Link.
Customer ibm.com/servers/resourcelink
Internet
Online permanent order
The following sample sequence on the IBM Resource Link initiates an order: 1. Sign on to Resource Link. 2. Select the Customer Initiated Upgrade option from the main Resource Link page. Customer and system details associated with the user ID are listed. 3. Select the system that will receive the upgrade. The current configuration (PU allocation and memory) is shown for the selected system. 4. Select the Order Permanent Upgrade function. The Resource Link limits options to those that are valid or possible for the selected configuration (system). 5. After the target configuration is verified by the system, accept or cancel the order. An order is created and verified against the pre-established agreement.
329
8049ch09.fm
6. Accept or reject the price that is quoted. A secondary order approval is optional. Upon confirmation, the order is processed. The LICCC for the upgrade should be available within hours. Figure 9-5 illustrates the process for a permanent upgrade. When the LICCC is passed to the remote support facility, you are notified through an e-mail that the upgrade is ready to be downloaded.
EC12
Figure 9-5 CIU-eligible order activation example
The two major components in the process are ordering and retrieval (along with activation).
9.4.1 Ordering
Resource Link provides the interface that enables you to order a concurrent upgrade for a system. You can create, cancel, view the order, and view the history of orders that were placed through this interface. Configuration rules enforce only valid configurations being generated within the limits of the individual system. Warning messages are issued if you select invalid upgrade options. The process allows only one permanent CIU-eligible order for each system to be placed at a time. For a tutorial see: https://www-304.ibm.com/servers/resourcelink/hom03010.nsf/pages/CIUInformation?Ope nDocument
330
8049ch09.fm
Figure 9-6 shows the initial view of the machine profile on Resource Link.
The number of CPs, ICFs, zAAPs, zIIPs, IFLs, SAPs, memory size, and unassigned IFLs on the current configuration are displayed on the left side of the web page. Resource Link retrieves and stores relevant data associated with the processor configuration, such as the number of CPs and installed memory cards. It allows you to select only those upgrade options that are deemed valid by the order process. It allows upgrades only within the bounds of the currently installed hardware.
331
8049ch09.fm
In the Perform Model Conversion panel, select the Permanent upgrades option to start the process. See Figure 9-7.
The panel provides several possible options. If you select the Retrieve and apply data option, you are prompted to enter the order activation number to initiate the permanent upgrade. See Figure 9-8.
9.5.1 Overview
The capacity for CPs is expressed in MSUs. Capacity for speciality engines is expressed in number of speciality engines. Capacity tokens are used to limit the resource consumption for all types of processor capacity. Capacity tokens are introduced to provide better control over resource consumption when On/Off CoD offerings are activated. Token are represented as follows: 332
IBM zEnterprise EC12 Technical Guide
8049ch09.fm
For CP capacity, each token represents the amount of CP capacity that will result in one MSU of software cost for one day (an MSU-day token). For speciality engines, each token is equivalent to one speciality engine capacity for one day (an engine-day token). Tokens are by capacity type, MSUs for CP capacity, and number of engines for speciality engines. Each speciality engine type has its own tokens, and each On/Off CoD record has separate token pools for each capacity type. During the ordering sessions on Resource Link, you decide how many tokens of each type should be created in an offering record. Each engine type must have tokens for that engine type to be activated. Capacity that has no tokens cannot be activated. When resources from an On/Off CoD offering record containing capacity tokens are activated, a billing window is started. A billing window is always 24 hours in length. Billing takes place at the end of each billing window. The resources billed are the highest resource usage inside each billing window for each capacity type. An activation period is one or more complete billing windows, and represents the time from the first activation of resources in a record until the end of the billing window in which the last resource in a record is deactivated. At the end of each billing window, the tokens are decremented by the highest usage of each resource during the billing window. If any resource in a record does not have enough tokens to cover usage for the next billing window, the entire record will be deactivated. On/Off CoD requires that the Online CoD Buying feature (FC 9900) be installed on the system that is to be upgraded. On/Off CoD to Permanent Upgrade Option is a new offering, which is an offshoot of On/Off CoD and takes advantage of the aspects of the architecture. The customer is given a window of opportunity to assess capacity additions to their permanent configurations using On/Off CoD. If a purchase is made, the hardware On/Off CoD charges during this window, 3 days or less, are waived. If no purchase is made, then the customer is charged for the temporary use. The resources eligible for temporary use are CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs. Temporary addition of memory and I/O ports is not supported. Unassigned PUs that are on the installed books can be temporarily and concurrently activated as CPs, ICFs, zAAPs, zIIPs, IFLs, SAPs through LICCC, up to twice the currently installed CP capacity and up to twice the number of ICFs, zAAPs, zIIPs, or IFLs. This means that an On/Off CoD upgrade cannot change the system model. The addition of new books is not supported. However, activation of an On/Off CoD upgrade can increase the model capacity identifier 4xx, 5yy, 6yy, or 7nn.
9.5.2 Ordering
Concurrently installing temporary capacity by ordering On/Off CoD is possible, as follows: CP features equal to the MSU capacity of installed CPs IFL features up to the number of installed IFLs ICF features up to the number of installed ICFs zAAP features up to the number of installed zAAPs zIIP features up to the number of installed zIIPs SAPs up to four for model H20, eight for an H43, twelve for an H66, sixteen for an H89 and HA1. On/Off CoD can provide CP temporary capacity in two ways:
333
8049ch09.fm
By increasing the number of CPs. For subcapacity models, capacity can be added by increasing the number of CPs or by changing the capacity setting of the CPs, or both. The capacity setting for all CPs must be the same. If the On/Off CoD is adding CP resources that have a capacity setting different from the installed CPs, then the base capacity settings are changed to match. On/Off CoD has the following limits associated with its use: The number of CPs cannot be reduced. The target configuration capacity is limited to: Twice the currently installed capacity, expressed in MSUs for CPs. Twice the number of installed IFLs, ICFs, zAAPs, and zIIPs. The number of SAPs that can be activated depends on the model described in 9.2.1, Model upgrades on page 317.
On/Off CoD can be ordered as prepaid or postpaid: A prepaid On/Off CoD offering record contains resource descriptions, MSUs, number of speciality engines, and tokens that describe the total capacity that can be used. For CP capacity, the token contains MSU-days; for speciality engines, the token contains speciality engine-days. When resources on a prepaid offering are activated, they must have enough capacity tokens to allow the activation for an entire billing window, which is 24 hours. The resources remain active until you deactivate them or until one resource has consumed all of its capacity tokens. When that happens, all activated resources from the record are deactivated. A postpaid On/Off CoD offering record contains resource descriptions, MSUs, speciality engines, and can contain capacity tokens describing MSU-days and speciality engine-days. When resources in a postpaid offering record without capacity tokens are activated, those resources remain active until they are deactivated, or until the offering record expires, which is usually 180 days after its installation. When resources in a postpaid offering record with capacity tokens are activated, those resources must have enough capacity tokens to allow the activation for an entire billing window (24 hours). The resources remain active until they are deactivated or until one of the resource tokens are consumed, or until the record expires, usually 180 days after its installation. If one capacity token type is consumed, resources from the entire record are deactivated. As an example, for a zEC12 with capacity identifier 502, two ways to deliver a capacity upgrade through On/Off CoD exist: The first option is to add CPs of the same capacity setting. With this option, the model capacity identifier could be changed to a 503, which would add one additional CP (making it a 3-way) or to a 504, which would add two additional CPs (making it a 4-way). The second option is to change to a different capacity level of the current CPs and change the model capacity identifier to a 602 or to a 702. The capacity level of the CPs is increased but no additional CPs are added. The 502 could also be temporarily upgraded to a 603 as indicated in the table, thus increasing the capacity level and adding another processor. The 420 does not have an upgrade path through On/Off CoD. Preferably, use the Large Systems Performance Reference (LSPR) information to evaluate the capacity requirements according to your workload type. LSPR data for current IBM processors is available at this website:
334
8049ch09.fm
https://www-304.ibm.com/servers/resourcelink/lib03060.nsf/pages/lsprindex The On/Off CoD hardware capacity is charged on a 24-hour basis. There is a grace period at the end of the On/Off CoD day. This allows up to an hour after the 24-hour billing period to either change the On/Off CoD configuration for the next 24-hour billing period or deactivate the current On/Off CoD configuration. The times when the capacity is activated and deactivated are maintained in the zEC12 and sent back to the support systems. If On/Off capacity is already active, additional On/Off capacity can be added without having to return the system to its original capacity. If the capacity is increased multiple times within a 24-hour period, the charges apply to the highest amount of capacity active in the period. If additional capacity is added from an already active record containing capacity tokens, a check is made to control that the resource in question has enough capacity to be active for an entire billing window (24 hours). If that criteria is not met, no additional resources will be activated from the record. If necessary, additional logical partitions can be activated concurrently to use the newly added processor resources. Attention: On/Off CoD provides a concurrent hardware upgrade, resulting in more enabled processors available to a system configuration. Additional planning tasks are required for nondisruptive upgrades. See Guidelines to avoid disruptive upgrades on page 352. To participate in this offering, you must have accepted contractual terms for purchasing capacity through the Resource Link, established a profile, and installed an On/Off CoD enablement feature on the system. Subsequently, you can concurrently install temporary capacity up to the limits in On/Off CoD and use it for up to 180 days. Monitoring occurs through the system call-home facility and an invoice is generated if the capacity has been enabled during the calendar month. The customer will continue to be billed for use of temporary capacity until the system is returned to the original configuration. If the On/Off CoD support is no longer needed, the enablement code must be removed. On/Off CoD orders can be pre-staged in Resource Link to allow multiple optional configurations. The pricing of the orders is done at the time of the order, and the pricing can vary from quarter to quarter. Staged orders can have different pricing. When the order is downloaded and activated, the daily costs are based on the pricing at the time of the order. The staged orders do not have to be installed in order sequence. If a staged order is installed out of sequence, and later an order that was staged that had a higher price is downloaded, the daily cost will be based on the lower price. Another possibility is to store unlimited On/Off CoD LICCC records on the Support Element with the same or different capacities at any given time, giving greater flexibility to quickly enable needed temporary capacity. Each record is easily identified with descriptive names, and you can select from a list of records that can be activated. Resource Link provides the interface that allows you to order a dynamic upgrade for a specific system. You are able to create, cancel, and view the order. Configuration rules are enforced, and only valid configurations are generated based on the configuration of the individual system. After completing the prerequisites, orders for the On/Off CoD can be placed. The order process is to use the CIU facility on Resource Link. You can order temporary capacity for CPs, ICFs, zAAPs, zIIPs, IFLs, or SAPs. Memory and channels are not supported on On/Off CoD. The amount of capacity is based on the amount of owned capacity for the different types of resources. An LICCC record is established and staged to Resource Link for this order. After the record is activated, it has no expiration date.
Chapter 9. System upgrades
335
8049ch09.fm
However, an individual record can only be activated once. Subsequent sessions require a new order to be generated, producing a new LICCC record for that specific order. Alternatively the customer can use an auto renewal feature to eliminate the need for a manual replenishment of the On/Off CoD order. The is feature is implemented in Resource Link and the customer will have to check this feature in the machine profile. See Figure 9-9 on page 336 for more details.
336
8049ch09.fm
The example order in Figure 9-10 is a On/Off CoD order for 100% more CP capacity and for one ICF, one zAAP, one zIIP, and one SAP. The maximum number of CPs, ICFs, zAAPs, zIIPs, and IFLs is limited by the current number of available unused PUs of the installed books. The maximum number of SAPs is determined by the model number and the number of available PUs on the already installed books.
337
8049ch09.fm
With this API, automation code can be used to send an activation command along with the order number to the HMC to enable the order.
9.5.5 Termination
A customer is contractually obligated to terminate the On/Off CoD right-to-use feature when a transfer in asset ownership occurs. A customer can also choose to terminate the On/Off CoD right-to-use feature without transferring ownership. Application of FC 9898 terminates the right to use the On/Off CoD. This feature cannot be ordered if a temporary session is already active. Similarly, the CIU enablement feature cannot be removed if a temporary session is active. Any time the CIU enablement feature is removed, the On/Off CoD right-to-use is simultaneously removed. Reactivating the right-to-use feature subjects the customer to the terms and fees that apply at that time.
Monitoring
When you activate an On/Off CoD upgrade, an indicator is set in vital product data. This indicator is part of the call-home data transmission, which is sent on a scheduled basis. A time stamp is placed into call-home data when the facility is deactivated. At the end of each calendar month, the data is used to generate an invoice for the On/Off CoD that has been used during that month.
Maintenance
The maintenance price is adjusted as a result of an On/Off CoD activation.
Software
Software Parallel Sysplex License Charge (PSLC) customers are billed at the MSU level represented by the combined permanent and temporary capacity. All PSLC products are billed at the peak MSUs enabled during the month, regardless of usage. Customers with WLC licenses are billed by product at the highest four-hour rolling average for the month. In this instance, temporary capacity does not necessarily increase the software bill until that capacity is allocated to logical partitions and actually consumed. Results from the STSI instruction reflect the current permanent and temporary CPs. See Store System Information (STSI) instruction on page 349 for more details.
338
8049ch09.fm
Ethernet Switch
WLM
WLM
Provisioning policy
(SNMP)
The z/OS WLM manages the workload by goals and business importance on each z/OS system. WLM metrics are available through existing interfaces and are reported through Resource Measurement Facility (RMF) Monitor III, with one RMF gatherer for each z/OS system. Sysplex-wide data aggregation and propagation occur in the RMF Distributed Data Server (DDS). The RMF Common Information Model (CIM) providers and associated CIM models publish the RMF Monitor III data. The Capacity Provisioning Manager (CPM), a function inside z/OS, retrieves critical metrics from one or more z/OS systems through the Common Information Model (CIM) structures and protocol. CPM communicates to (local or remote) Support Elements and HMCs through the SNMP protocol. CPM has visibility of the resources in the individual offering records, and the capacity tokens. When CPM decides to activate resources, a check is performed to determine whether enough capacity tokens remain for the specified resource to be activated for at least 24 hours. If not enough tokens remain, no resource from the On/Off CoD record is activated. If a capacity token is completely consumed during an activation driven by the CPM, the corresponding On/Off CoD record is deactivated prematurely by the system, even if the CPM has activated this record, or parts of it. You do, however, receive warning messages if capacity tokens are getting close to being fully consumed. You receive the messages five days before a capacity token is fully consumed. The five days are based on the assumption that the consumption will be constant for the 5 days. You will need to put operational
339
8049ch09.fm
procedures in place to handle these situations. You can either deactivate the record manually, let it happen automatically, or replenish the specified capacity token by using the Resource Link application. The Capacity Provisioning Control Center (CPCC), which resides on a workstation, provides an interface to administer capacity provisioning policies. The CPCC is not required for regular CPM operation. The CPCC will over time be moved into the z/OSMF. Parts of the CPCC has been included in z/OSMF V1R13.
The Capacity Provisioning Domain represents the central processor complexes (CPCs) that are controlled by the Capacity Provisioning Manager. The HMCs of the CPCs within a CPD must be connected to the same processor LAN. Parallel Sysplex members can be part of a CPD. There is no requirement that all members of a Parallel Sysplex must be part of the CPD, but participating members must all be part of the same CPD. The Capacity Provisioning Control Center (CPCC) is the user interface component. Administrators work through this interface to define domain configurations and provisioning policies, but it is not needed during production. The CPCC is installed on a Microsoft Windows workstation. CPM operates in four modes, allowing four different levels of automation:
340
8049ch09.fm
Manual mode: Use this command-driven mode when no CPM policy is active. Analysis mode: In analysis mode: CPM processes capacity-provisioning policies and informs the operator when a provisioning or deprovisioning action is required according to policy criteria. The operator determines whether to ignore the information or to manually upgrade or downgrade the system by using the HMC, the SE, or available CPM commands. Confirmation mode: In this mode, CPM processes capacity provisioning policies and interrogates the installed temporary offering records. Every action proposed by the CPM needs to be confirmed by the operator. Autonomic mode: This mode is similar to the confirmation mode, but no operator confirmation is required. A number of reports are available in all modes, containing information about workload and provisioning status and the rationale for provisioning guidelines. User interfaces are provided through the z/OS console and the CPCC application. The provisioning policy defines the circumstances under which additional capacity can be provisioned (when, which, and how). These are the three elements in the criteria: A time condition is when provisioning is allowed, as follows: Start time indicates when provisioning can begin. Deadline indicates that provisioning of additional capacity no longer allowed End time indicates that deactivation of additional capacity should begin. A workload condition is which work qualifies for provisioning. Parameters include: The z/OS systems that can execute eligible work Importance filter indicates eligible service class periods, identified by WLM importance Performance Index (PI) criteria: Activation threshold: PI of service class periods must exceed the activation threshold for a specified duration before the work is considered to be suffering. Deactivation threshold: PI of service class periods must fall below the deactivation threshold for a specified duration before the work is considered to no longer be suffering.
Included service classes are eligible service class periods. Excluded service classes are service class periods that should not be considered. Tip: If no workload condition is specified, the full capacity described in the policy will be activated and deactivated at the start and end times specified in the policy. Provisioning scope is how much additional capacity can be activated, expressed in MSUs. Specified in MSUs, number of zAAPs, and number of zIIPs must be one specification per CPC that is part of the Capacity Provisioning Domain. The maximum provisioning scope is the maximum additional capacity that can be activated for all the rules in the Capacity Provisioning Domain.
341
8049ch09.fm
The provisioning rule is as follows: In the specified time interval, if the specified workload is behind its objective, then up to the defined additional capacity can be activated. The rules and conditions are named and stored in the Capacity Provisioning Policy. For more information about z/OS Capacity Provisioning functions, see z/OS MVS Capacity Provisioning Users Guide, SA33-8299.
342
8049ch09.fm
or repair work, replacement capacity can be installed temporarily on another zEC12 in the customers environment. Attention: CPE is for planned replacement capacity only and cannot be used for peak workload management. The feature codes are: FC 6833 Capacity for Planned Event enablement FC 0116 - 1 CPE Capacity Unit FC 0117 - 100 CPE Capacity Unit FC 0118 - 10000 CPE Capacity Unit FC 0119 - 1 CPE Capacity Unit-IFL FC 0120 - 100 CPE Capacity Unit-IFL FC 0121 - 1 CPE Capacity Unit-ICF FC 0122 - 100 CPE Capacity Unit-ICF FC 0123 - 1 CPE Capacity Unit-zAAP FC 0124 - 100 CPE Capacity Unit-zAAP FC 0125 - 1 CPE Capacity Unit-zIIP FC 0126 - 100 CPE Capacity Unit-zIIP FC 0127 - 1 CPE Capacity Unit-SAP FC 0128 - 100 CPE Capacity Unit-SAP The feature codes are calculated automatically when the CPE offering is configured. Whether using the eConfig tool or the Resource Link, a target configuration must be ordered consisting of a model identifier or a number of speciality engines, or both. Based on the target configuration, a number of feature codes from the list above is calculated automatically and a CPE offering record is constructed. CPE is intended to replace capacity lost within the enterprise because of a planned event such as a facility upgrade or system relocation. CPE is intended for short duration events lasting up to a maximum of three days. Each CPE record, after it is activated, gives the you access to dormant PUs on the system that you have a contract for as described above by the feature codes. Processor units can be configured in any combination of CP or specialty engine types (zIIP, zAAP, SAP, IFL, and ICF). At the time of CPE activation the contracted configuration will be activated. The general rule of one zIIP and one zAAP for each configured CP will be controlled for the contracted configuration. The processors that can be activated by CPE come from the available unassigned PUs on any installed book. CPE features can be added to an existing zEC12 non-disruptively. A one-time fee is applied for each individual CPE event depending on the contracted configuration and its resulting feature codes. Only one CPE contract can be ordered at a time. The base system configuration must have sufficient memory and channels to accommodate the potential requirements of the large CPE-configured system. It is important to ensure that all required functions and resources are available on the system where CPE is activated, including CF LEVELs for coupling facility partitions, memory, cryptographic functions, and including connectivity capabilities. The CPE configuration is activated temporarily and provides additional PUs in addition to the systems original, permanent configuration. The number of additional PUs is predetermined by the number and type of feature codes configured as described above by the feature codes. The number PUs that can be activated is limited by the unused capacity available on the system. For example: A model H43 with 16 CPs, no IFLs, ICFs, or zAAPs, has 27 unassigned PUs available.
343
8049ch09.fm
A model H66 with 28 CPs, 1 IFL, and 1 ICF has 36 unassigned PUs available. When the planned event is over, the system must be returned to its original configuration. You can deactivate the CPE features at any time before the expiration date. A CPE contract must be in place before the special code that enables this capability can be installed on the system. CPE features can be added to an existing zEC12 non-disruptively.
9.7.1 Ordering
The CBU process allows for CBU to activate CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs. To be able to use the CBU process a CBU enablement feature (FC 9910) must be ordered and installed. You must order the quantity and type of PU that you require. The feature codes are: 6805: Additional test activations 6817: Total CBU years ordered 6818: CBU records ordered 6820: Single CBU CP-year 6821: 25 CBU CP-year 6822: Single CBU IFL-year 6823: 25 CBU IFL-year 6824: Single CBU ICF-year 6825: 25 CBU ICF-year 6826: Single CBU zAAP-year 6827: 25 CBU zAAP-year 6828: Single CBU zIIP-year 6829: 25 CBU zIIP-year 6830: Single CBU SAP-year 6831: 25 CBU SAP-year 6832: CBU replenishment The CBU entitlement record (6818) contains an expiration date that is established at the time of order and is dependent upon the quantity of CBU years (6817). You have the capability to
5
All new CBU contract documents contain new CBU Test terms to allow execution of production workload during CBU test. Existing CBU customers will need to execute IBM Customer Agreement Amendment for IBM System z Capacity Backup Upgrade Tests (US form #Z125-8145).
344
8049ch09.fm
extend your CBU entitlements through the purchase of additional CBU years. The number of 6817 per instance of 6818 remains limited to five and fractional years are rounded up to the near whole integer when calculating this limit. For instance, if there are two years and eight months to the expiration date at the time of order, the expiration date can be extended by no more than two additional years. One test activation is provided for each additional CBU year added to the CBU entitlement record. Feature code 6805 allows for ordering additional tests in increments of one. The total number of tests allowed is 15 for each feature code 6818. The processors that can be activated by CBU come from the available unassigned PUs on any installed book. The maximum number of CBU features that can be ordered is 101. The number of features that can be activated is limited by the number of unused PUs on the system. For example: A model H20 with Capacity Model Identifier 410 can activate up to 20 CBU features: ten to change the capacity setting of the existing CPs and ten to activate unused PUs. A model H43 with 15 CPs, four IFLs, and one ICF has twenty three unused PUs available. It can activate up to twenty three CBU features. However, the ordering system allows for over-configuration in the order itself. You can order up to 101 CBU features regardless of the current configuration, however at activation, only the capacity already installed can be activated. Note that at activation, you can decide to activate only a sub-set of the CBU features that are ordered for the system. Subcapacity makes a difference in the way the CBU features are done. On the full-capacity models, the CBU features indicate the amount of additional capacity needed. If the amount of necessary CBU capacity is equal to four CPs, then the CBU configuration would be four CBU CPs. The subcapacity models have multiple capacity settings of 4xx, 5yy, or 6yy; the standard models have capacity setting 7nn. The number of CBU CPs must be equal to or greater than the number of CPs in the base configuration, and all the CPs in the CBU configuration must have the same capacity setting. For example, if the base configuration is a 2-way 402, then providing a CBU configuration of a 4-way of the same capacity setting requires two CBU feature codes. If the required CBU capacity changes the capacity setting of the CPs, then going from model capacity identifier 402 to a CBU configuration of a 4-way 504 would require four CBU feature codes with a capacity setting of 5yy. If the capacity setting of the CPs is changed, then more CBU features are required, not more physical PUs. This means that your CBU contract requires more CBU features if the capacity setting of the CPs is changed. Note that CBU can add CPs through LICCC-only, and the zEC12 must have the proper number of books installed to allow the required upgrade. CBU can change the model capacity identifier to a higher value than the base setting, 4xx, 5yy, or 6yy, but does not change the system model. The CBU feature cannot decrease the capacity setting. A CBU contract must be in place before the special code that enables this capability can be installed on the system. CBU features can be added to an existing zEC12 non-disruptively. For each machine enabled for CBU, the authorization to use CBU is available for a definite number of years of 1-5 years. The alternate configuration is activated temporarily and provides additional capacity greater than the systems original, permanent configuration. At activation time, you determine the capacity required for a given situation, and you can decide to activate only a sub-set of the capacity specified in the CBU contract.
Chapter 9. System upgrades
345
8049ch09.fm
The base system configuration must have sufficient memory and channels to accommodate the potential requirements of the large CBU target system. Ensure that all required functions and resources are available on the backup systems, including CF LEVELs for coupling facility partitions, memory, and cryptographic functions, as well as connectivity capabilities. When the emergency is over (or the CBU test is complete), the system must be taken back to its original configuration. The CBU features can be deactivated by the customer at any time before the expiration date. Failure to deactivate the CBU feature before the expiration date can cause the system to degrade gracefully back to its original configuration. The system does not deactivate dedicated engines, or the last of in-use shared engines. Planning: CBU for processors provides a concurrent upgrade, resulting in more enabled processors or changed capacity settings available to a system configuration, or both. You decide, at activation time, to activate a sub-set of the CBU features ordered for the system. Thus, additional planning and tasks are required for nondisruptive logical upgrades. See Guidelines to avoid disruptive upgrades on page 352. For detailed instructions, see the System z Capacity on Demand Users Guide, SC28-6846.
CBU activation
CBU is activated from the SE by using the Perform Model Conversion task or through automation by using API on the SE or the HMC. In case of real disaster, use the Activate CBU option to activate the 90-day period.
Image upgrades
After the CBU activation, the zEC12 can have more capacity, more active PUs, or both. The additional resources go into the resource pools and are available to the logical partitions. If the logical partitions have to increase their share of the resources, the logical partition weight can be changed or the number of logical processors can be concurrently increased by configuring reserved processors online. The operating system must have the capability to concurrently configure more processors online. If necessary, additional logical partitions can be created to use the newly added capacity.
CBU deactivation
To deactivate the CBU, the additional resources have to be released from the logical partitions by the operating systems. In some cases, this is a matter of varying the resources offline. In other cases, it can mean shutting down operating systems or deactivating logical partitions. After the resources have been released, the same facility on the SE is used to turn off CBU. To deactivate CBU, click the Undo temporary upgrade option from the Perform Model Conversion task on the SE.
CBU testing
Test CBUs are provided as part of the CBU contract. CBU is activated from the SE by using the Perform Model Conversion task. Select the test option to initiate a 10-day test period. A standard contract allows one test per CBU year. However, you can order additional tests in
346
8049ch09.fm
increments of one up to a maximum of 15 for each CBU order. The test CBU has a 10-day limit and must be deactivated in the same way as the real CBU, using the same facility through the SE. Failure to deactivate the CBU feature before the expiration date can cause the system to degrade gracefully back to its original configuration. The system does not deactivate dedicated engines, or the last of in-use shared engine. Testing can be accomplished by ordering a diskette, calling the support center, or using the facilities on the SE. The customer has the possibility of purchasing additional tests.
CBU example
An example of a capacity backup operation could be as follows; 12 CBU features are installed on a backup model H43 with model capacity identifier 708. When a production model H20 with model capacity identifier 708 has an unplanned outage, the backup system can be temporarily upgraded from model capacity identifier 708 to 720, so that the capacity can take over the workload from the failed production system. Furthermore, you can configure systems to back up each other. For example, if you use two models of H20 model capacity identifier 705 for the production environment, each can have five or more features installed. If one system suffers an outage, the other one uses a temporary upgrade to recover approximately the total original capacity.
347
8049ch09.fm
configure more capacity online. z/OS operating systems have this capability. z/VM can concurrently configure new processors and I/O devices online, memory can be dynamically added to z/VM partitions. If the concurrent upgrade is intended to satisfy the need for more operating system images, additional logical partitions can be created concurrently on the zEC12 system, including all resources needed by such logical partitions. These additional logical partitions can be activated concurrently. These enhanced configuration options are made available through the separate HSA, which was introduced on the System z196. Linux operating systems in general do not have the capability of adding more resources concurrently. However, Linux, and other types of virtual machines running under z/VM, can benefit from the z/VM capability to non-disruptively configure more resources online (processors and I/O). With z/VM, Linux guests can manipulate their logical processors through the use of the Linux CPU hotplug daemon. The daemon can start and stop logical processors based on the Linux average load value. The daemon is available in Linux SLES 10 SP2 and Red Hat RHEL V5R4.
9.8.1 Components
The following components can be added, depending on considerations described here.
Processors
CPs, ICFs, zAAPs, zIIPs, IFLs, and SAPs can be concurrently added to a zEC12 if unassigned PUs are available on any installed book. The number of zAAPs cannot exceed the number of CPs plus unassigned CPs. The same holds true for the zIIPs. Additional books can also be installed concurrently, allowing further processor upgrades. Concurrent upgrades are not supported with PUs defined as additional SAPs. If necessary, additional logical partitions can be created concurrently to use the newly added processors. The Coupling Facility Control Code (CFCC) can also configure more processors online to coupling facility logical partitions by using the CFCC image operations window.
Memory
Memory can be concurrently added up to the physical installed memory limit. Additional books can also be installed concurrently, allowing further memory upgrades by LICCC, enabling memory capacity on the new books. Using the previously defined reserved memory, z/OS operating system images, and z/VM partitions, can dynamically configure more memory online, allowing nondisruptive memory upgrades. Linux on System z supports Dynamic Storage Reconfiguration.
I/O
I/O cards can be added concurrently if all the required infrastructure (I/O slots and HCAs) is present on the configuration.
348
8049ch09.fm
I/O drawers and PCIe I/O drawers can be added concurrently without preplanning if free space is available in one of the frames and the configuration permits. Dynamic I/O configurations are supported by certain operating systems (z/OS and z/VM), allowing nondisruptive I/O upgrades. However, having dynamic I/O reconfiguration on a stand-alone coupling facility system is not possible because there is no operating system with this capability running on this system.
Cryptographic adapters
Crypto Express4S and Crypto Express3 features can be added concurrently if all the required infrastructure is in the configuration.
Processor identification
Two instructions are used to obtain processor information: Store System Information instruction (STSI) STSI reports the processor model and model capacity identifier for the base configuration and for any additional configuration changes through temporary upgrade actions. It fully supports the concurrent upgrade functions and is the preferred way to request processor information. Store CPU ID instruction (STIDP) STIDP is provided for purposes of backward compatibility.
349
8049ch09.fm
The model capacity identifier contains the base capacity, the On/Off CoD, and the CBU. The Model Permanent Capacity Identifier and the Model Permanent Capacity Rating contain the base capacity of the system, and the Model Temporary Capacity Identifier and Model Temporary Capacity Rating contain the base capacity and the On/Off CoD.
8049ch09.fm
Table 9-6 STIDP output for zEC12 Description Version code CPU identification number Machine type number Logical partition 2-digit indicator
0-7 x00a
32 - 48 x2827
48 - 63 x8000c
a. The version code for zEC12 is x00. b. The logical partition identifier is a two-digit number in the range of 00 - 3F. It is assigned by the user on the image profile through the Support Element or HMC. c. High order bit on indicates that the logical partition ID value returned in bits 8 - 15 is a two-digit value.
When issued from an operating system running as a guest under z/VM, the result depends on whether the SET CPUID command has been used, as follows: Without the use of the SET CPUID command, bits 0 - 7 are set to FF by z/VM, but the remaining bits are unchanged, which means that they are exactly as they would have been without running as a z/VM guest. If the SET CPUID command has been issued, bits 0 - 7 are set to FF by z/VM and bits 8 31 are set to the value entered in the SET CPUID command. Bits 32 - 63 are the same as they would have been without running as a z/VM guest. Table 9-7 lists the possible output returned to the issuing program for an operating system running as a guest under z/VM.
Table 9-7 z/VM guest STIDP output for zEC12 Description Version code CPU identification number Machine type number Logical partition 2-digit indicator
Bit position Without SET CPUID command With SET CPUID command
0-7 xFF
8 - 15 Logical partition ID
32 - 48 x2827
48 - 63 x8000
xFF
x2827
x8000
351
8049ch09.fm
Installation of an I/O cage is disruptive. An I/O upgrade when the operating system cannot use the dynamic I/O configuration function is disruptive. Linux, z/VSE, z/TPF, and CFCC do not support dynamic I/O configuration.
8049ch09.fm
The zEC12 can have up to eight offerings installed at the same time, with the limitation that only one of them can be an On/Off Capacity on Demand offering; the others can be any combination. The installed offerings can be activated fully or partially, and in any sequence and any combination. The offerings can be controlled manually through command interfaces on the HMC, or programmatically through a number of APIs, so that IBM applications, ISV programs, or customer-written applications, can control the usage of the offerings. Resource consumption (and thus financial exposure) can be controlled by using capacity tokens in On/Off CoD offering records. The Capacity Provisioning Manager (CPM) is an example of an application that uses the Capacity on Demand APIs to provision On/Off CoD capacity based on the requirements of the executing workload. The CPM cannot control other offerings.
9.10 References
For more information, see the following publications: IBM System z10 Enterprise Class Capacity On Demand, SG24-7504 IBM zEnterprise 196 Capacity on Demand Users Guide, SC28-2605 IBM zEnterprise EC12 Capacity on Demand Users Guide, SCxx-xxxx
353
8049ch09.fm
354
8049ch10.fm
10
Chapter 10.
RAS
This chapter describes a few of the reliability, availability, and serviceability (RAS) features of the IBM zEnterprise EC12. The zEC12 design is focused on providing higher availability by reducing planned and unplanned outages. RAS can be accomplished with improved concurrent replace, repair, and upgrade functions for processors, memory, books, and I/O. RAS also extends to the nondisruptive capability for installing Licensed Internal Code (LIC) updates. In most cases a capacity upgrade can be concurrent without a system outage. As an extension to the RAS capabilities, we discuss environmental controls implemented in the system to help reduce power consumption and cooling requirements. The design of the memory on the zEC12 is implemented based a fully redundant memory infrastructure, Redundant Array of Independent Memory (RAIM), a concept similar to the RAID design used in external disk storage systems. RAIM was first introduced with z196. The zEnterprise CPCs are the only systems in the industry offering this level of memory design. To make the delivery and transmission of microcode (LIC), fixes and restoration/backup files are digitally signed. Any data transmitted to IBM Support is encrypted. The design goal for the zEC12 has been to remove all sources of planned outages. This chapter discusses the following topics: zEC12 availability characteristics on page 356 zEC12 RAS functions on page 357 zEC12 Enhanced book availability on page 360 zEC12 Enhanced driver maintenance on page 368 RAS capability for the HMC and SE on page 369 RAS capability for zBX on page 370 Considerations for PowerHA in zBX environment on page 371 IBM System z Advanced Workload Analysis Reporter (IBM zAware) on page 373 RAS capability for Flash Express on page 374
355
8049ch10.fm
8049ch10.fm
with two independent chilled water feeds. Similarly to the radiator cooling system, one WCU and one water feed can support the entire system load. Both radiator and water cooling systems are backed-up by an air cooling system in the rare event of a cooling system problem.
Planned
The difference between scheduled outages and planned outages is, perhaps, not very obvious. The general consensus is that scheduled outages are considered to take place somewhere in the near future. The near future is considered to be an approximate 2-week time frame. Planned outages are considered outages planned well in advance and go beyond this approximate 2-week time frame. In this chapter we will not make a difference between scheduled and planned outages. Preventing unscheduled, scheduled, and planned outages has been addressed by the IBM System z system design for many years. The zEC12 introduces a fixed size HSA of 32 GB. This helps eliminate pre-planning requirements for HSA and provides flexibility to dynamically update the configuration. Performing the following tasks dynamically is possible1: Add a logical partition Add a logical channel subsystem (LCSS) Add a subchannel set Add a logical CP to a logical partition Add a cryptographic coprocessor Remove a cryptographic coprocessor Enable I/O connections Swap processor types Add memory Add a physical processor In addition, by addressing the elimination of planned outages, the following tasks are also possible: Concurrent driver upgrades Concurrent and flexible customer-initiated upgrades
1
Some pre-planning considerations may exist, see Chapter 9, System upgrades on page 311 for more information
357
8049ch10.fm
For a description of the flexible customer-initiated upgrades see 9.2.2, Customer Initiated Upgrade facility on page 319.
8049ch10.fm
advantages are improving even more, as it makes the service network itself easier to handle and more flexible. The PCIe I/O drawer is available for thezEC12. It can be installed concurrently and I/O cards can be added to the PCIe drawers concurrently.
359
8049ch10.fm
360
8049ch10.fm
The I/O connectivity must also support book removal. The majority of the paths to the I/O have redundant I/O interconnect support in the I/O infrastructure (drawers and cages) that enable connections through multiple fanout cards. If sufficient resources are not present on the remaining books, certain non-critical logical partitions might have to be deactivated, and one or more CPs, specialty engines, or storage might have to be configured offline to reach the required level of available resources. Planning that addresses these possibilities can help reduce operational errors. We suggest you do this planning well in advance. EBA function: Single-book systems cannot make use of the EBA procedure. Include the planning as part of the initial installation and any follow-on upgrade that modifies the operating environment. A customer can use the Resource Link machine profile report to determine the number of books, active PUs, memory configuration, and the channel layout. If the zEC12 is installed, you can click the Prepare for Enhanced Book Availability option in the Perform Model Conversion panel of the EBA process on the HMC. This task helps you determine the resources required to support the removal of a book with acceptable degradation to the operating system images. The EBA process determines which resources, including memory, PUs, and I/O paths will have to be freed to allow for the removal of a given book. You can run this preparation on each book to determine which resource changes are necessary; use the results as input to the planning stage to help identify critical resources. With this planning information, you can examine the logical partition configuration and workload priorities to determine how resources might be reduced and allow for the book to be removed. In planning, include the following tasks: Review of the zEC12 configuration to determine the following values: Number of books installed and the number of PUs enabled. Note the following information: Use the Resource Link machine profile or the HMC to determine the model, number and types of PUs (CPs, IFL, ICF, zAAP, and zIIP). Determine the amount of memory, both physically installed and LICCC-enabled. Work with your IBM service personnel to determine the memory card size in each book. The memory card sizes and the number of cards installed for each installed book can be viewed from the SE under the CPC configuration task list, using the view hardware configuration option.
Channel layouts, HCA to channel connections: Use the Resource Link machine profile to review channel configuration including the HCA paths. This is a normal part of the I/O connectivity planning. The alternate paths have to be separated as far into the system as possible. Review the system image configurations to determine the resources for each. Determine the importance and relative priority of each logical partition. Identify the logical partition or workloads and the actions to be taken: Deactivate the entire logical partition. Configure PUs.
361
8049ch10.fm
Reconfigure memory, which might require the use of Reconfigurable Storage Units (RSU) value. Vary off of the channels. Review the channel layout and determine whether any changes are necessary to address single paths. Develop the plan to address the requirements. When performing the review, document the resources that can be made available if the EBA is to be used. The resources on the books are allocated during a power-on reset (POR) of the system and can change during a POR. Perform a review when changes are made to the zEC12, such as adding books, CPs, memory, or channels, or when workloads are added or removed, or if the HiperDispatch feature has been enabled and disabled since the last time a POR was done.
Processor availability
Processor resource availability for reallocation or deactivation is affected by the type and quantity of resources in use, as follows: Total number of PUs that are enabled through LICCC PU definitions in the profiles that can be: Dedicated and dedicated reserved Shared Active logical partitions with dedicated resources at the time of book repair or replacement To maximize the PU availability option, ensure that there are sufficient inactive physical resources on the remaining books to complete a book removal.
Memory availability
Memory resource availability for reallocation or deactivation depends on: Physically installed memory Image profile memory allocations Amount of memory enabled through LICCC Flexible memory option See 2.7.2, Enhanced book availability on page 60.
362
8049ch10.fm
Figure 10-1 Perform Model Conversion; select Prepare for Enhanced Book Availability
After you select Prepare for Enhanced Book Availability, the Enhanced Book Availability panel opens. Select the book that is to be repaired or upgraded, then select OK. See Figure 10-2. Only one target book can be selected at a time.
363
8049ch10.fm
The system verifies the resources required for the removal, determines the required actions, and presents the results for review. Depending on the configuration, the task can take from a few seconds to several minutes. The preparation step determines the readiness of the system for the removal of the targeted book. The configured processors and the memory that is in the selected book are evaluated against unused resources available across the remaining books. The system also analyzes I/O connections associated with the removal of the targeted book for any single path I/O connectivity. If not enough resources are available, the system identifies the conflicts so you can take action to free other resources. Three states can result from the preparation step: The system is ready to perform the enhanced book availability for the targeted book with the original configuration. The system is not ready to perform the enhanced book availability because of conditions indicated from the preparation step. The system is ready to perform the enhanced book availability for the targeted book. However, to continue with the process, processors are reassigned from the original configuration. Review the results of this reassignment relative to your operation and business requirements. The reassignments can be changed on the final window that is presented. However, before making changes or approving reassignments, ensure that the changes have been reviewed and approved by the correct level of support, based on your organizations business requirements.
Preparation tabs
The results of the preparation are presented for review in a tabbed format. Each tab indicates conditions that prevent the EBA option from being performed. Tabs are for processors, memory, and various single path I/O conditions. See Figure 10-3 on page 365. Possible tab selections are: Processors Memory Single I/O Single Domain I/O Single Alternate Path I/O
364
8049ch10.fm
Only the tabs that have conditions that prevent the book from being removed are displayed. Each tab indicates what the specific conditions are and possible options to correct the conditions.
365
8049ch10.fm
Important: Consider the results of these changes relative to the operational environment. Understand the potential impact of making such operational changes. Changes to the PU assignment, although technically correct, can result in constraints for critical system images. In certain cases, the solution might be to defer the reassignments to another time that might have less impact on the production system images. After reviewing the reassignment results, and making adjustments if necessary, click OK. The final results of the reassignment, which include changes made as a result of the review, are displayed. See Figure 10-5. These results will be the assignments when the book removal phase of the EBA is completed.
366
8049ch10.fm
I/O connectivity: Alternate paths to other books must be available on the remaining active books or the I/O path must be taken offline. By understanding both the system configuration and the LPAR allocation for memory, PUs, and I/O, you can make the best decision about how to free necessary resources and allow for book removal. To concurrently replace a book, follow these steps: 1. Run the preparation task to determine the necessary resources. 2. Review the results. 3. Determine the actions to perform in order to meet the required conditions for EBA. 4. When you are ready for the book removal, free the resources that are indicated in the preparation steps. 5. Rerun the step in Figure 10-1 on page 363 (Prepare for Enhanced Book Availability task) to ensure that the required conditions are all satisfied. 6. Upon successful completion (see Figure 10-6), the system is ready for the removal of the book.
The preparation process can be run multiple times to ensure that all conditions have been met. It does not reallocate any resources. All it does is to produce a report. The resources are not reallocated until the Perform Book Removal process is invoked.
367
8049ch10.fm
368
8049ch10.fm
Concurrent crossover from Driver level N to Driver level N+1, to Driver level N+2 must be done serially; no composite moves are allowed. Disruptive upgrades are permitted at any time and allow for a composite upgrade (Driver N to Driver N+2). Concurrent back off to the previous Driver level is not possible. The driver level must move forward to driver level N+1 after EDM is initiated. Catastrophic errors during an update can result in a scheduled outage to recover. The EDM function does not completely eliminate the need for planned outages for driver-level upgrades. Upgrades might require a system level or a functional element scheduled outage to activate the new LIC. The following circumstances require a scheduled outage: Specific complex code changes might dictate a disruptive driver upgrade. You are alerted in advance so you can plan for the following changes: Design data or hardware initialization data fixes CFCC release level change OSA CHPIDs code changes may require CHPID Vary OFF/ON in order to activate new code.
SE
The zEC12 is provided with two Laptops inside the Z Frame. One is always the primary SE and the other is the alternate SE. The primary SE is the active one, while the alternate acts as the backup. Once per day, information is mirrored. For more information, see 12.1, Introduction to HMC and SE on page 398.
HMC in an ensemble
The serviceability function for the components of an ensemble is delivered through the traditional HMC/SE constructs as for earlier System z systems. From a serviceability point of view all the components of the ensemble, including the zBX, are treated as zEC12 features, similar to the treatment of I/O cards and other traditional zEC12 features.
369
8049ch10.fm
The zBX receives all of its serviceability and problem management through the HMC/SE infrastructure, and all service reporting, including call home functions are delivered in a similar fashion. The primary HMC for the ensemble is where portions of the Unified Resource Manager routines execute. The Unified Resource Manager is an active part of the ensemble and zEC12 infrastructure. Thus, the HMC is in a stateful state that needs high availability features to assure survival of the system in case of failure. Each ensemble must therefore be equipped with two HMC workstations: a primary and an alternate. While the primary HMC can do all HMC activities (including Unified Resource Manager activities), the alternate can only be the backup and cannot be used for tasks or activities. Failover: The primary HMC and its alternate must be connected to the same LAN segment to allow the alternate HMC to take over the IP address of the primary HMC during failover processing. For more information, see 12.6, HMC in an ensemble on page 422.
2 3
PS701 8406-71Y blades AIX 6.1 TL06 SP3 preferred with RSCT 3.1.0.4 (packaged in CSM PTF 1.7.1.10 installed w/ AIX 6.1.6.3) is the preferred baseline for zBX Virtual Servers running AIX.
370
8049ch10.fm
Table 10-1 PowerHA and required AIX levels IBM zBX model 003 PowerHA V5.5 AIX V5.3 AIX V6.2 AIX V7.1
PowerHA V5.5 SP8 AIX V7.1 RSCT V3.1.0.3 PowerHA V6.1 SP3 AIX V7.1 RSCT V3.1.0.3 AIX V7.1 RSCT V3.1.0.3
PowerHA V6.1
PowerHA V7.1
Not Supported
zEnterprise BladeCenter Extension (zBX) Model 003 also introduces a new version for the Advanced Management Module (AMM) and major firmware changes compared to the zBX Model 002. This takes the RAS concept of the zBX Model 003 to higher, unprecedented levels.
4 5
High Availability Cluster Multi-Processing Servers can be also virtual servers; one server = one instance of the AIX Operating System
371
8049ch10.fm
The virtual servers defined and managed in zBX use only virtual I/O resources. PowerHA can manage both physical and virtual I/O resources (virtual storage and virtual network interface cards). PowerHA can be configured to perform automated service recovery for the applications running in virtual servers deployed in zBX. PowerHA automates application failover from one virtual server in a System p blade to another virtual server in a different System p blade with similar configuration. Failover protects service (masks service interruption) in case of unplanned or planned (scheduled) service interruption. During failover, clients might experience a short service unavailability, while resources are configured by PowerHA on the new virtual server. Power HA configuration for zBX environment is similar to standard Power environments, with the particularity that it uses only virtual I/O resources. Currently, PowerHA for zBX support is limited to failover inside the same zBX. PowerHA configuration must include the following tasks: Network planning (VLAN and IP configuration definition and for server connectivity). Storage planning (shared storage must be accessible to all blades that provide resources for a PowerHA cluster). Application planning (start/stop/monitoring scripts and OS, CPU and memory resources). PowerHA SW installation and cluster configuration. Application integration (integrating storage, networking, and application scripts). PowerHA cluster testing and documentation. A typical PowerHA cluster is shown in Figure 10-7 on page 373
372
8049ch10.fm
For more information about IBM PowerHA System Mirror for AIX, see this website: http://www-03.ibm.com/systems/power/software/availability/aix/index.html
373
8049ch10.fm
Drill down to identify the cause of the problem Improve problem determination in near real time Reduce problem determination efforts significantly For detailed information on IBM zAware, see Appendix A, IBM zAware on page 429.
374
8049ch11.fm
11
Chapter 11.
Environmental requirements
This chapter briefly describes the environmental requirements for the IBM zEnterprise EC12. It lists the dimensions, weights, power and cooling requirements as an overview of what is needed to plan for the installation of a IBM zEnterprise EC12 and zEnterprise BladeCenter Extension. There are a number of options for the physical installation of the server, including air or water cooling, installation on raised floor, or non-raised floor, I/O and power cables exiting under the raised floor, or off the top of the server frames, and the possibility of having a high-voltage DC power supply as an alternative to the usual AC power supply. For comprehensive physical planning information, see Installation Manual - Physical Planning 2827, GC28-6914. In this chapter, the following topics are discussed: zEC12 power and cooling on page 376 IBM zEnterprise EC12 physical specifications on page 381 IBM zEnterprise EC12 physical planning on page 382 zBX environmental requirements on page 385 Energy management on page 390
375
8049ch11.fm
Note: Power will be lower in normal ambient temperature room and for configuration which do not have every I/O slot plugged, maximum memory and maximum configured processors. Power will also be slightly lower for DC input voltage. Actual power for any configuration, power source and room condition can be obtained using the power estimation tool (see Resource Link). Table 11-2 lists the maximum power requirements for the water cooled models.
376
8049ch11.fm
Table 11-2 Power requirements: Water cooled models Number of I/O units Power requirement kVA 0 1 2 3 4 5 6
Table 11-3 Bulk Power Regulator (BPR) requirements for books and I/O units. Note that a second pair of line cords will be installed if the number of BPR pairs is 4 or higher. If your initial configuration needs one line cord pair, but for growth would need a second pair, you can order the Line Cord Plan Ahead feature (FC 2000), which will install four line cords at the initial configuration. Also, if Balanced Power Plan Ahead (FC 3003) is ordered, four line cords are shipped and all 12 possible BPRs will be installed. Finally, if the zEC12 is configured with the Internal Battery Feature (IBF), Balanced Power Plan Ahead will automatically supply the maximum number of batteries, six IBFs, with the system.
Table 11-3 Number of BPRs requirements Number of I/O units Number of BPRs per side 0 1 2 3 4 5 6
1 2 3 4
1 3 3 4
1 3 4 5
2 3 4 5
3 3 4 5
3 4 4 5
3 4 4 5
Systems that specify two line cords can be brought up with one line cord and will continue to run. The larger machines, that is a minimum of 4 BPR pairs are installed, four line cords are installed. Four line cords offer power redundancy, so that when a line cord fails, the remaining cords deliver sufficient power to keep the system up and running.
H20
7.7
5.0
4.0
7.9
11.1
11.0
11.0
377
8049ch11.fm
Note: System holdup time given above assumes both sides functional and fresh batteries under normal room ambient conditions. Holdup times will be greater for configurations which do not have every I/O slot plugged, maximum memory and maximum configured processors. Holdup times for actual configurations are given in the power estimation tool (see Resource Link)
8049ch11.fm
379
8049ch11.fm
Allowable system inlet water temperature range is 6-20 C (43-68 F), using standard building chilled water (BCW). A special water system is typically not required. Required flow rate to the frame is 3.7 - 79.4 lpm (1- 21 gpm), depending on inlet water temperature and the number of nodes populated in the server. Colder inlet water temperatures require less flow then warmer water temperatures. Fewer nodes require less flow then maximum populated processors. Minimum water pressure required across the IBM hose ends is 0.34 - 2.32BAR (5 - 33.7 psi), depending on the minimum flow required. Maximum water pressure supplied at the IBM hose connections to the customer water supply should not exceed 6.89 BAR (100 psi). The requirements for water cooling options are indicated in detail in Installation Manual Physical Planning 2827, GC28-6914.
pl t y Su p F acili
F ac il
urn ity R et
Quick Disconnect
380
8049ch11.fm
A and Z frames with IBF (3212) and Top Exit Cabling Features (FC 7942 + FC 7901)
Weight kg (lbs) Width mm (in) Depth mm (in) Height mm (in) Height Reduction mm (in) 1568 (61.7) 1806 (71.1) 2013 (79.3) 1803 (71.0)
Water cooled servers
Weight kg (lbs) Width mm (in) Depth mm (in) Height mm (in) Height Reduction mm (in) 1568 (61.7) 1908 (75.1) 2013 (79.3) 1809 (71.2) 1847 (72.7) 1908 (75.1) 2154 (84.8) 1809 (71.2)
Notes: 1. Weight includes covers. Width, depth, and height are indicated without covers. 2. Weight is based on maximum system configuration, not the addition of the maximum weight of each frame.
381
8049ch11.fm
On raised floor, either air cooled models or water cooled models are supported.
None T op Exit features
RF Tailg ate
RF Tailgate
RF Tailga te
RF Tail gate
RF Tail gate
RF Tailgate
RF Ta lgate i
Note: No top exit feature support of water hoses for water cooled models, water hoses must go through underneath the raised floor. Non-raised floor
382
8049ch11.fm
If you want to install zEC12 server on non-raised floor environment, you can only select air cooled models and Non-Raised Floor Support feature code (FC 7998) is required. Top Exit I/O Cabling feature code (FC 7942) and Top Exit Power feature code (FC 7901) need to be ordered also. All cables should exit from the top frame of zEC12 server, no cables may exit at floor level. See Figure 11-4
383
8049ch11.fm
The customer drop must come down to the frame to meet the system input plug. These concepts are illustrated in the Figure 11-5.
Figure 11-5 Top Exit Power
cut cords plugged connections
A frame
384
8049ch11.fm
7 14 28
1 1 2
1 1 1
385
8049ch11.fm
Number of blades
Number of BladeCenters
Number of racks
42 56 70 84 98 112
3 4 5 6 7 8
2 2 3 3 4 4
A zBX can be populated by up to 112 Power 701 blades. A maximum of 28 IBM BladeCenter HX5 blades can be installed in a zBX. For DataPower blades the maximum number is 28. Note that the DataPower blade is double-wide.
1 2 3
2 4 6
386
8049ch11.fm
Number of BladeCenters
4 5 6 7 8
8 10 12 14 16
For the maximum availability, the line-cords on each side of the racks should be powered from different building power distribution units. Actual power consumption is dependent on the zBX configuration in terms of the number of BladeCenters and blades installed. Input power in kVA is equal to the out power in kW. Heat output expressed in kBTU per hour is derived by multiplying the table entries by a factor of 3.4. For 3-phase installations, phase balancing is accomplished with the power cable connectors between the BladeCenters and the PDUs.
7 14 28 42 56 70 84 98
387
8049ch11.fm
Number of blades
112
79.3
269.62
zBX rack
Heatrise
Cold air leaves the rack Computer Room Water Conditioner (CRWC)
Heatrise
CL
Figure 11-7 Rear Door Heat eXchanger (left) and functional diagram
The IBM Rear Door Heat eXchanger also offers a convenient way to handle hazardous hot spots, which can help you lower the total energy cost of the data center.
388
8049ch11.fm
zBX weight
Table 11-10 shows the maximum weights of fully populated zBX racks and BladeCenters.
Table 11-10 Weights of zBX racks Rack description Weight kg (lbs.)
Rack weight: A fully configured Rack B is heavier than a fully configured Rack C, D, or E because Rack B has the TOR switches installed.
For a complete view of the physical requirements, see zBX Installation Manual for Physical Planning 2458-003, GC27-2619.
389
8049ch11.fm
( e .g. Ca pa c ity P rov is ioning M a a ge r) n (e .g. Ca pa c ity P rov is ionin g M a na ge r) ( e .g. Ca pa c ity P rov is ioning M a a ge r) n
GUI
API
Energy M anagement HMC/SE Energy M oring Dashboard onit Hard ware Man agemen t Console
GUI
Energy Management
EC12
Supp ort Element
power code power and ambi ent temperat ure snapshots read ambient temp read V, I set V set f Power Su b-system Con l ler tro System z server
AMM
Bl adeCenter
The hardware components in the zEC12 and the optional zBX are monitored and managed by the Energy Management component in the Support Element (SE) and HMC. The GUIs of the SE and HMC provide views, for instance with the System Activity Display or Monitors Dashboard. Through an SNMP API energy information is available to, for instance Active Energy Manager, a plug-in of IBM Systems Director, see 11.5.4, IBM Systems Director Active Energy Manager on page 392 for more information. When Unified Resource Manager features are installed (see 12.6.1, Unified Resource Manager on page 422), several monitoring and control functions can be used to perform Energy Management. More details are discussed in section 11.5.5, Unified Resource Manager: Energy Management on page 393. A few aids are available to plan and monitor the power consumption and heat dissipation of the zEC12. This section summarizes the tools that are available to plan and monitor the energy consumption of the zEC12. Power estimation tool Query maximum potential power System activity display and Monitors Dashboard IBM Systems Director Active Energy Manager
390
8049ch11.fm
Label Power
Wasted Power / Power budget not converted into Cooling compute cycles Allocation
Power (watts)
Time
Trending (weeks, months)
391
8049ch11.fm
The Monitors Dashboard of the HMC allows you to display power and other environmental data and also allows you to start a Dashboard Histogram Display, where you can trend a particular value of interest, such as the power consumption of a blade or the ambient temperature of the zEC12.
392
8049ch11.fm
Active Energy Manger is a management software tool that can provide a single view of the actual power usage across multiple platforms as opposed to the benchmarked or rated power consumption. It can effectively monitor and control power in the data center at the system, chassis, or rack level. By enabling these power management technologies, data center managers can more effectively power manage their systems while lowering the cost of computing. The following data is available through Active Energy Manager: System name, machine type, model, serial number and firmware level of System z servers and optional zBX attached to IBM zEnterprise Systems or IBM zEnterprise EC12. Ambient temperature Exhaust temperature Average power usage Peak power usage Limited status and configuration information. This information helps explain changes to the power consumption, called Events, which can be: Changes in fan speed Radiator/WCU failures Changes between power off, power on, and IML complete states Number of books and I/O cages CBU records expiration(s)
IBM Systems Director Active Energy Manager provides customers with the intelligence necessary to effectively manage power consumption in the data center. Active Energy Manager, which is an extension to IBM Director systems management software, enables you to meter actual power usage and trend data for any single physical system or group of systems. Active Energy Manager uses monitoring circuitry, developed by IBM, to help identify how much actual power is being used and the temperature of the system.
Choice of suites
The energy management capabilities for Unified Resource Manager that can be used in an ensemble depends on which suite is installed in the ensemble: Manage suite (feature code 0019) Automate/Advanced Management suite (feature code 0020)
Manage suite
For energy management the manage suite focuses on the monitoring capabilities. Energy monitoring can help better understand the power and cooling demands of the zEC12 system, by providing complete monitoring and trending capabilities for the zEC12 and the zBX using one or more of the following options: Monitor dashboard Environmental Efficiency Statistics Details view
393
8049ch11.fm
394
8049ch11.fm
More information about Energy Management with Unified Resource Manager is available in the Redbooks publication, IBM zEnterprise Unified Resource Manager, SG24-7921.
395
8049ch11.fm
396
8049ch12.fm
12
Chapter 12.
397
8049ch12.fm
398
8049ch12.fm
Table 12-1 zEC12 HMC - System z support summary System z family name Machine type Minimum SE driver Minimum SE version
12 93 86 79 79 67 67 55 55
For the IBM zAware management, tasks and panels are implemented to configure and support it as the example shows in the following Figure 12-2 on page 400:
399
8049ch12.fm
For details see Appendix A, IBM zAware on page 429 and IBM zAware Concept Guide , SG24-8070. STP NTP broadband security Authentication is added to the HMCs NTP communication with NTP time servers. Refer to HMC NTP broadband authentication support for zEC12 on page 415 for details. The environmental task has usability improvements regarding the time frame. See Environmental Efficiency Statistic Task on page 412. Crypto Function Integration in Monitors Dashboard See The Monitors Dashboard task on page 410. Removal of modem support from the HMC This change impacts customers that may have set up the modem for Remote Support Facility (RSF) or for STP NTP access. For details see 12.3, Remote Support Facility on page 405 and 12.5.8, NTP client/server support on HMC on page 414 Install and activate by MCL bundle target. Microcode installation by MCL bundle target on page 409 provides further informations. A confirmation panel before processing an Alt-Ctrl-Delete request is added. Note: If an HMC needs to be rebooted, always use the Shutdown and Restart task on the HMC to avoid any file corruption. The capability to modify the time of the SE mirror scheduled operation is added. It is now possible to allow mass delete of messages from the Operating System Messages task. The Network Settings task is updated to clearly show the ordering of the routing table entries. 400
IBM zEnterprise EC12 Technical Guide
8049ch12.fm
Remove support for the Coprocessor Group Counter Sets In zEC12 each physical processor has its own crypto coprocessor, and no longer has to share this coprocessor with another PU. The Coprocessor Group Counter Sets of the counter facilities will not be available. All of the necessary crypto counter information will be available in the crypto activity counter sets directly. The check-box selection for the Coprocessor Group Counter Sets was removed from the Image profile definition and the Change Logical Partition Security task.
12.1.3 Tree Style User Interface and Classic Style User Interface
Two User Interface styles are provided with an HMC. The Tree Style User Interface (default) uses a hierarchical model popular in newer operating systems and features context-based task launching. The Classic Style User Interface uses the drag and drop interface style. Tutorials: The IBM Resource Linka provides tutorials which demonstrate how to change from the Classic to the Tree Style Interface, and will introduce you to the function of the Tree Style Interface on the HMC: https://www-304.ibm.com/servers/resourcelink/hom03010.nsf/pages/education?Op enDocument
a. Registration is required to access the IBM Resource Link
401
8049ch12.fm
HMC
HMC HMC Customer Network 1
HMC
H MC
HMC
HMC Customer Network 2 J0 1 J05 J05 J02 J01 J02
SE
SE
SE
Various methods are available for setting up the network. It is the customer responsibility to plan and conceive the HMC and SE connectivity. The method that should be selected depends on the customer connectivity and security requirements. Security: Configuration of network components, such as routers or firewall rules, is beyond the scope of this document. Any time networks are interconnected, security exposures can exist. Network security is a clients responsibility. The document IBM System z HMC Security provides information about HMC security. It is available on the IBM Resource Linka: https://www-304.ibm.com/servers/resourcelink/lib03011.nsf/pages/zHmcSecurity/$f ile/zHMCSecurity.pdf The Hardware Management Console Operations Guide (V2.12.0), SC28-6919 covers the possibilities to set on the HMC regarding access and security.
a. Registration is required to access the IBM Resource Link.
The HMC and SE network connectivity should be planned carefully to allow for current and future use. Many of the System z capabilities benefit from the various network connectivity options available. For example, functions available to the HMC that depend on the HMC connectivity include: LDAP support, that can be used for HMC user authentication NTP client/server support Remote Support Facility (RSF) via broadband HMC remote web browser
402
8049ch12.fm
Enablement of the SNMP and CIM APIs to support automation or management applications such as IBM System Director Active Energy Manager (AEM) The above examples are shown in Figure 12-4:
More information can be found in the following documentation: Hardware Management Console Operations Guide (V2.12.0), SC28-6919 11.5.4, IBM Systems Director Active Energy Manager on page 392 Installation Manual - Physical Planning 2827, GC28-6914
403
8049ch12.fm
404
8049ch12.fm
For remote operations using a web browser, if an IPv6 address is assigned to the HMC, navigate to it by specifying that address. The address must be surrounded with square brackets in the browsers address field: https://[fdab:1b89:fc07:1:201:6cff:fe72:ba7c] Using link-local addresses must be supported by browsers.
405
8049ch12.fm
SE access
It is not necessary to be physically close to an SE in order to use it. The HMC can be used to access the SE remotely via the Single Object Operations (SOO). The same interface as in the SE is provided. For details please consult the Hardware Management Console Operations Guide (V2.12.0), SC28-6919.
406
8049ch12.fm
The Hardware messages task displays hardware-related messages at CPC level, at logical partition level, SE level, or hardware messages related to the HMC itself.
407
8049ch12.fm
MCLs should be reviewed continuously and customers should decide whether to wait for the next scheduled apply session, or schedule one earlier if their risk assessment of hiper MCLs warrants. Note: The following link in IBM Resource Linka provides access to the machine information for your System z according the scheduled sent system availability data. There you can find further information about the MCL status of your zEC12: https://www-304.ibm.com/servers/resourcelink/svc0303a.nsf/fwebsearchstart?openf orm
a. Registration is required to access the IBM Resource Link.
Microcode terms
The driver contains EC streams. Each EC stream covers the code for a specific component from the zEC12. It has a specific name and an ascending number. The EC stream name and a specific number is one Microcode Change Level (MCL). MCLs from the same EC stream must be installed in sequence. MCLs can have installation dependencies with other MCLs. Combined MCLs from one or more EC streams are in one bundle. A MCL contains one or more Microcode Fixes (MCF). Figure 12-5 shows how the driver, bundle, EC stream, MCL, and MCF interact with each other.
408
8049ch12.fm
12.5.5 Monitoring
This section discusses monitoring considerations.
409
8049ch12.fm
410
8049ch12.fm
With zEC12 the monitor dashboard has been enhanced with an adapters table. The crypto utilization percentage is displayed on the monitors dashboard according to the PCHID number. The associated crypto number (Adjunct Processor Number) for this PCHID is also shown in the table. It provides information about utilization rate in a system wide basis, not per logical partition, as shown in Figure 12-9.
Also for Flash Express there is a new panel, shown in Figure 12-10 on page 412.
411
8049ch12.fm
8049ch12.fm
activate and deactivate a temporary upgrade. The task help managing all installed or staged LICCC records by showing a list of them. It also shows a history of recorded activities. HMC for IBM zEnterprise EC12 CoD capabilities include: SNMP API support: API interfaces for granular activation and deactivation API interfaces for enhanced Capacity On Demand query information API Event notification for any Capacity On Demand change activity on the system Capacity On Demand API interfaces (such as On/Off CoD and CBU) SE panel features (accessed through HMC Single Object Operations): Panel controls for granular activation and deactivation History panel for all Capacity On Demand actions Descriptions editing of Capacity On Demand records HMC/SE version 2.12.0 provides CoD information such as: MSU and processor tokens shown on panels Last activation time shown on panels Pending resources are shown by processor type instead of just a total count Option to show details of installed and staged permanent records More details for the Attention! state on panels (by providing seven additional flags) New in SE version 2.12.0: Some panels preselected defaults are removed. Specifying each selection in the panel is required. HMC and SE are an integral part of the z/OS Capacity Provisioning environment. The Capacity Provisioning Manager (CPM) communicates with the HMC through System z APIs and enters CoD requests. For this reason, SNMP must be configured and enabled on the HMC. For additional information about using and setting up CPM, see the publications: z/OS MVS Capacity Provisioning Users Guide, SC33-8299 zEnterprise System Capacity on Demand Users Guide, SC28-2605
413
8049ch12.fm
z9 EC or earlier - CF & OS
EC12
Not Supported
Not Supported
IMPORTANT
z2120 can be in sam e STP CTN as z196, z114 and z10 servers but NOT in a STP CTN with z9 servers
z9 BC or earlier - CF and OS
(Sysplex and STP)
In an STP-only CTN, the HMC can be used to perform the following tasks: Initialize or modify the CTN ID. Initialize the time, manually or by contacting an NTP server. Initialize the time zone offset, daylight saving time offset, and leap second offset. Assign the roles of preferred, backup, and current time servers, as well as arbiter. Adjust time by up to plus or minus 60 seconds. Schedule changes to the offsets listed. STP can automatically schedule daylight saving time, based on the selected time zone. Monitor the status of the CTN. Monitor the status of the coupling links initialized for STP message exchanges. For diagnostic purposes the Pulse per Second port state on a zEC12 can be displayed and fenced ports can be reset individually.
8049ch12.fm
Note: External Time Source (ETS) connection through modem is not supported on the zEC12 HMC. This capability addresses the following requirements: Customers who want time accuracy for the STP-only CTN Customers using a common time reference across heterogeneous platforms NTP server becomes the single time source, ETS for STP, as well as other servers that are not System z (such as AIX, Windows, and others) that have NTP clients. The HMC can act as an NTP server. With this support, the zEC12 can get time from the HMC without accessing other than the HMC/SE network. When the HMC is used as an NTP server, it can be configured to get the NTP source from the Internet. For this type of configuration, a LAN separate from the HMC/SE LAN can be used.
415
8049ch12.fm
Symmetric key (NTP V3-V4) authentication Symmetric key authentication as described in RFC-1305 (made available in NTP Version 3). Symmetric key encryption uses same key for both encryption and decryption. Users exchanging data keep this key to themselves. Messages encrypted with a secret key can be only decrypted with the same secret key. Symmetric does support network address translation (NAT). Symmetric key autokey (NTP V4) authentication Autokey which uses public key cryptography as described in RFC-5906 (made available in NTP Version 4). The key generation for the HMC NTP is performed by clicking the Generate Local Host Key button on the Autokey Configuration panel. Pressing this button will issue the ntp-keygen command to generate the specific key and certificate for this system. Autokey authentication is not available with NAT firewall. Issue NTP commands NTP Command support is also added to display the status of remote NTP servers and the current NTP server (HMC). For additional planning and setup information for STP and NTP, see the following publications: Server Time Protocol Planning Guide, SG24-7280 Server Time Protocol Implementation Guide, SG24-7281 Server Time Protocol Recovery Guide, SG24-7380
8049ch12.fm
Time-of-Day (TOD) clock every hour and allows the SE's clock to maintain a time accuracy of 100 milliseconds to an NTP server configured as the External Time Source in an STP-only CTN. This is shown on Figure 12-14.
417
8049ch12.fm
418
8049ch12.fm
The System Input/Output Configuration Analyzer is a view-only tool. It does not offer any options other than viewing options. With the tool, data is formatted and displayed in five different views, various sort options are available, and data can be exported to a USB flash memory drive (UFD) for a later viewing. The following five views are available: PCHID Control Unit View, which shows PCHIDs, CSS, CHPIDs and their control units. PCHID Partition View, which shows PCHIDS, CSS, CHPIDs and the partitions they are in. Control Unit View, which shows the control units, their PCHIDs, and their link addresses in each CSS. Link Load View, which shows the Link address and the PCHIDs that use it. Node ID View, which shows the Node ID data under the PCHIDs.
Cryptographic hardware
The IBM zEnterprise EC12 includes both standard cryptographic hardware and optional cryptographic features for flexibility and growth capability. The HMC/SE interface provides the following capabilities: Define the cryptographic controls Dynamically add a Crypto feature to a partition for the first time Dynamically add a Crypto feature to a partition already using Crypto Dynamically remove Crypto feature from a partition The Crypto Express4S, a new Peripheral Component Interconnect Express (PCIe) Cryptographic Coprocessor, was introduced. It is an optional and zEC12 exclusive feature. This provides a secure programming and hardware environment wherein crypto processes are performed. Each Crypto Express4S adapter can be configured by the installation as a Secure IBM CCA coprocessor, as a Secure IBM Enterprise PKCS #11 (EP11) coprocessor or as an accelerator.
419
8049ch12.fm
When EP11 mode is selected, a unique Enterprise PKCS #11 firmware is loaded into de cryptographic coprocessor. It is separate from the CCA firmware currently loaded when CCA coprocessor is selected. CCA firmware and PKCS #11 firmware cannot coexist at same time in a card. Trusted Key Entry (TKE) Workstation with smart card reader feature is required to support the administration of the Crypto Express4S when configured as an Enterprise PKCS #11 coprocessor. Crypto Express3 is also available in a carry forward only basis when upgrading from earlier generations tozEC12. In order to support the new Crypto Express4S card, the Cryptographic Configuration panel was changed to support the card modes as follows: Accelerator mode (CEX4A) CCA Coprocessor mode (CEX4C) PKCS #11 Coprocessor mode (CEX4P) Other Cryptographic Configuration panel updates include: Support for a Customer Initiated Selftest (CIS) for Crypto running EP11 Coprocessor mode. TKE commands always permitted for EP11 mode. The Test RN Generator function was modified/generalized to also support CIS, depending on the mode of the crypto card. Crypto Details panel changed to display the crypto part number. Support is now provided for up to 4 UDX files. Only UDX CCA is currently supported for zEC12. UDX import now only supports importing from DVD-ROM The next Figure 12-15 shows an example of the Cryptographic Configuration.
A Usage Domain Zeroize task is provided to clear the appropriate partition crypto keys for a given usage domain when removing a crypto card from a partition. Crypto Express4S in EP11 mode will be configured to the standby state after Zeroize.
420
8049ch12.fm
For detailed set-up information, see IBM zEnterprise EC12 Configuration Setup, SG24-8034.
421
8049ch12.fm
Hy pe rv is or M an age m ent Hyp ervi sors (except z/VM) shipped and serviced as firmware. Integrate d depl oyment and config uration of hypervisors Mana ge and control communication betw en virtu al server operating e systems and the hypervisor.
Ene rgy M a nage m ent Moni toring and trend reporting of energy efficiency. Ab ility to query maximum pote ntial power. Po wer saving Po wer capping
Hype rvisors Ope r ationa l C ontr ols HMC provi des a si ngle con sol idate d and consiste nt vi ew of resources Auto-discove ry and configuration supp ort for ne w resource s. Cross platform hardware probl em detection, reporti ng and call home. Physical hardware configurati on, backup and re store. Delivery of system acti vi ty u sin g new use r.
Energy
Operations
Performance
Network s
Virtual Serve rs
Work load A war e nes s a nd Platf orm Pe rfor m anc e Ma nage m ent Wizard-driven man agemen t of resources in acco rda nce with specified business servi ce level ob jectives HMC p rovid es a sin gle consoli dated and consistent view of resources Monitor resource use withi n the context of a busine ss worklo ad Define workloads and associated performance po licies
Ne twor k M ana ge me nt Creation of vitual networks Man ageme nt of vi rtual networks including access control
Key
Manag e suite
Automate suite
Figure 12-16 Unified Resource Manager functions and suites
V ir tua l S er ve r Life c yc le M ana ge me nt Sin gle vie w of virtu alization across platforms. Abi lity to deplo y mul tip le, cro ss-platform virtua l servers within mi nutes Manag ement of virtual networks includ ing access contro l
Overview
Unified Resource Manager provides the following functions:
422
8049ch12.fm
Hypervisor Management: Provides tasks for managing hypervisor life-cycle, managing storage resources, performing RAS (reliability, availability, and serviceability) and FFDC (First Failure Data Capture) features and monitoring the supported hypervisors. Ensemble membership management: Provides tasks for creating an ensemble and controlling membership of the ensemble. Storage Management: Provides a common user interface for allocation and deallocation of physical and virtual storage resources for an ensemble. Virtual server management: Provides life-cycle management to create, delete, activate, deactivate, and modify definitions of virtual servers. Virtual network management: Provides for the management of networking resources for an ensemble. Performance management: Provides a global performance view of all the virtual servers supporting workloads deployed in an ensemble. The virtual server workload performance goal is like a simplified z/OS WLM policy: You can define, monitor, report, and manage the performance of virtual servers based on workload performance policies. Policies are associated to the workload: From the overall Workload performance health report, you can review contributions of individual virtual servers. You can manage resources across virtual servers within a hypervisor instance.
Energy management: Monitors energy usage and control power-saving settings, accessed through the new monitors dashboard. Monitoring of virtual server resources for CPU use and delays, with capability of creating a graphical trend report Unified Resource Manager supports different levels of system management. A feature determines these management functions and operational controls that are available for a zEnterprise mainframe and any attached zBX. These named features are Manage and Automate/Advanced Management: Manage suite Provides Unified Resource Managers function for core operational controls, installation, and energy monitoring. It is configured by default and activated when an ensemble is created. workload definition and performance policy monitoring and reporting. The Automate functionality adds, on top of the Advanced Management functionality, goal oriented resource monitoring management and energy management for CPC components, IBM Smart Analytic Optimizer, POWER7 Blade, and DataPower XI50z.
Automate/Advanced Advanced Management functionality for IBM System x Blades delivers Management suite
423
8049ch12.fm
Table 12-2 lists the feature codes that must be ordered for enabling Unified Resource Manager. To get ensemble membership, make sure that you also order FC 0025 for zEC12.
Table 12-2 Unified Resource Manager feature codes and charge indicator. Unified Resource Manager managed component Managea - per connection Advanced Managementa - per connection Automatea - per connection
base features IFL POWER7 Blade DataPower Blade IBM System x Blades
a. Yes = charged feature, N/C = no charge, N/A = not applicable. All components are either managed through the
Manage suite, or the Automate/Advanced Management suite. The Automate/Advanced Management suite contains the functionality of the Managed suite. b. Feature code 0019 is a prerequisite for feature codes 0020, 0039, 0040, 0041, and 0042. c. Feature code 0020 is a prerequisite for feature codes 0043, 0044, 0045, 0046, and 0052.
Feature code
Feature code 0025 (Ensemble Membership Flag) is associated with an HMC when a zEC12 is ordered. This feature code is required on the controlling zEC12 to be able to attach a zBX. A new task called Create Ensemble will allow the Access Administrator to create an ensemble that contains CPCs, Images, workloads, virtual networks and storage pools, either with or without an optional zBX. If a zEC12 has been entered into an ensemble, then the CPC details task on the SE and HMC will reflect the ensemble name.
424
8049ch12.fm
Unified Resource Manager actions for the ensemble are conducted from a single primary HMC. All other HMCs connected to the ensemble will be able to perform system management tasks (but not ensemble management tasks) for any CPC within the ensemble. The primary HMC can also be used to perform system management tasks on CPCs that are not part of the ensemble, such as Load, Activate, and so on. The ensemble-specific managed objects include: Ensemble Members Blades BladeCenters Hypervisors Storage Resources Virtual Servers Workloads When another HMC accesses an ensemble nodes CPC, the HMC can do the same tasks as if the CPC were not a part of an ensemble. A few of those tasks have been extended to allow you to configure certain ensemble-specific properties (such as setting the virtual network associated with OSAs for a LPAR). Showing ensemble-related data in certain tasks is allowed. Generally, if the data affects the operation of the ensemble, then the data is read-only on another HMC. The tasks that show ensemble-related data on another HMC are as follows: Scheduled operations: Displays ensemble introduced scheduled operations, but you can only view these scheduled operations. User role: Shows ensemble tasks and you can modify and delete those roles. Event monitoring: Displays ensemble-related events, but you cannot change or delete the event.
8049ch12.fm
The alternate HMCs role is to mirror ensemble configuration and policy information from the primary HMC. When failover happens, the alternate HMC will become the primary HMC. This behavior is the same as the current primary and alternate Support Elements.
426
8049ch12.fm
Primary
Client Network
A lternate
1 HMC
O S M
O S M
EC12 CPC2
2
1. HMC to SE Network 2. INMN (EC12) 3. INMN (zBX) 4. IEDN 5. FC Disk Storage (zBX)
OSX
zBX
Fibre Channel Disk Storage
5
Fibre Channel Switch
For the stand-alone CPC ensemble node (CPC2), two OSA-Express4S 1000BASE-T ports (CHPID type OSM) connect to the Bulk Power Hubs (port J07) with 3.2 meter Category 6 Ethernet cables. The HMCs also communicate with all the components of the ensemble by the BPHs in the CPC. The OSA-Express4S 10 GbE ports (CHPID type OSX) are plugged with client provided 10 GbE cables (either SR or LR, depending on the OSA feature). Details for zBX can be found in Chapter 7, zBX zEnterprise BladeCenter Extension Model 003 on page 211.
427
8049ch12.fm
428
8049ax01.fm
Appendix A.
IBM zAware
Instead of trying to produce a programme to simulate the adult mind, why not rather try to produce one which simulates the child's? If this were then subjected to an appropriate course of education one would obtain the adult brain.
Computing Machinery and Intelligence by Alan Turing
In this appendix we introduce IBM System z Advanced Workload Analysis Reporter (IBM zAware), the next generation of system monitoring. It is a new feature that is designed to offer a near real-time, continuous learning, diagnostics and monitoring capability. IBM zAware helps you pinpoint and resolve potential problems quickly enough to minimize impacts to your business. In this appendix, the following topics are discussed: A.1, Troubleshooting in Complex IT Environments on page 430 A.2, Introducing the IBM zAware on page 431 A.3, IBM zAware Technology on page 432 A.4, IBM zAware PreRequisites on page 438 A.5, Configuring and using the IBM zAware virtual appliance on page 439 For comprehensive information about IBM zAware, see IBM zAware Concept Guide , SG24-8070 and Advanced Workload Analysis Reporter (IBM zAware), SC27-2623.
429
8049ax01.fm
430
8049ax01.fm
Checks configurations Programmatic, applies to IBM and ISV tools Can escalate notifications
Yes
Rules based
431
8049ax01.fm
Solutions Available
Rules based
Self Learning
Method
Trending analysis of z/OS system resources, and performance Can invoke z/OS RTD Real time diagnostics of specific z/OS system issues Pattern based message analysis Self learning Provides aid in diagnosing complex z/OS problems, including cross sysplex Yes
Yes
Yes
Diagnosis
a. Included in z/OS
Any large and complex z/OS installation with mission-critical applications and middleware, is suggested to exploit the IBM zAware along with z/OS included problem diagnosis solutions.
432
8049ax01.fm
Vie w IB M zA wa re re sults
z /OSM F
me sto Cu
o rk e tw rn
z/OS
I BM zAware Host Parti ti on
z/OS
z/OS
z/OS
z/OS
Pe rsiste nt Stora ge
File Syste m
T IP
n un
el
IBM zAware runs in a logical partition as firmware. IBM zAware; requires the zEC12 configuration to have a priced feature code. needs processor, memory, disk, and network resources to be assigned to the LPAR it runs Shows similarity with Coupling Facility in many ways is updated like all other firmware, with a separate Engineering Change stream is loaded from the Support Element Hard Disk demonstrates out of band monitoring with minimal effect on z/OS product workloads Figure A-2 on page 434 shows IBM zAware Image Profile on HMC
433
8049ax01.fm
IBM zAware analyzes massive amounts of OPERLOG messages (including all z/OS console messages, ISV and application generated messages) to build Sysplex and detailed views in the IBM zAware GUI. Figure A-3 on page 435 and Figure A-4 on page 436 show samples for both Sysplex and Detailed views consecutively.
434
8049ax01.fm
Figure A-3 IBM zAware Sysplex view; showing all connected managed z/OS clients
435
8049ax01.fm
Figure A-4 IBM zAware detailed view; drilled down to a single z/OS image
The analytics creates a statistical model of the normal message traffic generated by each individual z/OS. Using this model which is stored in a database, unexpected messages and patterns of messages are identified. Using a sliding ten minute interval which is updated every two minutes, a current score for the interval is created based on how unusual the message traffic is. Stable system requires lower interval score to be marked as interesting or rare Unstable system requires larger interval score to be marked as interesting or rare For each interval IBM zAware provides details of all of the unique and unusual message IDs found within interval including how many, how rare, how much they contributed to the intervals score (anomaly score, interval contribution score, rarity score, appearance count), when they first appeared. IBM zAware also handles burst of messages and identifies; if it is one component emitting unusual message IDs, if the message is a critical z/OS kernel message, if the messages are related to changes like new software levels (operating system, middleware, applications) or updated system settings/configurations. The choice of unique message IDs is embedded in the domain knowledge of IBM zAware. IBM zAware detects things typical monitoring systems miss due to Message suppression (message too common) - useful for long-term health issues 436
8049ax01.fm
Uniqueness (message not common enough) - useful for real-time event diagnostics IBM zAware decides on the color of an interval using distribution of interval score: Normal: Blue color - interval score between 1- 99.5 Interesting: Orange color - interval score between 99.5 - 100 Rare: Red color - interval score 101
Training Period
The IBM zAware server starts receiving current data from the z/OS system logger running on z/OS monitored clients. However, the server cannot use this data for analysis until a model of normal system behavior exists. The estimated amount of data for building the most accurate models is 90 days of data for each client. By default, training automatically runs every 30 days. Your installation can modify the number of days required for this training period, based on your knowledge of the workloads running on z/OS monitored clients. This training period applies for all monitored clients; different training period for each client cannot be defined.
437
8049ax01.fm
For IBM messages; IBM zAware GUI has a link to the message description which often includes a recommended action to correct the issue highlighted by the message. There also is a z/OSMF link on the navigation bar.
IBM zAware is complementary to your existing tools Compared to existing tools, IBM zAware works out of the box with relatively little
customization. It does not depend on other solutions or manual coding of rules and is always enabled to watch your system. The XML output built by IBM zAware can be queued up to existing system monitoring tools like Tivoli via published API.
8049ax01.fm
FC 0101 is a priced feature code that represents the quantity (in decimal) of client/host connections. IBM zAware must be ordered from the host to select both host and client connections, so there is no need to order FC 0101 for client machines. Minimum quantity available would be the quantity of Processor Books present on the host (ordering) machine. For instance, if thezEC12 to host IBM zAware is a model H49, minimum quantity will be 2. Maximum quantity that can be ordered as FC 0101 is 99. A Disaster Recovery option is also available and will represent that IBM zAware is installed on a Disaster Recovery zEC12 server. In this case, zero priced feature FC 0102 will be assigned to represent the quantity of connections (in decimal). Note: zEC12 resource requirements are dependent on the number of monitored clients, amount of message traffic, length of time data retained. Processors General Purpose CP or IFL that can be shared with other LPARs in zEC12. Usage estimates between a partial engine to two engines depending on the size of the configuration Memory Minimum 4GB initial memory + 200MB for each z/OS LPAR being monitored Flash Express is not supported Direct Access Storage Devices 250-300 GB persistent DASD storage Only ECKD format, FCP devices are not supported IBM zAware manages its own data store and uses Logical Volume Manager (LVM) to aggregate multiple physical devices into a single logical device Network (for both instrumentation data gathering and outbound alerting/communications) HiperSockets for the z/OS LPARs running on the same zEC12 as the IBM zAware LPAR OSA ports for the z/OS LPARs running on a different CPC than where IBM zAware LPAR runs z/OS LPARs running on other System z servers (IBM zEnterprise 196, IBM System z10 Business Class, etc.) can be monitored clients by sending log files via TCP/IP network to host zEC12 server as long as they fulfill z/OS requirements
z/OS Requirements:
z/OS V1.13 with additional PTFs 90 days historical SYSLOG or formatted OPERLOG data to prime IBM zAware
439
8049ax01.fm
Verify that your installation meets the prerequisites for using the IBM zAware virtual appliance Configure network connections for the IBM zAware partition through the Hardware Configuration Definition (HCD) or the Input/Output Configuration Program (IOCP). Configure persistent storage for the IBM zAware partition through the HCD or IOCP. Define the LPAR characteristics of the IBM zAware partition through the Hardware Management Console (HMC). Define network settings for the IBM zAware partition through the HMC. Activate the IBM zAware partition through the HMC. Phase 3: Configuring the IBM zAware server and its monitored clients: Assign storage devices for the IBM zAware server through the IBM zAware GUI. (Optional) Replace the self-signed Certificate Authority (CA) certificate that is configured in the IBM zAware server. (Optional) Configure an LDAP directory or local file-based repository for authenticating users of the IBM zAware GUI. (Optional) Authorize users or groups to access the IBM zAware GUI. (Optional) Modify the configuration values that control IBM zAware analytics operation. Configure a network connection for each z/OS monitored client through the TCP/IP profile. If necessary, update firewall settings. Verify that each z/OS system meets the sysplex configuration and OPERLOG requirements for IBM zAware virtual appliance monitored clients. Configure the z/OS system logger to send data to the IBM zAware virtual appliance server. Prime the IBM zAware server with prior data from monitored clients. Build a model of normal system behavior for each monitored client. The IBM zAware server uses these models for analysis.
440
8049ax02.fm
Appendix B.
Channel options
The following two tables describe all channel attributes, the required cable types, the maximum unrepeated distance, and the bit rate for the zEC12. For all optical links the connector type is LC Duplex, except the 12xIFB connection is established with an MPO connector. The electrical Ethernet cable for the OSA connectivity are connected through an RJ45 jack. Table B-1 lists the attributes of the channel options supported on zEC12.
Statement of Direction: The zEC12 is the last IBM System z server to support ISC-3 Links. The zEC12 is the last IBM System z server to support Ethernet half-duplex operation and 10 Mbps link data rate on 1000BASE-T Ethernet features. The FICON Express8S features, supporting a 2, 4, or 8 Gbps link data rate, are planned to be the last FICON features to support a 2 Gbps link data rate.
Table B-1 zEC12 channel feature support Channel feature Feature codes Bit rate Cable type in Gbps (or stated) Maximum unrepeated distancea Ordering information
3321 3325
Carry forward Carry forward New build Carry forward Carry forward New build
FICON Express8S 10KM LX 0409 FICON Express4 SX FICON Express8 SX FICON Express8S SX 3322 3326 0410
441
8049ax02.fm
Channel feature
Feature codes
Ordering information
SM 9 m MCP 50 m
5 km 550 m (500)
OSA-Express3 1000BASE-T OSA-Express4 1000BASE-T OSA-Express3 10 GbE LR OSA-Express4S 10 GbE LR OSA-Express3 10 GbE SR
Carry forward New build Carry forward New build Carry forward New build
Parallel Sysplex
IC ISC-3 (peer mode) ISC-3 (RPQ 8P2197 Peer mode at 1 Gbps)b HCA2-O (12x IFB) HCA2-O LR (1x IFB) HCA3-O (12x IFB) HCA3-O LR (1x IFB)
Cryptography
N/A 2 0217 0218 0219 0163 0168 0171 0170 1 6 GBps 2.5 or 5 Gbps 6 GBps 2.5 or 5 Gbps
N/A Carry forward Carry forward Carry forward Carry forward New build New build
0864 0865
N/A N/A
N/A N/A
N/A N/A
a. were applicable, Minimum fiber bandwidth distance product in MHzkm for multi-mode fiber optic links are included in parentheses. b. RPQ 8P2197 enables the ordering of a daughter card supporting 20 km unrepeated distance for 1 Gbps peer mode. RPQ 8P2262 is a requirement for that option, and other than the normal mode, the channel increment is two (that is, both ports (FC 0219) at the card must be activated).
Table B-2 Maximum unrepeated distance for FICON SX features Cable type\bit rate 1 Gbps 2 Gbps 4 Gbps 8 Gbps
21 meters 69 feet
442
8049ax02.fm
1 Gbps
2 Gbps
4 Gbps
8 Gbps
443
8049ax02.fm
444
Appendix C.
Flash Express
This appendix introduces the IBM Flash Express feature available on the zEC12 server. Flash memory is a non-volatile computer storage technology. It was introduced on the market some decades ago. Flash memory is commonly used today in memory cards, USB Flash drives, Solid State Drives (SSDs), and similar products for general storage and transfer of data. Until recently, the high cost per gigabyte and limited capacity of SSDs restricted deployment of these drives to specific applications. Recent advances in SSD technology and economies of scale have driven down the cost of SSDs, making them a viable storage option for I/O intensive enterprise applications. A Solid State Drive (SSD), sometimes called a solid-state disk or electronic disk, is a data storage device that uses integrated circuit assemblies as memory to store data persistently. SSD technology uses electronic interfaces compatible with traditional block I/O hard disk drives. SSDs do not employ any moving mechanical components, which distinguishes them from traditional magnetic disks such as hard disk drives (HDDs), which are electromechanical devices containing spinning disks and movable read/write heads. With no seek time or rotational delays, SSDs can deliver substantially better I/O performance than HDDs. Flash SSDs demonstrate latencies that are at least 10 to 50 times lower than the fastest hard disk drives (HDDs), often enabling dramatically improved I/O response times.
445
CPU Cache Random Access Memory (RAM) Flash Express Solid State Drive (SSD) Storage Spinning Dis k Driv e WORM, Tape Library
< 20 ns
< 200 ns
5-20 s 1-3 ms
< 10 ms seconds
Flash Express is as optional PCIe card feature available on zEC12 servers. Flash Express cards are supported in PCIe I/O drawers and can be mixed with other PCIe I/O cards like FICON Express8S, Crypto Express4S or OSA Express4S cards. You must order a minimum of two features (FC 0408) and a maximum of eight. The cards are ordered in increments of two. Flash Express cards are assigned one PCHID even though they have no ports. There is no HCD/IOCP definition required for Flash Express installation. Flash uses sub-channels that are allocated from the .25K reserved in sub-channel set 0. Similar to other PCIe I/O cards, redundant PCIe paths to Flash Express cards are provided by redundant I/O interconnect. Unlike other PCIe I/O cards, they can only be accessed to the host by a unique protocol. A Flash Express PCIe adapter card integrates four SSD cards of 400 GB each for a total of 1.6 TB of usable data per card as showed in Figure C-2.
446
Each card is installed in a PCIe I/O drawer in two different I/O domains. A maximum of two pairs are installed in a drawer with only one flash card per domain only. Greater than two pairs will require a second PCIe I/O drawer. Cards first installed in the front of the installed drawers on slots 1 and 14, before using the rear slots 25 and 33. Each pair of card must formatted prior to utilization. Figure C-3 shows a PCIe I/O cage fully populated with Flash Express cards.
Top View
Card Slot 01 Card Slot 02 Card Slot 03 Card Slot 04 PCI-IN 0 Card Slot 06 Card Slot 07 Card Slot 08 Card Slot 09 FSP Card Slot 11 Card Slot 12 Card Slot 13 Card Slot 14 PCI-IN 2 D o m a i n 2 D o m a i n 0 Card Slot 38 D Card Slot 37 o m Card Slot 36 a Card Slot 35 i PCI-IN 1 n Card Slot 33 1 Card Slot 32 Card Slot 31 Card Slot 30 FSP Card Slot 28 D Card Slot 27 o Card Slot 26 m a Card Slot 25 i n 3 PCI-IN 3 Card Slot 23 Card Slot 22 Card Slot 21 Card Slot 20
Front
Rear
2 interconnect cables
Figure C-3 PCIe I/O cage fully populated with Flash Express cards.
For higher resiliency and high availability, Flash Express cards are always installed by pairs. A maximum of four pairs are supported in a zEC12 system providing with a maximum of 6.4 TB of storage. In each Flash Express card, data is stored in a RAID configuration. If a solid state disk fails, data is reconstructed dynamically. The cards pair are mirror to each other over a pair of cables in RAID 10 configuration mixing mirroring and stripping RAID capabilities. If either card fails, the data is available on the other card. Card replacement is concurrent with customer operations. In addition, Flash Express support concurrent firmware upgrades and card replacement is concurrent with customer operations.
Appendix C. Flash Express
447
The data written on the Flash Express cards is always stored encrypted with a volatile key and the card is only usable on the system with the key that encrypted it. For key management, both the primary and alternate Support Element (SE) have a smart card installed. Smart Card contains both a unique key personalized for each system and a small Crypto engine that can perform a set of security functions within the card.
z/OS
z/OS V1R13a
On the SE or HMC by using the Flash Express allocation panels, you can define the initial and maximum amount of Flash Express available to a logical partition. The maximum memory allocated to a LPAR can be dynamically changed. From the SE or HMC, additional Flash memory up to the maximum can be configured online to a logical partition. On z/OS this can also be done using an operator command. Flash memory can also be configured offline to a logical partition. Figure C-4 on page 449 gives a sample SE/HMC panel interface to used for Flash Express allocation.
448
Figure C-4 Sample SE/HMC panel for Flash Express allocation to LPAR
The new SE user interface for Flash Express provides four new types of actions you can perform: Flash status and control: displays the list of adapters that are installed in the system and their state Manage Flash allocation: display the amount of Flash memory on the system View Flash allocations: displays a table of Flash information for one partition View Flash: display information for one pair of flash adapters. Physical Flash Express PCIe cards are fully virtualized across logical partitions. Each logical partition can be configured with its own SCM address space. The size of Flash Express memory allocated to a partition is done by amount, not by card size. The hardware supports error isolation, transparent mirroring, centralized diagnostic, hardware logging, recovery, independently from the software. At IPL, z/OS detects if flash is assigned to the partition. z/OS automatically uses Flash Express for paging unless specified otherwise via the new z/OS PAGESCM=NONE parameter. All paging data can reside on Flash Express memory. The function is easy to use, and there is no need for capacity planning or placement of data on Flash Express cards. Figure C-5 gives an example of Flash Express allocation between two z/OS LPARs.
449
SC M SPACE
z/OS
Main Memory
z/OS
Main Memory
SC M SPACE
Pa rtition Ma xim um
Pa rtition M ax im um
L P1
LP2
D ata transfer be tween Main Memory and Sto rage Class Memory is via EAD MF (4KB or 1MB blocks)
Pa rtition Initia l Va lue
Flash Express memory is a faster paging device as compared to HDD. It is not replacing memory, but replacing disks. It is suitable for workloads that can tolerate paging and will not benefit to workloads that cannot afford to page. The z/OS design for Flash Express memory does not completely remove the virtual constraints created by a paging spike in the system.z/OS paging subsystem works with a mix of internal Flash Express and external disks. Flash Express improves paging performance. Currently 1 MB large pages are not pageable. With the introduction of Flash Express, 1 MB large pages can reside on Flash and becomes then pageable. Table C-2 on page 450 introduces, for a few z/OS data types supported by Flash Express the choice criteria for data placement on Flash Express or on disk.
Table C-2 Flash Express z/OS supported data types Data type Data page placement
At IPL/NIP time PLPA pages will be placed both on flash and disk. VIO data will always be placed on disk (First to VIO accepting data sets with any spillover flowing to non VIO data sets) If flash space is available, all virtual pages belonging to a HyperSwap Critical Address Space will be placed on flash memory. If flash space is not available, these pages will be kept in memory and only paged to disk when the system is real storage constrained and no other alternatives exist If contiguous flash space is available, pageable large pages will be preferentially written to flash. If available space exists on both flash and disk then make a selection based on response time.
450
Flash Express is used by the Auxiliary Storage Manager (ASM) in conjunction with paging data sets to satisfy page-out and page-in requests received from the Real Storage Manager (RSM). It supports 4KB and 1MB page sizes. ASM determines where to write a page based on space availability, data characteristics, and performance metrics. ASM still requires definition of a PLPA, Common, and at least one local paging data set. VIO pages are only written to DASD (persistence needed for warm starts). A new PAGESCM keyword in IEASYSxx member defines the minimum amount of flash to be reserved for paging. Value may be specified in units of MB, GB, or TB. NONE indicates do not used flash for paging. ALL (default) indicates all flash defined to the partition is available for paging. New messages issued during z/OS IPL indicate the status of SCM: IAR031I USE OF STORAGE-CLASS MEMORY FOR PAGING IS ENABLED - PAGESCM=ALL, ONLINE=00001536M or IAR032I USE OF STORAGE-CLASS MEMORY FOR PAGING IS NOT ENABLED - PAGESCM=NONE. The D ASM and D M commands are enhanced to display flash-related information/status: D ASM lists SCM status along with paging data set status. D ASM,SCM displays summary of SCM usage. D M=SCM display SCM online/offline and increment information. D M=SCM(DETAIL) displays detailed increment-level information. The CONFIG ONLINE command is enhanced to allow bringing additional SCM online: CF SCM (amount), ONLINE.
451
Cryptography Standards #1 (PKCS #1) and then encrypted by the smart card RSA public key. The encrypted content is then written to a file on the Support Element hard disk. This design defines a tight-coupling of the file on the Support Element to the smart card which ensures that any other Support Element will not be able to share the file or the smart card associated with a given SE. It guarantees that the encrypted files are unique and all such smart cards are uniquely tied to their Support Elements. All key generation, encryption, and decryption take place on the smart card. Keys are never in clear. The truly sensitive key the flash encryption key / authentication key only resides in the file on the Support Element until be served to the firmware management of the Flash Express adapter. Figure C-6 shows the cryptographic keys involved in creating this tight-coupling design.
SE
Hard Disk
The flash encryption key / authentication key can be served to the firmware management of the Flash Express adapter either upon request from the firmware at IML time or from the Support Element as the result of a request to change or roll the key. During the alternate Support Element initialization, APIs are called to initialize the alternate smart card in it with the applet code and create the RSA public/private key pair. The API will return the public key of the smart card associated with the alternate Support Element to be used to encrypt the KEK and the Flash encryption key / authentication key from the primary Support Element, sending the resulting encrypted file to the alternate SE for redundancy.
452
HSA
Public Key
Private Key
Private Key
AES Flash Encryption Key / Authentication Key
Firmware Management of the Flash Express Adapter Keys Generated in the Smart Card
AES Flash Encryption Key / Authentication Key Flash Encryption Key / Authentication Key
Flash
The firmware management of the Flash Express adapter requests the flash encryption key / authentication key from the Support Element at Initial Microcode Load (IML) time. When this request arrives, the firmware public key is passed to the Support Element to be used as the transport key. The file containing the KEK-encrypted flash encryption key / authentication key and the firmware public key are passed to the IKC which sends the file contents and the public key to the smart card. The applet on the smart card decrypts the file contents and the flash encryption key / authentication key and then re-encrypts the flash encrypting key / authentication key with the firmware public key. This is then passed back to the Support Element which forwards it on to the firmware management of the Flash Express adapter code. This flow is shown at Figure C-7 on page 453.
453
Primary Support Element failure during IML serving of the flash key
If the primary Support Element fails during the serving of the key, the alternate SE takes over as the primary and restarts the key serving operation.
Alternate Support Element failure during switch over from the primary
If the alternate Support Element fails during switch over when the primary SE fails, the key serving state is lost. When the primary comes back up, the key serving operation may be restarted.
454
8049bibl.fm
Related publications
The publications listed in this section are considered particularly suitable for a more detailed discussion of the topics covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide additional information about the topic in this document. Note that some publications referenced in this list might be available in softcopy only. ????full title???????, xxxx-xxxx ????full title???????, SG24-xxxx ????full title???????, REDP-xxxx ????full title???????, TIPS-xxxx You can search for, view, download or order these documents and other Redbooks, Redpapers, Web Docs, draft and additional materials, at the following website: ibm.com/redbooks
Other publications
These publications are also relevant as further information sources: ????full title???????, xxxx-xxxx ????full title???????, xxxx-xxxx ????full title???????, xxxx-xxxx
Online resources
These websites are also relevant as further information sources: Description1 http://????????.???.???/ Description2 http://????????.???.???/ Description3 http://????????.???.???/
455
8049bibl.fm
456
8049IX.fm
Index
Numerics
1x InfiniBand 35 240V line 386 64-I/128-D KB 76 central processor (CP) 6, 357 central processor complex (CPC) 9 central storage (CS) 105 CFCC 6 channel subsystem 6, 195, 198 logical partitions 195, 198 channel-to-channel (CTC) 11 Chinese Remainder Theorem (CRT) 202, 292294 CHPID number 208 CHPID type OSM 427 Cipher Block Chaining (CBC) 200 CIU application 315316 CIU facility 312314, 328 given server 312, 329 IFLs processors 328 Permanent upgrade 319320 CoD can provide (CP) 312313 Common Cryptographic Architecture 186, 209 Common Cryptographic Architecture (CCA) 186 concurrent book add (CBA) 318 concurrent book replacement 363, 366367 Concurrent Driver Upgrade (CDU) 357358, 368 concurrent hardware upgrade 318 concurrent upgrade 17, 92, 313, 316 cooling requirements 378 Coordinated Server Time (CST) 16 Coordinated Time Network (CTN) 16 Coordinated Timing Network (CTN) 16 coprocessor 1314, 167, 193, 198, 419 coupling facility (CF) 6, 16, 35, 114115 coupling link 4, 10, 35, 163 CP 76, 93, 183184, 194, 317 conversion 9 CP Assist 8, 13, 167 CP capacity 65, 67, 317, 319 CP Cryptographic Assist Facility (CPACF) 87 CP rule 67 CPACF 194 cryptographic capabilities 13 feature code 194 PU design 87 CPC cage 7 CPCs 19, 426 CPE 313, 316 CPM 313 CPs 196, 199, 313314, 317 capacity identifier 317, 324 concurrent and temporary activation 321322 Crypto enablement 193 Crypto Express coprocessor 186, 205 coprocessor feature 207 feature 185186
A
activated capacity 313 active core 8 Active Energy Manager (AEM) 390, 392, 395 Advanced Encryption Standard (AES) 13, 184, 209 application preservation 103 application program interface (API) 320 application program interface (API) 320
B
billable capacity 313 BladeCenter 385386, 425 BladeCenter chassis 17 book ring topology 39, 82 upgrade 61 Broadband RSF 327 Bulk Power Hub (BPH) 12 Bulk Power Regulator (BPR) 377
C
cache level 79 cage I/O cage 198, 326 capacity 313 Capacity Backup 65, 312, 316 See CBU Capacity for Planned Event See CPE Capacity for Planned Event (CPE) 312313 Capacity On Demand (COD) 67, 312 Capacity on Demand (CoD) 313, 322323, 325326 Capacity Provisioning Manager (CPM) 313 capacity setting 65, 313314, 317 full capacity 67 capacity token 321 CBU 313, 316, 321322 contract 322 conversions 67 feature 322 record 65 test 66 5-year CBU contract 66 CBU for CP 65 CBU for IFL 65 CBU test 322
457
8049IX.fm
tamper-resistant feature 183 Crypto Express2 11, 13, 35, 185, 196198, 200, 202 accelerator 14, 200, 202, 208 coprocessor 14, 199, 201, 207208 Crypto Express3 187, 194 operational keys 186 cryptographic asynchronous functions 184 domain 196, 198, 201202 feature codes 193 Cryptographic Function 8, 13, 167, 183184 cryptographic function security-relevant portion 201 cryptography Secure Hash Algorithm (SHA) 13 CSS definition 4 CSSs 4, 196, 198 Customer Engineer (CE) 363 Customer Initiated Upgrade (CIU) 312 activation 331 ordering 330 customer profile 314
F
fanout 35 FC 0864 167, 194, 197 FC 2000 377 FC 3863 167 feature code FC 28xx 367 flexible memory option 360 feature code (FC) 193, 376, 381383, 424 Federal Information Processing Standard (FIPS) 12, 183, 194 FICON channel 11, 35 FICON Express 35 channel 35 FICON Express2 11, 35 FICON Express4 10km LX 151 FICON Express8 11 feature 11 field replaceable unit (FRU) 358 FIPS 140-2 Level 4 183 five-model structure 6 flexible memory option 54, 315, 318, 358 flexible memory option 360, 362 full capacity CP feature 314
D
Data Encryption Standard (DES) 13, 183185, 204 data integrity 184 decimal floating point (DFP) 8 dial modem 327 Digital Signature Verify (CSFNDFV) 186 direct memory access (DMA) 12 double-key DES 184185 double-key MAC 184 dynamic coupling facility dispatching 95 dynamic SAP sparing and reassignment 103 dynamic storage reconfiguration (DSR) 116, 326
G
Gbps 10 granular capacity 65, 67, 93 granularity of both (GB) 326
H
hardware management 17, 376, 425 console 12, 17, 187, 319, 331, 361 hardware management console (HMC) 121 Hardware Security Module (HSM) 200 hardware system area See HSA hardware system area (HSA) 187 HCA2-C fanout 35 HCA2-O LR 35 high water mark 314 HMC 331, 426 host channel adapter (HCA) 9, 21 HSA 3
E
EBA process 361 Model Conversion panel 362 Elliptic Curve Cryptography 186 Digital Signature Algorithm 186 emergency power-off 377 enhanced book availability (EBA) 21, 323, 325, 363 prepare 363 enhanced driver maintenance (EDM) 21 ESA/390 Architecture mode 113 ESA/390 TPF mode 114 ESCON channel 35 estimated capacity (EC) 7 Europay Mastercard VISA (EMV) 2000 185 expanded storage 106 extended translation facility 90 external time source (ETS) 16
I
I/O cage, I/O slot 326 card 323 I/O cage 4, 1011, 35, 199, 315, 361362, 376, 381 installed I/O domains 362 I/O card 9, 312, 359, 369 I/O connectivity 361 I/O domain 362 I/O drawer 312, 318, 359, 376 I/O feature 5, 10 I/O unit 376
458
8049IX.fm
IBM representative 312 IBM Systems Director Active Energy Manager 390 ICF 76 CBU 65 IEC 60793 236 IEDN 12, 17 IEEE Floating Point 90 IFB cable 76 I/O interface 76 IFL 6, 76, 318 IFLs, SAPs, book, memory (IBM) 311 InfiniBand coupling links LR 164 InfiniBand coupling 35 link 10 links 164 InfiniBand Double Data Rate (IB-DDR) 10 initial configuration 318, 323, 327, 377 line-cord pairs 377 initial machine load (IML) 368 initial order (I/O) 167, 196 initial program load (IPL) 327 installed book 4, 318319, 361 additional memory capacity 323 PU count 319 installed record 314 instruction decoding 90 fetching 90 grouping 90 set extensions 91 Instruction fetch and branch (IFB) 358 Integrated Cluster Bus-4 (ICB-4) 15 Integrated Cryptographic Service Facility (ICSF) 13, 184, 188, 204, 290 Integrated Facility for Linux 6 See IFL Intelligent resource director (IRD) 117 inter-book communication 4 Internal Battery Feature (IBF) 7, 69, 76, 377378 Internal Coupling Channels 15 internal coupling facility See ICF internal coupling facility (ICF) 8, 318 internal QDIO (IQDIO) 12 InterSystem Channel-3 15 Intraensemble data network 12, 17 intranode management network 12, 17 IP address 370, 404, 426 ISC-3 link 35 ISC-3 coupling link 164 ISO 16609 CBC Mode 186
L
Large Systems Performance Reference (LSPR) 3 LICCC 318 I/O 318 memory 318 processors 318 Licensed Internal Code (LIC) 5, 16, 193, 318, 323, 359 See also LICCC line-cords 376 Linux 6, 8, 13, 184, 209210, 327 mode 114 Linux-only mode 111 loading of initial ATM keys 185 Local Area Network (LAN) 12, 193, 204, 426 logical partition 4, 12, 107, 195, 198, 201, 317318, 324, 326, 357 dynamic add and delete 111 mode 110 relative priority 361 reserved storage 326 logical processor 4, 306, 324 Long Reach (LR) 35 LPAR single storage pool 106 LPAR mode 111 LSPR 3
M
M80 model 381 machine type 6 master key entry 201 maximum number 196, 198 MBA fanout card 9 Mbps 193 MCI 314, 317, 323 Capacity on Demand 314 list of identifiers 64 model upgrade 317 updated 323 MCM 3, 6, 8 memory allocation 104 card 318, 323, 325 physical 360, 367 size 76 memory hierarchy 23 performance sensitive area 24 memory upgrade 53, 313, 315 MES order 35 MES upgrade 322 message authentication code (MAC) 200 message authentication code (MAC) 184, 200 Message-Security Assist (MSA) 13, 184, 194 MIDAW facility 11 MIF image ID (MIF ID) 201 miscellaneous equipment specification (MES) 312, 322 model capacity 65, 67, 314315
K
key exchange 186 kW 387
Index
459
8049IX.fm
identifier 6465, 68, 317 model M15 317 model M32 50, 317, 324 model M49 50, 317 model M80 50, 318 Model Permanent Capacity Identifier (MPCI) 314 model S08 76, 324 Model Temporary Capacity Identifier (MTCI) 314 model upgrade 317 Modulus Exponent (ME) 202, 292294 MPCI 314 MSU value 22 MTCI 314 multi-chip module 3, 6 multimode fiber 236 multiple platform 393
N
network security considerations 233 Network Time Protocol (NTP) 16 NTP client 16 NTP server 16
O
On/Off Capacity 67, 312, 316 On/Off CoD 15, 6768, 312, 314, 316, 322 activation 321 CP6 temporary CPs 68 granular capacity 67 offering 320 rules 68 Open Systems Adapter (OSA) 11 operating system 6, 18, 305, 313, 316, 318 optionally assignable SAPs 101 OSA-Express 11 OSA-Express2 10 Gb Ethernet LR 35 1000BASE-T Ethernet 35 OSA-Express3 1000BASE-T Ethernet 35 OSA-Express3 feature 12 data router function present 12 oscillator 39 out-of-order (OOO) 9
P
parallel access volume (PAV) 4 Parallel Sysplex 5, 307 certain coupling link configurations 12 system images 117 payment card industry (PCI) 187 PCHID 195, 198, 209 PCI Cryptographic Accelerator (PCICA) 195, 198 PCI Express adapter 35 cryptographic adapter 167, 209 PCICC 195, 198
PCI-e cryptographic adapter level 202 number 196, 199 PCI-X cryptographic adapter 35, 193198 cryptographic coprocessor 194, 197 Peripheral Component Interconnect Express (PCIE) 183 permanent capacity 314 permanent entitlement record (PER) 314 permanent upgrade 314315 retrieve and apply data 332 personal identification number (PIN) 193, 200, 208 physical memory 9, 360, 367 PKA Encrypt 202 PKA Key Import (CSNDPKI) 185 PKA Key Token Change (CSNDKTC) 186 plan-ahead memory 318, 325 capacity 55 planned event 312314 capacity 312, 316 point of sale (POS) 185 point unit (PU) 3, 316, 318, 360 port J07 427 power consumption 390391 power distribution unit (PDU) 386 power estimation tool 390 POWER7 blade system support 19 power-on reset (POR) 196, 199, 362 PR/SM 107 primary HMC 370, 426 explicit action 424 policy information 426 processing unit (PU) 6, 9, 76, 102, 318, 329 characterization 102 concurrent conversion 318 conversion 318 feature code 6 maximum number 6 pool 92 sparing 92 type 111, 318, 367 Processor Resource/Systems Manager (PR/SM) 4, 201 processor unit (PU) 3 production workload 322 Provide Cryptographic Key Management Operation (PCKMO) 188, 194 Pseudo Random Number Generation (PRNG) 184 pseudorandom number generator (PRNG) 184185, 209 PU chip 8, 42 PU type 316, 367 public key algorithm 185186, 200, 204 decrypt 209 encrypt 185, 209 public key algorithm (PKA) 201 Pulse per Second (PPS) 17 purchased capacity 314 PUs 3, 317318, 360
460
8049IX.fm
Q
queued direct input/output (QDIO) 12
R
Redbooks website 455 Contact us xxii reducing all sources (RAS) 5, 12 redundant array of independent drives (RAID) 359 redundant array of independent memory (RAIM) 3, 9, 359 redundant I/O 361 redundant I/O interconnect (RII) 9, 21, 362 relative nest intensity (RNI) 24 reliability, availability, serviceability (RAS) 1920 remaining book 360 sufficient dormant resources 360 sufficient inactive physical resources 362 Remote Support Facility (RSF) 313, 319320, 330331 replacement capacity 313314 reserved storage 115 Resource Access Control Facility (RACF) 188 Resource Link 313314, 329 CIU application 316 CIU facility 319 machine profile 331 Rivest-Shamir-Adelman (RSA) 185, 200, 202, 210
subcapacity setting 4 subchannel 357 Summary 328 superscalar 86 superscalar processor 86 Support Element (SE) 7, 314, 331, 362 Change LPAR Cryptographic Controls task 196, 199 logical partition 207 support element (SE) 17, 76, 196, 199, 202, 368, 395, 426 SUSE Linux Enterprise Server (SLES) 116 system activity display (SAD) 390 system assist processor variable number 8 system assist processor (SAP) 6, 9, 317 System Management Facilities (SMF) 204 System Storage Interoperation Center (SSIC) 237 system upgrade 312 System z 3, 32, 35, 184, 311, 319 distribution 19 hardware platform management 19 High Performance FICON 11 New Application License Charge 307 UDX toolkit 186 System z BladeCenter Extension (zBX) 1718
T
target configuration 318 TB 1, 9 temporary capacity 6768, 314 CP count 68 temporary entitlement record (TER) 314 temporary upgrade 312 time synchronization 16 TKE 201 additional smart cards 193 Smart Card Reader 193 workstation 1415, 21, 193, 203 workstation feature 203 Top of Rack (TOR) switches 370 total number 6, 65, 319, 362 Transaction Processing Facility (TPF) 327 translation look-aside buffer (TLB) 90 Transport Layer Security (TLS) 14, 183 triple-key DES 185 Trusted Key Entry (TKE) 14, 193, 203
S
SAP 6, 101 concurrent book replacement 367 number of 76, 318, 331 SC chip 8, 45 SE 426 secondary approval 314 Secure Hash Algorithm (SHA) 13 Secure Sockets Layer (SSL) 14, 183, 185, 194, 202, 208 Server Time Protocol (STP) 1516, 165 SHA-1 184 SHA-1 and SHA-256 184 SHA-256 184 single I/O path 365 rule 367 single-key MAC 184 small form factor pluggable (SFP) 232 soft capping 306 specialty engine 4, 67, 315, 361 SSL/TLS 14, 183 staged CoD records 15 staged record 314 stand-alone z196 ensemble node 427 Standard SAP 76 storage area network (SAN) 11 store system information (STSI) instruction 323 STP-only CTN 17 subcapacity 314 subcapacity model 317 subcapacity models 317
U
unassigned IFL 6364 unassigned PUs 323 Unified Resource Manager 2, 12, 370 unplanned upgrades 320 unused PUs 318, 322, 360 upgrade 62 for I/O 326 for memory 325 for processors 323 permanent upgrade 328
Index
461
8049IX.fm
V
virtual server 12, 18, 427 VPD 315
W
wild branch 89 Workload License Charge (WLC) 305, 329 CIU 325
Z
z/Architecture 6, 194 z/OS 290, 313 Capacity Provisioning Manager 15 z/OS logical partition EBA function 368 z/TPF 19 z/VM V5R4 4, 103 z10 EC 7 z10 server 10 zAAP 61, 76, 95, 361, 367 CBU 65 LPAR definitions 96 zBX 17 zBX Model 003 upgrade 327 zBX Rack-B 228 zEnterprise 196 12, 4, 6, 375 zEnterprise BladeCenter Extension (ZBX) 2, 241, 312, 322, 358, 375, 385, 427 zEnterprise System 2, 370 environmental requirements 375 zIIP 6, 61, 76, 318319
462
To determine the spine width of a book, you divide the paper PPI into the number of pages in the book. An example is a 250 page book using Plainfield opaque 50# smooth which has a PPI of 526. Divided 250 by 526 which equals a spine width of .4752". In this case, you would use the .5 spine. Now select the Spine width for the book and hide the others: Special>Conditional Text>Show/Hide>SpineSize(-->Hide:)>Set . Move the changed Conditional text settings to all files in your book by opening the book file with the spine.fm still open and File>Import>Formats the
8049spine.fm
463
To determine the spine width of a book, you divide the paper PPI into the number of pages in the book. An example is a 250 page book using Plainfield opaque 50# smooth which has a PPI of 526. Divided 250 by 526 which equals a spine width of .4752". In this case, you would use the .5 spine. Now select the Spine width for the book and hide the others: Special>Conditional Text>Show/Hide>SpineSize(-->Hide:)>Set . Move the changed Conditional text settings to all files in your book by opening the book file with the spine.fm still open and File>Import>Formats the
8049spine.fm
464
Back cover
Describes the zEnterprise System and related features and functions Discusses hardware and software capabilities Explains virtualizing and managing the infrastructure for complex applications
This IBM Redbooks publication discusses the new IBM zEnterprise System, which consists of the IBM zEnterprise EC12 (zEC12), an updated IBM zEnterprise Unified Resource Manager, and the IBM zEnterprise BladeCenter Extension (zBX) Model 003. The zEC12 is designed with improved scalability, performance, security, resiliency, availability, and virtualization. The superscalar design allows the zEC12 to deliver a record level of capacity over the prior System z servers. It is powered by 120 of the worlds most powerful microprocessors running at 5.5 GHz and is capable of executing more than 75,000 millions of instructions per second (MIPS). The zEC12 Model HA1 is estimated to provide up to 50% more total system capacity than the z196 Model M80. The zBX Model 003 infrastructure works with the zEC12 to enhance System z virtualization and management through an integrated hardware platform that spans mainframe, POWER7, and System x technologies. Through the Unified Resource Manager, the zEnterprise System is managed as a single pool of resources, integrating system and workload management across the environment. This book provides information about the zEnterprise System and its functions, features, and associated software support. Greater detail is offered in areas relevant to technical planning. This book is intended for systems engineers, consultants, planners, and anyone wanting to understand the zEnterprise System functions and plan for their usage. It is not intended as an introduction to mainframes. Readers are expected to be generally familiar with existing IBM System z technology and terminology.