SG 248951
SG 248951
Redbooks
Note: Before using this information and the product it supports, read the information in “Notices” on
page xiii.
This edition applies to IBM z16 Model A01, Machine Type 3931.
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiv
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xviii
Contents iii
2.5.6 Virtual Flash Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
2.5.7 Flexible Memory Option . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
2.6 Reliability, availability, and serviceability. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
2.6.1 RAS in the CPC memory subsystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
2.6.2 General IBM z16 A01 RAS features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
2.7 Connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
2.7.1 Redundant I/O interconnect . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
2.7.2 Enhanced drawer availability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
2.7.3 CPC drawer upgrade . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
2.8 Model configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
2.8.1 Upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
2.8.2 Model capacity identifier . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
2.9 Power and cooling. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
2.9.1 PDU-based configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
2.9.2 Bulk Power Assembly-based configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
2.9.3 Power estimation tool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
2.9.4 Cooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
2.9.5 Radiator Cooling Unit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
2.10 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
Contents v
Chapter 5. Central processor complex channel subsystem . . . . . . . . . . . . . . . . . . . . 199
5.1 Channel subsystem. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
5.1.1 Multiple logical channel subsystems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
5.1.2 Multiple subchannel sets. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
5.1.3 Channel path spanning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 206
5.2 I/O configuration management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 208
5.3 Channel subsystem summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 209
Contents vii
8.9 Capacity for Planned Event. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 367
8.10 Capacity Backup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 367
8.10.1 Ordering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 368
8.10.2 CBU activation and deactivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 370
8.10.3 Automatic CBU enablement for GDPS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 371
8.11 Flexible Capacity Cyber Resiliency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 372
8.12 Planning for nondisruptive upgrades. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 374
8.12.1 Components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375
8.12.2 Concurrent upgrade considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376
8.13 Summary of Capacity on-Demand offerings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 380
Contents ix
12.9.1 Collect CPU MF counter data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 486
12.9.2 Creating EDF files with CP3KEXTR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 486
12.9.3 Loading EDF files to the capacity planning tool . . . . . . . . . . . . . . . . . . . . . . . . 487
12.9.4 IBM z16 Performance Best Practices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 488
12.9.5 IBM zPCR HiperDispatch Report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 489
12.9.6 IBM zPCR Topology Report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 492
12.9.7 IBM z16 HMC - View Partition Resource Assignments. . . . . . . . . . . . . . . . . . . 493
12.9.8 IBM zPCR Large Partition Support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 495
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 515
Tailored Fit Pricing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 516
Software Consumption Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 517
International Program License Agreement in the Software Consumption Model . . . . . 518
Hardware Consumption Model. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 518
Tailored Fit Pricing for IBM Z hardware in more detail . . . . . . . . . . . . . . . . . . . . . . . . . 519
Efficiency benefits of TFP-Hardware. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 521
Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 521
IBM Z Flexible Capacity for Cyber Resiliency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 522
Use cases of IBM Flexible Capacity for Cyber Resiliency . . . . . . . . . . . . . . . . . . . . . . . . . 522
Disaster recovery and DR testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 522
Frictionless compliance. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 522
Facility maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 523
Pro-active avoidance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 523
How does IBM Flexible Capacity for Cyber Resiliency work? . . . . . . . . . . . . . . . . . . . . . . 523
Set up process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 523
Transferring capacity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 524
Multi-system environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 526
Tailored fit pricing for hardware and IBM Z Flexible Capacity for Cyber Resiliency. . . . . . 527
Ordering and installing IBM Z Flexible Capacity for Cyber Resilience . . . . . . . . . . . . . . . . 528
Terms and conditions of IBM Z Flexible Capacity for Cyber Resiliency. . . . . . . . . . . . . . . 529
IBM Z Flexible Capacity for Cyber Resilience versus Capacity Back Up. . . . . . . . . . . . . . 530
Contents xi
xii IBM z16 (3931) Technical Guide
Notices
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
Notices xiii
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines
Corporation, registered in many jurisdictions worldwide. Other product and service names might be
trademarks of IBM or other companies. A current list of IBM trademarks is available on the web at “Copyright
and trademark information” at http://www.ibm.com/legal/copytrade.shtml
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
AIX® IBM Z® System z®
CICS® IBM z Systems® System z10®
Connect:Direct® IBM z13® System z9®
DB2® IBM z13s® VTAM®
Db2® IBM z14® WebSphere®
DS8000® IBM z16™ z Systems®
FICON® Interconnect® z/Architecture®
FlashCopy® Language Environment® z/OS®
GDPS® OMEGAMON® z/VM®
Guardium® Parallel Sysplex® z/VSE®
HyperSwap® Passport Advantage® z13®
IBM® PIN® z13s®
IBM Blockchain™ RACF® z15™
IBM Cloud® Redbooks® z16™
IBM Security® Redbooks (logo) ® z9®
IBM Sterling® Resource Link® zEnterprise®
IBM Watson® Sterling™
The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.
Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States,
other countries, or both.
Java, and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its
affiliates.
Red Hat, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the United States and
other countries.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Other company, product, or service names may be trademarks or service marks of others.
This IBM® Redbooks® publication describes the features and functions of the latest member
of the IBM Z® platform that was built with the IBM Telum processor: the IBM z16™ (machine
type 3931). It includes information about the IBM z16 processor design, I/O innovations,
security features, and supported operating systems.
The IBM Z platform is recognized for its security, resiliency, performance, and scale. It is relied
on for mission-critical workloads and as an essential element of hybrid cloud infrastructures.
The IBM z16 server adds capabilities and value with innovative technologies that are needed
to accelerate the digital transformation journey.
The IBM z16 is a state-of-the-art data and transaction system that delivers advanced
capabilities, which are vital to any digital transformation. The IBM z16 is designed for
enhanced modularity, which is in an industry standard footprint.
This book explains how this system uses new innovations and traditional IBM Z strengths to
satisfy growing demand for cloud, analytics, and open source technologies. With the IBM z16
as the base, applications can run in a trusted, reliable, and secure environment that improves
operations and lessens business risk.
Authors
This book was produced by a team of specialists from around the world working at IBM
Redbooks, Poughkeepsie Center.
Octavian Lascu is an IBM Redbooks Project Leader with over 25 years of IT infrastructure
experience. He specializes in designing, implementing, and supporting complex IT
infrastructure environments (systems, storage, and networking), including high availability
and disaster recovery solutions and high-performance computing deployments. He has
developed materials for and taught over 50 workshops for technical audiences around the
world. He is the author of several IBM publications.
Bill White is an IBM Redbooks Project Leader and Senior Networking and Connectivity
Specialist at IBM Redbooks, Poughkeepsie Center.
Ewerson Palacio is an IBM Redbooks Project Leader. He holds Bachelor’s degree in Math
and Computer Science. Ewerson worked for IBM Brazil for over 40 years and retired in 2017
as an IBM Distinguished Engineer. Ewerson co-authored many IBM Z Redbooks, and created
and presented ITSO seminars around the globe.
Preface xv
John Troy is an IBM Z and storage hardware National Top Gun in the northeast area of the
US. He has over 40 years of experience in the service field. His areas of expertise include
IBM Z servers and high-end storage systems technical and customer support and services.
John has also been an IBM Z hardware technical support course designer, developer, and
instructor for the last eight generations of IBM high-end servers.
Jannie Houlbjerg is a Systems Programmer working at JN Data in Denmark. She has more
than 20 years of experience in the IBM Z field. Her areas of expertise include IBM Z hardware
and infrastructure, IBM Parallel Sysplex®, connectivity, performance, IBM GDPS®, and
technical project management and documentation.
Martijn Raave is an IBM Z and LinuxONE Client Architect and Hardware Technical Specialist
for IBM Northern Europe. Over a period of over 20 years, his professional career has revolved
around the mainframe platform, supporting several large Dutch customers in their technical
and strategic journey on IBM Z. His focus areas are hardware, resiliency, availability,
architecture, and any other topics about IBM Z.
Kazuhiro Nakajima is a Senior IT Specialist in IBM Japan. He has almost 30 years career in
IBM Japan and he has been active as an advanced Subject Matter Expert of IBM Z products
for over 20 years. His areas of expertise include IBM Z hardware, performance, z/OS®, and
IBM Z connectivity. He has been a co-author of several IBM Z configuration set up IBM
Redbooks publications from the IBM zEC12 to the IBM z14®.
Paul Schouten is an IBM Z Client Technical Specialist based in Sydney, Australia. During his
40 years supporting mainframe systems, he has performed many roles, including Certified IT
Architect, Systems Software developer, and Systems Programming. He has extensive
experience developing and documenting high availability solutions for IBM’s enterprise
customers.
André Spahni is an IBM Z Client Technical Specialist based in Zurich, Switzerland. He has
20 years of experience working with and supporting IBM Z clients. André has worked for
EMEA 2nd level supporter and national Top Gun. His areas of expertise include IBM Z
hardware, HMC/SE, and connectivity.
Anna Shugol is a mainframe technical specialist with IBM UK. She has over 8 years of
Mainframe experience and has worked with clients in various geographies. Large system
performance is one of her key areas of expertise. Anna holds a Computer Science degree
from Bauman Moscow State Technical University.
Hervey Kamga is an IBM Z Product Engineer with the EMEA I/O Connectivity Team in
Montpellier, France. Before serving in his current role, he was a Support Engineer and
Engineer On Site for 13 years with Sun MicroSystems and Oracle in EMEA. Hervey’s areas of
expertise include Oracle Solaris (operating system and hardware products), virtualization
(VMware and virtualBox), Linux, and IBM Z I/O features and protocols (IBM FICON® and
OSA) while co-authoring several IBM Redbooks publications.
Slav Martinski works as a hardware technical support specialist for IBM Europe. He has
more than 5 years experience working on IBM Power Servers. Currently, he is working in
IBM’s Z Systems support team. Slav holds a Computer Science degree from Delaware
County College.
Markus Ertl is a Senior IT Specialist in IBM Germany. He has more than 15 years of
experience with IBM Z, working with clients from various industries. His area of expertise
includes IBM Z hardware and infrastructure, performance, and Capacity on Demand topics.
Robert Haimowitz
IBM Redbooks, Poughkeepsie Center
Dave Surman, Darelle Gent, David Hutton, Eysha Powers, Patty Driever, Anthony Sofia, Brian
Valentine, Purvi Patel, Bill Bitner, Christine Smith, Barbara Weiler, Tom Morris, Marna Walle,
Franco Pinto, Walter Niklaus, Martin Recktenwald, Ron Geiger, Frank Miele, Lisa Schloemer,
Richard Gagnon, Susan Farrell, Les Geer III, Yamil Rivera.
IBM
Preface xvii
Now you can become a published author, too!
Here’s an opportunity to spotlight your skills, grow your career, and become a published
author—all at the same time! Join an IBM Redbooks residency project and help write a book
in your area of expertise, while honing your experience using leading-edge technologies. Your
efforts will help to increase product acceptance and customer satisfaction, as you expand
your network of technical contacts and relationships. Residencies run from two to six weeks
in length, and you can participate either in person or as a remote resident working from your
home base.
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an email to:
[email protected]
Mail your comments to:
IBM Corporation, IBM Redbooks
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
The new member of the IBM Z family, IBM z16, continues that commitment and adds value
with innovative technologies that can help accelerate the digital transformation journey.
The IBM z16 is the first IBM Z system that is built with the IBM Telum processor1. It is
designed to help businesses meet the following goals:
Create value in every interaction and to optimize decision making, with the on-chip
Artificial Intelligence (AI) accelerator. The Accelerator for AI can deliver the speed and
scale that is required to infuse AI inferencing into workloads with no impact on service
delivery.
Act now to protect today’s data against current and future threats with quantum-safe
protection immediately through quantum-safe cryptography APIs and crypto discovery
tools.
Enhance resiliency with flexible capacity to dynamically shift system resources across
locations to proactively avoid disruptions.
Modernize and integrate applications and data in a hybrid cloud environment with
consistent and flexible deployment options to innovate with speed and agility.
Reduce cost and keep up with changing regulations through a solution that helps simplify
and streamline compliance tasks.
This chapter describes the basic characteristics of the IBM z16 platform. It includes the
following topics:
1.1, “IBM z16 A01 highlights” on page 2
1.2, “IBM z16 A01 technical overview” on page 6
1.3, “Hardware management” on page 12
1.4, “Reliability, availability, and serviceability” on page 13
1 IBM Telum Processor: the next-gen microprocessor for IBM Z and IBM LinuxONE
The new processor chip design has a new cache hierarchy, on-chip AI accelerator that is
shared by the PU cores, transparent memory encryption, and increased uniprocessor
capacity (single thread and SMT similar).
The on-chip AI scoring logic provides submicrosecond AI inferencing for deep learning and
complex neural network models.
The result is improved system performance and scalability with 1.5x more cache capacity per
core over the IBM z15 and reduced average access latency through a flatter topology.
The IBM z16 (machine type 3931) has one model: the A01. The maximum number of
characterizable processor units (PUs) with the IBM z16 is represented by feature names:
Max39, Max82, Max125, Max168, and Max200.
The number of characterizable PUs, spare PUs, System Assist Processors (SAPs), and
Integrated Firmware Processors (IFP) are included with each IBM z16 feature (see
Table 1-1).
Max39 1 0667 1 - 39 5 2
Max82 2 0668 1 - 82 10 2
The IBM z16 memory subsystem uses proven redundant array of independent memory
(RAIM) technology to ensure high availability. Up to 40 TB (10 TB per CPC drawer) of
addressable memory per system can be ordered.
The IBM z16 also has unprecedented capacity to meet consolidation needs with innovative
I/O features for transactional and hybrid cloud environments.
The IBM z16 (maximum configuration) can support up to 12 PCIe+ I/O drawers. Each I/O
drawer can support up to 16 I/O or special purpose features for storage, network, and
clustering connectivity, and cryptography.
The IBM z16 is more flexible and features simplified on-demand capacity to satisfy peak
processing demands and quicker recovery times with built-in resiliency capabilities. The
Capacity on Demand (CoD) function can dynamically change available system capacity. This
function can help respond to new business requirements with flexibility and precise
granularity.
The IBM Tailored Fit Pricing for IBM Z options delivers unmatched simplicity and predictability
of hardware capacity and software pricing, even in the constantly evolving era of hybrid cloud.
Consider the following points:
The IBM z16 enhancements in resiliency include a capability that is called IBM Z Flexible
Capacity for Cyber Resiliency. With Flexible Capacity for Cyber Resiliency, you can
remotely shift capacity and production workloads between IBM z16 systems at different
sites on demand with no onsite personnel or IBM intervention. This capability is designed
to help you proactively avoid disruptions from unplanned events and planned scenarios,
such as site facility maintenance.
IBM z16 System Recovery Boost enhancements provide boosted processor capacity and
parallelism for specific events. Client-selected middleware starts and restarts to expedite
recovery for middleware regions and restore steady-state operations as soon as possible.
IBM SAN Volume Controller memory dump processing and HyperSwap configuration load
and reload are boosted to minimize the effect on running workloads.
On IBM z16, the enhanced ICA-SR coupling link protocol provides up to 10%
improvement for read requests and lock requests, and up to 25% for write requests and
duplexed write requests, compared to CF service times on IBM z15 systems. The
improved CF service times for CF requests can translate into better Parallel Sysplex
coupling efficiency; therefore, the software costs can be reduced for the attached z/OS
images in the Parallel Sysplex.
IBM z16 provides improved CF processor scalability for CF images. Compared to IBM
z15, the relative scaling of a CF image beyond a 9-way is significantly improved; that is,
that the effective capacity of IBM z16 CF images continues to increase meaningfully all the
way up to the maximum of 16 processors in a CF image.
The IBM z16 also added functions to protect today’s data now, and from future cyberattacks
that can be initiated by quantum computers. The IBM z16 provides the following
quantum-safe capabilities:
Key generation
Encryption
Key encapsulation mechanisms
Hybrid key exchange schemes
Dual digital signature schemes
The IBM z16 provides increased processing and enhanced I/O capabilities over its
predecessor, the IBM z15 T01. This capacity is achieved by increasing the number of PUs per
system, redesigning the system cache, and introducing new I/O technologies.
The IBM z16 feature Max200 is estimated to provide up to 17% more total system capacity
than the IBM z15 Model Max190, with the same amount of memory and power requirements.
With up to 40 TB of main storage and enhanced SMT, the performance of the IBM z16 A01
processors deliver considerable improvement. Uniprocessor performance also increased
significantly. An IBM z16 Model 701 offers average performance improvements of up to 11%
over the IBM z15 Model 701.2
The Integrated Facility for Linux (IFL) and IBM Z Integrated Information Processor (zIIP)
processor units on the IBM z16 can be configured to run two simultaneous threads in a single
processor (SMT). This feature increases the capacity of these processors with 25% in
average2 over processors that are running single thread. SMT is also enabled by default on
System Assist Processors (SAPs).
Within each single drawer, IBM z16 provides 25% greater capacity than IBM z15 for standard
models and 40% greater capacity on the maximum configured model, which enables efficient
partition scaling.
For more information about millions of service units (MSUs) ratings, see the IBM Z Software
Contracts website.
IBM plans to support 21st Century Software VSEn V6.3 on IBM z16. For more information,
see 7.2.4, “21st Century Software z/VSEn V6.3” on page 251.
IBM plans to support the following Linux on IBM Z distributions on IBM z16:
SUSE SLES 15 SP3 and SUSE SLES 12 SP5
Red Hat RHEL 8.4 and Red Hat RHEL 7.9
Ubuntu 22.04 LTS, Ubuntu 21.10, and Ubuntu 20.04.1 LTS
The support statements for the IBM z16 also cover the KVM hypervisor on distribution levels
that have KVM support.
For more information about the features and functions that are supported on IBM z16 by
operating system, see Chapter 7, “Operating system support” on page 247.
The compilers increase the return on your investment in IBM Z hardware by maximizing
application performance by using the compilers’ advanced optimization technology for
IBM z/Architecture®.
To fully use the capabilities of the IBM z16, you must compile it by using the minimum level of
each compiler. To obtain the best performance, you must specify an architecture level of 14 by
using the ARCH(14) option.
For more information, see 7.5.8, “z/OS XL C/C++ considerations” on page 321.
1.2.1 Frames
The IBM z16 A01 uses 19-inch frames and industry-standardized power and hardware. It can
be configured as a one-, two-, three-, or four-frame system. The IBM z16 A01 packaging is
compared to the two previous IBM Z platforms in Table 1-2.
Table 1-2 IBM z16 configuration options compared to IBM z14 and IBM z15 configurations
System Number of Number of CPC Number of I/O I/O and power Power Cooling
frames drawers drawers connections options options
IBM z16 1-4 1-4 0 - 12a Rear only PDU or Radiator (air)
A01 BPA only
IBM z15 1-4 1-5 0 - 12b Rear only PDU or Radiator (air)
T01 BPA or water-
cooling unit
(WCU)
IBM z14 2 1-4 0-5 Front and rear BPA Radiator (air)
M0x or WCU
a. Maximum of 12 if ordered with a PDU or maximum of 10 if ordered with a BPA.
b. Maximum of 12 if ordered with a PDU or maximum of 11 if ordered with a BPA.
With the IBM z16, Virtual Flash Memory (VFM) feature is offered from the main memory
capacity in 0.5 TB units to increase granularity for the feature. VFM can provide much simpler
management and better performance by eliminating the I/O to the adapters in the PCIe+ I/O
drawers.
For a four CPC drawer system, up to 48 PCIe+ fan-out slots can be populated with fan-out
cards for data communications between the CPC drawers and the I/O infrastructure, and for
coupling. The multiple channel subsystem (CSS) architecture allows up to six CSSs, each
with 256 channels.
FICON Express
FICON Express features follow the established Fibre Channel (FC) standards to support data
storage and access requirements, along with the latest FC technology in storage and access
devices. FICON Express features support the following protocols:
FICON
This enhanced protocol (as compared to FC) provides for communication across
channels, channel-to-channel (CTC) connectivity, and with FICON devices, such as disks,
tapes, and printers. It is used in z/OS, z/VM, IBM z/VSE® (Virtual Storage Extended),
z/TPF (Transaction Processing Facility), and Linux on IBM Z environments.
Fibre Channel Protocol (FCP)
This standard protocol is used for communicating with disk and tape devices through FC
switches and directors. The FCP channel can connect to FCP SAN fabrics and access
FCP/SCSI devices. FCP is used by z/VM, KVM, z/VSE, and Linux on IBM Z environments.
FICON Express32S features are implemented by using PCIe cards, and offer better port
granularity and improved capabilities over the previous FICON Express features. FICON
Express32S features support a link data rate of 32 gigabytes per second (Gbps) (8, 16, or
32 Gbps auto-negotiate), and it is the preferred technology for new systems.
zHyperLink Express
zHyperLink was created to provide fast access to data by way of low-latency connections
between the IBM Z platform and storage.
The zHyperLink Express1.1 feature allows you to make synchronous requests for data that is
in the storage cache of the IBM DS8900F. This process is done by directly connecting the
zHyperLink Express1.1 port in the IBM z16 to an I/O Bay port of the IBM DS8000. This short
distance (up to 150 m [492 feet), direct connection is designed for low-latency reads and
writes, such as with IBM DB2® for z/OS synchronous I/O reads and log writes.
Working with the FICON SAN Infrastructure, zHyperLink can improve application response
time, which cuts I/O-sensitive workload response time in half without requiring application
changes.3
Note: The zHyperLink channels complement FICON channels, but they do not replace
FICON channels. FICON remains the main data driver and is mandatory for zHyperLink
usage.
3 The performance results can vary depending on the workload. Use the zBNA tool for the zHyperLink planning.
OSA-Express
The OSA-Express features provide local area network (LAN) connectivity and comply with
IEEE standards. In addition, OSA-Express features assume several functions of the TCP/IP
stack that normally are performed by the PU, which allows significant performance benefits by
offloading processing from the operating system.
OSA-Express7S 1.2 features continue to support copper and fiber optic (single-mode and
multimode) environments.
HiperSockets
IBM HiperSockets is an integrated function of the IBM Z platforms that supplies attachments
to up to 32 high-speed virtual LANs, with minimal system and network overhead.
HiperSockets is a function of the Licensed Internal Code (LIC). It provides LAN connectivity
across multiple system images on the same IBM Z platform by performing
memory-to-memory data transfers in a secure way.
The HiperSockets function eliminates the use of I/O subsystem operations. It also eliminates
having to traverse an external network connection to communicate between LPARs in the
same IBM Z platform. In this way, HiperSockets can help with server consolidation by
connecting virtual servers and simplifying the enterprise network.
RoCE Express
The 25 GbE and 10 GbE RoCE Express3 features4 use Remote Direct Memory Access
(RDMA) over Converged Ethernet (RoCE) to provide fast memory-to-memory
communications between two IBM Z platforms.
These features help reduce consumption of CPU resources for applications that use the
TCP/IP stack (such as IBM WebSphere® that accesses an IBM Db2® database). They also
can help reduce network latency with memory-to-memory transfers by using Shared Memory
Communications over RDMA (SMC-R).
With SMC-R, you can transfer huge amounts of data quickly and at low latency. SMC-R is
transparent to the application and requires no code changes, which enables rapid time to
value.
The RoCE Express3 features also can provide LAN connectivity for Linux on IBM Z, and
comply with IEEE standards. In addition, RoCE Express features assume several functions of
the TCP/IP stack that normally are performed by the PU, which allows significant
performance benefits by offloading processing from the operating system.
4 RoCE Express features also can be used as general-purpose Network Interface Cards (NICs) with Linux on IBM Z.
SMC-D requires no extra physical resources (such as RoCE Express features, PCIe
bandwidth, ports, I/O slots, network resources, or Ethernet switches). Instead, SMC-D uses
LPAR-to-LPAR communication through HiperSockets or an OSA-Express feature for
establishing the initial connection.
z/OS and Linux on IBM Z support SMC-R and SMC-D. Now, data can be shared by way of
memory-to-memory transfer between z/OS and Linux on IBM Z.
Coupling connectivity on the IBM z16 use Coupling Express2 Long Reach (CE2 LR) and
Integrated Coupling Adapter Short Reach (ICA SR and ICA SR1.1) features. The ICA SR
feature supports distances up to 150 meters (492 feet); the CE2 LR feature supports
unrepeated distances of up to 10 km (6.21 miles) between IBM Z platforms. ICA SR features
provide sysplex and timing connectivity direct to the CPC drawer, while Coupling Express2
LR features connect into the PCIe+ I/O Drawer.
Coupling links also can carry timing information (such as Server Time Protocol [STP]) for
synchronizing time across multiple IBM Z CPCs in a Coordinated Time Network (CTN).
For more information about coupling and clustering features, see 4.5, “I/O features” on
page 158.
1.2.7 Cryptography
IBM z16 provides two main cryptographic functions: CP Assist for Cryptographic Functions
(CPACF) and Crypto-Express8S.
CPACF
CPACF is a high-performance, low-latency coprocessor that performs symmetric key
encryption operations and calculates message digests (hashes) in hardware. The following
algorithms are supported:
AES
Data Encryption Standard (DES) and Triple Data Encryption Standard (TDES)
Secure Hash Algorithm (SHA)-1
SHA-2
SHA-3
Crypto-Express8S
The tamper-sensing and tamper-responding Crypto-Express8S features provide acceleration
for high-performance cryptographic operations and support up to 85 domains with the IBM
z16 A01. This specialized hardware performs AES, DES and TDES, RSA, Elliptic Curve
(ECC), SHA-1, and SHA-2, and other cryptographic operations.
It supports specialized high-level cryptographic APIs and functions, including those functions
that are required with quantum-safe cryptography and in the banking industry.
Crypto-Express8S features are designed to meet the Federal Information Processing
Standards (FIPS) 140-2 Level 4 and PCI HSM security requirements for hardware security
modules.
IBM z16 is the industry’s first quantum-safe system5. Consider the following points:
IBM z16 quantum-safe secure boot technology helps to protect IBM Z firmware from
quantum attacks by using a build-in dual signature scheme with no changes required.
IBM z16 quantum-safe technology and key management services were developed to help
you protect data and keys against a potential future quantum attack, such as harvest now,
decrypt later.
IBM z16 positions customers to use quantum-safe cryptography along with classic
cryptography as they begin modernizing existing applications and building new
applications.
For more information about cryptographic features and functions, see Chapter 6,
“Cryptographic features” on page 209.
The IBM z16 delivers a range of features and functions that allows PUs to concentrate on
computational tasks, while distinct, specialized features take care of the rest. For more
information about these features and other IBM z16 features, see in 3.5, “Processor unit
functions” on page 98.
The HMC is an appliance that provides a single point of control for managing local or remote
hardware elements.
For IBM z16 new built systems, IBM Z Hardware Management Appliance (FC 0129) is the
only available HMC. The HMC Appliance and SE Appliance run virtualized on the SE
hardware.
Older stand-alone HMCs (rack-mounted or tower) can be carried forward during an MES
upgrade to IBM z16.
For more information, see Chapter 10, “Hardware Management Console and Support
Element” on page 407.
6
The Crypto Express8S is available with either one or two hardware security modules (HSM). The HSM is the IBM
4770 PCIe Cryptographic Coprocessor (PCIeCC).
7 The Crypto Express7S comes with either one (1-port) or two (2-port) hardware security modules (HSM). The HSM
is the IBM 4769 PCIe Cryptographic Coprocessor (PCIeCC).
8
ICA SR and ICA SR1.1 features connect directly into the CPC (processor) Drawer, while Coupling Express2 Long
Reach connects into the PCIe+ I/O Drawer.
IBM Z platforms are designed to enable highest availability and lowest downtime. These facts
are recognized by various IT analysts, such as ITIC9 and IDC10. A comprehensive,
multi-layered strategy includes the following features:
Error Prevention
Error Detection and Correction
Error Recovery
System Recovery Boost
With a suitably configured IBM z16, further reduction of outages can be attained through First
Failure Data Capture (FFDC), which is designed to reduce service times and avoid
subsequent errors. It also improves nondisruptive replace, repair, and upgrade functions for
memory, drawers, and I/O adapters. IBM z16 supports the nondisruptive download and
installation of LIC updates.
IBM z16 RAS features provide unique high-availability and nondisruptive operational
capabilities that differentiate IBM Z in the marketplace. IBM z16 RAS enhancements are
made on many components of the CPC (processor chip, memory subsystem, I/O, and
service) in areas, such as error checking, error protection, failure handling, error checking,
faster repair capabilities, sparing, and cooling.
The ability to cluster multiple systems in a Parallel Sysplex takes the commercial strengths of
the z/OS platform to higher levels of system management, scalable growth, and continuous
availability.
The IBM z16 builds on the RAS of the IBM z15™ family with the following RAS improvements:
System Recovery Boost
– System Recovery Boost was introduced with IBM z15. It offers customers more Central
Processor (CP) capacity during system recovery operations to accelerate the startup
(IPL), shutdown, or stand-alone memory dump operations (at image level - LPAR11).
System Recovery Boost requires operating system support. No other IBM software
charges are made during the boost period.
System Recovery Boost can be used during LPAR shutdown or start to make the
running operating system and services available in a shorter period.
The System Recovery Boost provides the following options for the capacity increase:
• Subcapacity CP speed boost: During the boost period, subcapacity engines that
are allocated to the boosted LPAR are transparently activated at their full capacity
(CP engines).
• zIIP Capacity Boost: During the boost period, all active zIIPs that are assigned to an
LPAR are used to extend the CP capacity (CP workload is dispatched to zIIP
processors during the boost period).
9
For more information, see ITIC Global Server Hardware, Server OS Reliability Report.
10
For more information, see Quantifying the Business Value of IBM Z.
11 LPAR that is running an Operating System image.
12 z/OS that is configured as a guest system under z/VM does not use the boost.
Naming: The IBM z16 Model A01, Machine Type (M/T) 3931, is further identified in this
document as IBM z16, unless otherwise specified.
The redesigned CPC drawer and I/O infrastructure also lower power consumption, reduces
the footprint, and allows installation in virtually any data center. The IBM z16 server is rated
for ASHRAE class A31 data center operating environment.
The IBM z16 server is similar to IBM z15, but differentiates itself from previous IBM Z server
generations through the following significant changes to the modular hardware:
All external cabling (power, I/O, and management) is performed at the rear of the system
Flexible configurations: Frame quantity is determined by the system configuration (1 - 4
frames)
Choice of power Intelligent Power Distribution Unit (iPDU or PDU) or Bulk Power Assembly
(BPA)
Feature codes that reserve slots for plan-ahead CPC drawers
Internal water cooling plumbing for systems with more than three CPC drawers (Frames A
and B)
PCIe+ Gen3 I/O drawers (19-inch format) supporting 16 PCIe adapters
The power options include PDU-based power or BPA-based power. The IBM z16 A01 system
is designed as a radiator (air) cooled system.
The IBM z16 A01 includes the following basic hardware building blocks:
19-inch 42u frame (1 - 4)
CPC (Processor) drawers (1 - 4)
PCIe+ Gen3 I/O drawers (up to 12)
Radiator cooling assembly (RCA) for CPC drawers cooling
Power, with choice of:
– Intelligent Power Distribution Units (iPDU) pairs (2 - 4 per frame, maximum 8,
depending on the configuration).
– Bulk Power Regulators (1 - 6 pairs, depending on the configuration)
Support Elements combined with optional Hardware Management Appliance (two):
– Single Keyboard, Mouse, Monitor (KMM) device (USB-C connection)
– Optional IBM Hardware Management Appliance feature
24-port 1 GbE Switches (two or four, depending on the system configuration)
Hardware for cable management at the rear of the system
1
For more information, see Chapter 2, Environmental specifications in IBM 3931 Installation Manual for Physical
Planning, GC28-7015.
Z A B C
I/O 9 I/O 12
I/O 4
1-4 Frames
I/O 8 I/O 11 2 or 4 GbE Switches
2 SEs
1-4 CPCs
CPC2 0-12 I/Os
I/O 7 I/O 10 I/O 15 2,4,6,8 PDUs
CPC1 1 or 2 RCAs
An example of a BPA-based powered system, and a maximum of four CPC drawers and 10
PCIe+ I/O drawers is shown in Figure 2-2.
2981 CPC1 reserve Reserve A15 location for future add CPC1
(Max39 to Max82 upgrade)
2982 CPC2 reserve Reserve A20 location for future add CPC2
(Max82 to Max125 upgrade)
PDU power
0646 380 - 415 V 32A 3 Phase (Wye) Worldwide (except North America and Japan)
BPA power
3017 Bulk Power Regulator (BPR) Quantity per BPA feature: Min. 2, max 6 (in
pairs)
I/O
7816 Top Exit Cabling without Tophat Uses rear slide plates at top of frame
Power Options
The 3931 (z16 A01) has the power options shown in Table 2-2 on page 19:
Water Cooling No No
DC Power available No No
a. Wye cords require five wires, three for Power phases, one for Neutral and one for Ground.
Caution: Installation of a low voltage system to a high voltage facility will cause significant
damage to the system’s power components.
Considerations
Consider the following points:
A-Frame is always present in every configuration
1u Support Elements (x2) are always in A-Frame at locations A41 and A42
1u 24-port internal Ethernet switches (x2) are always at locations A39 and A40
More Ethernet switches (x2) are available when necessary in Frame C or B
I/O PCHID numbering starts with 0100 and increments depending on the number of
features that is ordered. No PCHID number affinity is available to a fixed PCIe+ I/O drawer
location as with previous systems.
All external I/O cabling enters the system at the rear of the frames for all adapters,
management LAN, and power connections.
The Top Exit feature code (FC 7898) provides an optional Top Exit cover enclosure. The
optional Top Exit cover enclosure provides a fiber cable organizer within the enclosure to
optimize the fiber cable storage and strain relief. It also provides mounting locations to secure
Fiber Quick Connector (FQC) MPO2 brackets (FC 5827) on the top of the frames.
Note: Overhead I/O cabling is contained within the frames. Extension “chimneys” that were
featured with systems before IBM z15 systems are no longer used.
A view of the top rear of the frame and the openings for top exit cables and power is shown in
Figure 2-4. When FC 7898 is installed, the plated adjustable shields are removed, and the top
exit enclosure is installed.
Care must be taken when ordering the Feature Codes (see Table 2-3) for cables that are
entering the frames from above the floor, below the floor, or both, and if the Top Exit feature is
wanted, which provides Top Hat.
Table 2-3 IBM z16 Model A01 cabling Feature Code combinations
Environment Bottom exit Top exit Feature Code Comments
Raised Floor Yes Yes (no Top Hat) 7899 and 7816 Bottom FQC support only
Raised Floor Yes Yes (with Top Hat) 7899 and 7898 Top and Bottom FQC support
Raised Floor No Yes (with Top Hat) 7898 Top FQC support
Ships bottom seal plate
Non-Raised Floor No Yes (no Top Hat) 7998 and 7816 No FQC support
Ships bottom seal plate
Non-Raised Floor No Yes (with Top Hat) 7990 and 7898 Top FQC support
Ships bottom seal plate
A vertical cable management guide (“spine”) can assist with proper cable management for
fiber, copper, and coupling cables. A top to bottom spine is present from manufacturing with
cable organizer clips that are installed for frames Z and C when present. Frames A and B
contain mini-spines that serve the same purpose.
The cable retention clips can be relocated for best usage. All external cabling to the system
(from top or bottom) can use the spines to minimize interference with the PDUs that are
mounted on the sides of the rack.
Figure 2-5 I/O cable management spine and fiber cable organizer (rear view)
The IBM z16 Model A01 5u CPC drawer always contains four Processor Unit (PU) DCMs,
and up to 48 memory DIMMs.
Depending on the feature, the IBM z16 A01 contains the following CPC components:
The number of CPC drawers installed is driven by the following feature codes:
– FC 0667: One CPC drawer, Max39, up to 39 characterizable PUs
– FC 0668: Two CPC drawers, Max82, up to 82 characterizable PUs
– FC 0669: Three CPC drawers, Max125, up to 125 characterizable PUs
– FC 0670: Four CPC drawers, Max168, up to 168 characterizable PUs
– FC 0671: Four CPC drawers, Max200, up to 200 characterizable PUs
The following Processor Unit DCM is used:
PU DCM contains two PU chips (Telum) on one single module that use 7 nm silicon wafer
technology, 22.5 billion transistors, and a core that is running at 5.2 GHz (designed with 8
cores per chip, 16 cores per PU DCM).
Memory plugging:
– Six memory controllers per drawer (two each on DCM2/DCM1; one each on
DCM3/DCM0
– Each memory controller supports eight DIMM slots
– All eight DIMMs on one memory controller are the same size
– 4 - 6 memory controllers per drawer are populated (up to 48 DIMMs)
– Different memory controllers can have different size DIMMs
The front view of the CPC drawer, which includes the cooling fans, BMC/OSC and processor
power cards (PPC), is shown in Figure 2-7.
Dual port I/O fan-outs and ICA SR adapters are plugged in specific slots for best performance
and availability. Redundant power supplies and six SMP9 ports also are shown. Each pair of
SMP9 ports is redundant. If a single cable fails, the repair can be performed concurrently.
The CPC drawer logical structure, component connections are shown in Figure 2-9.
Memory is connected to the DCMs through memory control units (MCUs). Up to six MCUs
are available in a CPC drawer (one or two per DCM) and provide the interface to the DIMM
controller. A memory controller uses eight DIMM slots.
The CPC drawers that are installed in Frame A and Frame B are populated from bottom to
top.
On IBM z16, the external time source (PTP or NTP) is connected directly to the CPC and
bypasses the SE connection. This IBM z16 feature allows more accurate sync with the
external time source.
The accuracy of an STP-only CTN is improved by using an NTP or PTP server with the PPS
output signal as the External Time Source (ETS). Devices with PPS output are available from
several vendors that offer network timing solutions.
Tip: STP is available as FC 1021. It is implemented in the Licensed Internal Code (LIC),
and allows servers to maintain time synchronization with each other and synchronization to
an ETS. In a multi-server STP Coordinated Timing Network (CTN) coupling/timing links are
required for STP communication
For more information, see IBM Z Server Time Protocol Guide, SG24-8480.
With IBM z16 Model A01, the CPC drawer BMC is combined with the Oscillator card in a
single Field Replaceable Unit (FRU). Two combined BMC/OSC cards are used per CPC
drawer.
Also, the PCIe+ I/O drawer has an improved BMC. Each BMC card has one Ethernet port that
connects to the internal Ethernet LANs through the internal network switches (SW1, SW2,
and SW3, SW4, if configured). The BMCs communicate with the SEs and provide a
subsystem interface (SSI) for controlling components.
Note: The maximum IBM z16 A01 system configuration features four GbE switches, four
CPC drawers, and up to 12 PCIe I/O drawers.
A typical BMC operation is to control a power supply. An SE sends a command to the BMC to
start the power supply. The BMC cycles the various components of the power supply,
monitors the success of each step and the resulting voltages, and reports this status to the
SE.
SEs are duplexed (N+1), and each element has at least one BMC. Two internal Ethernet LAN
switches and two SEs, for redundancy, and crossover connectivity between the LANs, are
configured so that both SEs can operate on both LANs.
The Hardware Management Consoles (HMCs) and SEs are connected directly to one or two
Ethernet Customer LANs. One or more HMCs can be used.
With IBM z16 current ordered systems, the Hardware Management Appliance (HMA, FC
0129) is the only HMC-orderable feature. The HMA is running on the same hardware (virtual
appliance) with the Support Elements (the SE code runs as a guest of the HMA).
The PU DCM is shown in Figure 2-13. The DCM features a thermal cap that is placed over
the chips. Each DCM is water-cooled by way of a cold plate manifold assembly.
PU chips use 7 nm FinFET process, each DCM socket contains 4753 pins, and the module
size is 71.5 mm x 79 mm.
The DCMs are each plugged into a socket that is part of the CPC drawer packaging. Each
DCM is cooled by water flowing through a manifold assembly where cold plates are attached
and used to help remove the heat from the processor chips.
The IBM z16 Model A01 PU chip (two chips packaged on each DCM) includes the following
features and improvements:
7 nm silicon lithography wafer technology
530 mm2 chip size
22.5 billion transistors
5.2 GHz base clock frequency
Eight core design (versus 12 for IBM z15) with increased on-chip cache sizes
Two PCIe Gen4 interfaces
DDR4 memory controller
2x M-Bus to support DCM internal chip to chip connectivity
6x X-Bus to support DCM to DCM connectivity in the CPC drawer
1x A-Bus to support drawer to drawer connectivity
New cache structure design compared to IBM z15 PU:
– L1D(ata) and L1I(nstruction) cache - ON-core (128 kB each)
– L2 - 32 MB dense SRAM - outside the core, semi-private to the core
– L3 (virtual) - up to 256 MB
– L4 (virtual) - up to 2048 MB
New Core-Nest Interface compared to IBM z15:
2.3.3 PU characterization
The PUs are characterized for client use. The characterized PUs can be used in general to
run supported operating systems, such as z/OS, z/VM, and Linux on IBM Z. They also can
run specific workloads, such as Java, XML services, IPsec, and some Db2 workloads, or
clustering functions, such as the Coupling Facility Control Code (CFCC).
The maximum number of characterizable PUs depends on the IBM z16 Model A01 CPC
drawer feature code. Some PUs are characterized for system use; some are characterized for
client workload use.
By default, one spare PU is available to assume the function of a failed PU. The maximum
number of PUs that can be characterized for client use are listed in Table 2-5.
Max39 0 - 39 0 - 39 0 - 38 0 - 38 0 - 39 2 5 0-8 2
Max82 0 - 82 0 - 82 0 - 81 0 - 81 0 - 82 10
The rule for the CP to zIIP purchase ratio, that for every CP purchased, up to two zIIPs can be
purchased has been removed on z16.
Converting a PU from one type to any other type is possible by using the Dynamic Processor
Unit Reassignment process. These conversions occur concurrently with the system
operation.
Note: The addition of ICFs, IFLs, zIIPs, and SAPs to the IBM z16 Model A01 does not
change the system capacity setting or its million service units (MSU) rating.
MEMORY
Virtual L4 1792 MB
DCM 0 DCM 3
CHIP 0 CHIP 1
4 DCMs CHIP 1
CHIP 0
MBUS MBUS
L2
ON-CHIP L2 L2 ON-CHIP L2 L2 ON-CHIP L2 L2 ON-CHIP L2
32 MB RING RING RING RING
Figure 2-16 Cache structure comparison: IBM z15 versus IBM z16
A total of 16 I/O cards are spread over two I/O domains (0 and 1):
– Each I/O slot reserves four PCHIDs.
The following configuration examples and how the configurations are different from the power
selection (PDU or BPA), number of CPC drawers and I/O features ordered, the layout of the
PCIe+ I/O drawers, and PCHID numbering are shown in Figure 2-18:
The first, a single frame system, ordered with PDU power, radiator cooling, one CPC
drawer, and greater than 32 I/O features to drive three I/O drawers.
PCHID numbering is consecutive from top to bottom.
The vertical card collapsed horizontal and the awareness of the port and PCHID layout where
the top of the adapter (port D1) is closest to the location panel on both sides of the drawer are
shown in Figure 2-19.
Figure 2-19 I/O feature orientation in PCIe I/O drawer (rear view)
Note: The CHPID Mapping Tool (available on ResourceLink) can be used to print a CHPID
Report that displays the drawer and PCHID/CHPID layout.
Note: IBM z16 with Bulk Power Adapter option can be configured with only 5 TB of
memory per CPC drawer (BPA systems maximum memory is 20 TB).
The minimum and maximum memory sizes that you can order for each IBM z16 Model T01
feature are listed in Table 2-6.
The memory granularity, which is based on the installed customer memory, is listed in
Table 2-7.
64 512 - 768
The CPC drawer memory topology of an IBM z16 is shown in Figure 2-20.
Servers are configured with the most efficient configuration of memory DIMMs that can
support Addressable Memory that is required for Customer Ordered Memory plus HSA. In
some cases, Available Addressable Memory might be available to support one or more
concurrent LIC CC Customer Memory upgrades with no DIMM changes.
Supported drawer memory configurations are listed in Table 2-8. Each CPC drawer is
included from manufacturing with one of these memory configurations.
22 0 32 32 32 32 0 1024 768
DIMM plugging for the configurations in each CPC drawer do not have to be similar. Each
memory 8-slot DIMM bank must have the same DIMM size; however, a drawer can have a
mix of DIMM banks.
The support element View Hardware Configuration task can be used to determine the size
and quantity of the memory plugged in each drawer. Figure 2-21 on page 41 shows an
example of configuration number 27 from the previous tables, and displays the location and
description of the installed memory modules.
Table 2-9 on page 42 lists the physical memory plugging configurations by feature code from
manufacturing when the system is ordered. Consider the following points:
The CPC drawer columns for the specific feature contain the Memory Plug Drawer
Configuration number that us referenced in Table 2-8 on page 39 and the Population by
DIMM Bank that is listed in Table 2-9 on page 42.
Dial Max indicates the maximum memory that can be enabled by way of the LICC
concurrent upgrade.
If more storage is ordered by using other feature codes, such as Virtual Flash Memory or
Flexible Memory, the extra storage is installed and plugged as necessary.
For example, a customer orders FC 2857 that features 9472 GB customer memory and FC
0669 Max125 (3 CPC drawers). The drawer configurations include the following components:
CPC0 (3584 GB), CPC2 (3584 GB) - Configuration #27 and CPC1 (3072 GB) -
Configuration #26 (from Table 2-8 on page 39)
Total 3584 + 3584 + 3072 - 256 HSA= 9984 GB (Dial Max)
Increments
Increments
Memory
Customer
Dial Dial Dial Max Dial
Drawer CPC0
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Max Max Max
Increments
Increments
Memory
Customer
Dial Dial Dial Max Dial
Drawer CPC0
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Max Max Max
Increments
Increments
Memory
Customer
Dial Dial Dial Max Dial
Drawer CPC0
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Max Max Max
For a model upgrade that results in the addition of a CPC drawer, the minimum memory
increment is added to the system. Each CPC drawer has a minimum physical memory size of
1024 GB.
Note: Memory downgrades within an IBM z16 are not supported. Feature downgrades
(removal of a CPC quantity feature) are not supported.
For more information, see 2.7.1, “Redundant I/O interconnect” on page 52.
Removing a CPC drawer often results in removing active memory. With the flexible memory
option, removing the affected memory and reallocating its use elsewhere in the system is
possible. (For more information, see 2.5.7, “Flexible Memory Option” on page 46.) This
process requires more available memory to compensate for the memory that becomes
unavailable with the removal of the drawer.
No application changes are required to change from IBM Flash Express to VFM. Consider the
following points:
Dialed memory + VFM = total hardware plugged
Dialed memory + VFM + Flex memory option = total hardware plugged
VFM helps improve availability and handling of paging workload spikes when z/OS V2.1,
V2.2, V2.3, V2.4, or V2.5 is run. With this support, z/OS helps improve system availability and
responsiveness by using VFM across transitional workload events, such as market openings
and diagnostic data collection. z/OS also helps improve processor performance by supporting
middleware use of pageable large (1 MB) pages and to help eliminate delays that can occur
when collecting diagnostic data.
VFM also can be used by coupling facility images to provide extended capacity and
availability for workloads that use IBM WebSphere MQ Shared Queues structures.
VFM can help organizations meet their most demanding service level agreements and
compete more effectively. VFM is easy to configure in the LPAR Image Profile and provides
rapid time to value.
4
CPC 1 and CPC 2 can be added concurrently in the field if FC 2981 and FC 2982 are ordered with the initial
configuration. The addition of a fourth CPC Drawer (CPC 3) is not supported. Four CPC drawer systems are
factory-built only.
When you order memory, you can request extra flexible memory. The extra physical memory,
(if required) is calculated by the configuration and priced accordingly.
The required memory DIMMs are installed before shipping and are based on a target
capacity that is specified by the customer. These memory DIMMs are enabled by using a
Licensed Internal Code Configuration Control (LICCC) order that is placed by the customer
when they determine more memory capacity is needed.
The flexible memory sizes that are available for the IBM z16 A01 are listed in Table 2-10.
Increments
Increments
Memory
Customer
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Max Max Max
Code
Feature
Increments
Increments
Memory
Customer
Dial Dial Dial
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Max Max Max
Code
Feature
Increments
Increments
Memory
Customer
Dial Dial Dial
Drawer CPC0
Drawer CPC1
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC0
Drawer CPC1
Drawer CPC2
Drawer CPC3
Max Max Max
The IBM Z hardware has decades of intense engineering behind it, which results in a robust
and reliable platform. The hardware has many RAS features that are built into it.
For more information, see Chapter 9, “Reliability, availability, and serviceability” on page 379.
The internal intelligent Power Distribution Unit (iPDU) provides the following capabilities:
Individual outlet control by way of Ethernet:
– Provide a System Reset capability
– Power cycle an SE if a hang occurs
– Verify a power cable at installation
System Reset Function:
– No physical EPO switch is available on the IBM z16, which provides a means for the
service technician6 to put a server into a known state.
5
The air density sensor measures air pressure and is used to control blower speed.
6 This function is available to IBM System Service Representatives (SSRs) only.
The power service and control network (PSCN) is used to control and monitor the elements in
the system and include the following components:
Ethernet Top of Rack (TOR) switches provide the internal PSCN connectivity:
– Switches are redundant (N+1)
– Concurrently maintainable
– Each switch has one integrated power supply
– BMCs are cross wired to the Ethernet switches
Redundant SEs
Each SE has two power supplies (N+1) and input power is cross-coupled from the PDUs.
Concurrent CPC upgrades
CPC1 to (CPC1 + CPC2) and (CPC1 + CPC2) to (CPC1+CPC2+CPC3) if CPC1 Reserve
or CPC2 Reserve features are part of the initial system order (FC 2981 or FC 2982)
All PCIe+ I/O drawer MESs are concurrent
All LICC model changes are concurrent
Service restoration involves speeding up IPL and shutdown operations of an image (LPAR),
and short-duration recovery process boosts for specific sysplex and operating system events.
For more information see Introducing IBM Z System Recovery Boost, REDP-5563.
IBM z16 servers continue to deliver robust server designs through new technologies,
hardening innovative and classic redundancy. For more information, see Chapter 9,
“Reliability, availability, and serviceability” on page 379.
2.7 Connectivity
Connections to PCIe+ I/O drawers and Integrated Coupling Adapters are driven from the CPC
drawer fan-out cards. These fan-outs are installed in the rear of the CPC drawer.
Figure 2-22 shows the location of the fan-out slots. Each slot is identified with a location code
(label) of LGxx.
A fan-out can be repaired concurrently with the use of redundant I/O interconnect. For more
information, see 2.7.1, “Redundant I/O interconnect” on page 52.
When configured for availability, the channels and coupling links are balanced across CPC
drawers. In a system that is configured for maximum availability, alternative paths maintain
access to critical I/O devices, such as disks and networks. The CHPID Mapping Tool can be
used to assist with configuring a system for high availability.
EDA allows a single CPC drawer in a multidrawer CPC to be removed and reinstalled
(serviced) concurrently for an upgrade or a repair. Removing a CPC drawer means that the
connectivity to the I/O devices that are connected to that CPC drawer is lost. To prevent
connectivity loss, the redundant I/O interconnect feature allows you to maintain connection to
critical I/O devices installed in PCIe+I/O Drawers when a CPC drawer is removed. ICA SR1.1
(and ICA SR) can also be installed for coupling and timing links redundancy.
The PCIe+ I/O drawer supports up to 16 PCIe features, which are organized in two hardware
domains (for each drawer). The infrastructure for the fan-out to I/O drawers and external
coupling is shown in Figure 2-23.
Figure 2-23 Infrastructure for PCIe+ I/O drawer (system with two PCIe+ I/O drawers)
The PCIe+ Gen3 fan-out cards are used to provide the connection from the PU DCM PCIe
Bridge Unit (PBU), which uses split the PCIe Gen4 (@32GBps) processor busses into two
PCIe Gen3 x16 (@16 GBps) interfaces to the PCIe switch card in the PCIe+ I/O drawer.
In the PCIe+ I/O drawer, the two PCIe switch cards (LG06 and LG16, see Figure 2-24)
provide a backup path (Redundant I/O Interconnect [RII]) for each other through the passive
connection in the PCIe+ I/O drawer backplane.
During a CPC Drawer PCIe+ Gen3 fan-out or cable failure, all 16 PCIe cards in the two
domains can be driven through a single PCIe switch card (see Figure 2-25).
Note: The PCIe Interconnect (switch) adapter must be installed in the PCIe+ I/O drawer to
maintain the interconnect across I/O domains. If the adapter is removed (for a service
operation), the I/O cards in that domain (up to eight) become unavailable.
Before removing the CPC drawer, the contents of the PUs and memory of the drawer must be
relocated. PUs must be available on the remaining CPC drawers to replace the deactivated
drawer. Also, sufficient redundant memory must be available if no degradation of applications
is allowed.
To ensure that the CPC configuration supports removal of a CPC drawer with minimal effect
on the workload, consider the flexible memory option. Any CPC drawer can be replaced,
including the first CPC drawer that initially contains the HSA.
Removal of a CPC drawer also removes the CPC drawer connectivity to the PCIe I/O
drawers, and coupling links. The effect of the removal of the CPC drawer on the system is
limited by the use of redundant I/O interconnect. (For more information, see 2.7.1,
“Redundant I/O interconnect” on page 52.) However, all ICA SR1.1 links that are installed in
the removed CPC drawer must be configured offline.
If the enhanced drawer availability and flexible memory options are not used when a CPC
drawer must be replaced, the memory in the failing drawer also is removed. This process
might be necessary during an upgrade or a repair action.
Until the removed CPC drawer is replaced, a power-on reset of the system with the remaining
CPC drawers is supported. The CPC drawer then can be replaced and added back into the
configuration concurrently.
A minimum of one PU that is characterized as a CP, IFL, or ICF is required per system. The
maximum number of characterizable PUs is 200. The maximum number of zIIPs is up to twice
the number of PUs that are characterized as CPs.
The following components are present in the IBM z16, but they are not part of the PUs that
clients purchase and require no characterization:
SAP to be used by the channel subsystem. The number of predefined SAPs depends on
the IBM z16 model.
Two IFPs, which are used in the support of designated features and functions, such as
RoCE (all features), Coupling Express2 LR, zHyperLink Express 1.1, Internal Shared
Memory (ISM) SMC-D, and other management functions.
Two spare PUs, which can transparently assume any characterization during a permanent
failure of another PU.
drawer
PUs per
SAPs
Opt
SAPs
Base
Spares
CPs IFLs ICFs uIFLs
Max39 1 48 0 - 39 0 - 39 0 - 39 0 - 38 0 - 25 2 0-8 5 2
Max82 2 48 0 - 82 0 - 82 0 - 82 0 - 81 0 - 54 2 10 2
Not all PUs available on a model are required to be characterized with a feature code.
Only the PUs purchased by a customer are identified with a feature code.
zIIP maximum quantity for new build systems follows the 2:1 ratio. It might be greater if
present during MES upgrades.
All PU conversions can be performed concurrently.
A capacity marker identifies the number of CPs that were purchased. This number of
purchased CPs is higher than or equal to the number of CPs that is actively used. The
capacity marker marks the availability of purchased but unused capacity that is intended to be
used as CPs in the future. This status often is present for software-charging reasons.
Unused CPs are not a factor when establishing the millions of service units (MSU) value that
is used for charging monthly license charge (MLC) software, or when charged on a
per-processor basis.
2.8.1 Upgrades
Concurrent upgrades of CPs, IFLs, ICFs, zIIPs, or SAPs are available for the IBM z16.
However, concurrent PU upgrades require that more PUs are installed but not activated.
Spare PUs are used to replace defective PUs. Two spare PUs always are on an IBM z16 A01
server. In the rare event of a PU failure, a spare PU is activated concurrently and
transparently and is assigned the characteristics of the failing PU.
Max39 N Y Y N N
Max82 N N Y N N
Max125 N N N N N
Max168 N N N N N
Max200 N N N N N
Characterization of a PU as an IFL, ICF, or zIIP is not reflected in the output of the STSI
instruction because characterization has no effect on software charging. For more information
about STSI output, see “Processor identification” on page 374.
The following distinct model capacity identifier ranges are recognized (one for full capacity
and three for granular capacity):
For full-capacity engines, model capacity identifiers 701 - 7K0 are used. They express
capacity settings for 1 - 200 characterized CPs.
Three model subcapacity identifier ranges offer a unique level of granular subcapacity
engines at the low end. They are available for up to 39 characterized CPs. These three
subcapacity settings are applied to up to 39 CPs, which combined offer 117 more capacity
settings, as described next.
Granular capacity
The IBM z16 A01 server offers 117 capacity settings (granular capacity) for up to 39 CPs.
When subcapacity settings are used, PUs beyond 39 can be characterized only as specialty
engines. For models with more that 39 CPs, all CPs are running at full capacity (7xx).
The three defined ranges of subcapacity settings include model capacity identifiers numbered
401- 439, 501 - 539, and 601 - 639.
Consideration: All CPs have the same capacity identifier. Specialty engines (IFLs, zIIPs,
and ICFs) operate at full speed.
Max39 701 - 739, 601 - 639, 501 - 539, and 401 - 439
Max82 701 - 782, 601 - 639, 501 - 539, and 401 - 439
Max125 701 - 7C5, 601 - 639, 501 - 539, and 401 - 439
Max168 701 - 7G8, 601 - 639, 501 - 539, and 401 - 439
Max200 701 - 7K0, 601 - 639, 501 - 539, and 401 - 439
For more information about temporary capacity increases, see Chapter 8, “System upgrades”
on page 327.
The PDUs are controlled by using an Ethernet port and support the following input:
3-phase 200 - 240 V AC (4-wire “Delta”)
3-phase 380 - 415 V AC (5-wire “Wye”)
The power supply units convert the AC power to DC power that is used as input for the Point
of Loads (POLs) in the CPC drawer and the PCIe+ I/O drawers.
The power requirements depend on the number of CPC drawers (1 - 4), number of PCIe I/O
drawers (0 - 12) and I/O features that are installed in the PCIe I/O drawers.
PDUs are installed and serviced from the rear of the frame. Unused power ports never should
be used by any external device.
Each PDU installed requires a customer-supplied power feed. The number of power cords
that are required depends on the system configuration.
Note: For initial installation, all power sources are required to run the system checkout
diagnostics successfully.
PDUs are installed in pairs. A system can have 2, 4, 6, or 8 PDUs, depending on the
configuration. Consider the following points:
Paired PDUs are A1/A2, A3/A4, B1/B2, and C1/C2.
From the rear of the system, the odd-numbered PDUs are on the left side of the rack; the
even-numbered PDUs are on the right side of the rack.
The total loss of one PDU in a pair does not affect the system operation.
Components that plug into the PDUs for redundancy (by using two power cords) include the
following features:
CPC Drawers, PCIe+ I/O drawers, Radiators, and Support Elements
The redundancy for each component is achieved by plugging the power cables into the
paired PDUs.
For example, the top Support Element (1), has one power supply plugged into PDU A1
and the second power supply plugged into the paired PDU A2 for redundancy.
As a best practice, connect the odd-numbered PDUs (A1, B1, C1, and D1) to one power
source or distribution panel, and the even-numbered PDUs (A2, B2, C2, and D2) to a
separate power source or distribution panel.
The frame count rules (number of frames) for IBM z16 A01 are listed in Table 2-14.
CPC drawers 0 1 2 3 4 5 6 7 8 9 10 11 12
1 1 1 1 1 2 2 2 2 2 3 3 3 3
2 1 1 1 2 2 2 2 2 3 3 3 3 3
3 1 1 2 2 2 2 2 3 3 3 3 3 N/A
4 2 2 2 2 2 3 3 3 3 3 4 4 4
The number of CPC drawers and I/O drawers determines the number of racks in the system
and the number of PDUs in the system.
The PDU/line cord rules (number of PDU/Cord pairs) for IBM z16 A01 are listed in Table 2-15.
Table 2-15 PDU/line cord rules (# PDU/Cord pairs) for IBM z16 A01
PDU/Linecord I/O drawers
CPC drawers 0 1 2 3 4 5 6 7 8 9 10 11 12
1 1 1 1 1 2 2 2 2 2 3 3 3 3
2 2 2 2 2 2 2 2 2 3 3 3 3 3
3 2 2 2 2 2 2 2 3 3 3 3 3 N/A
4 3 3 3 3 3 3 3 3 3 4 4 4 4
No PDUs are installed in this configuration; instead, 1 or 2 pairs of BPAs are installed that
house the bulk power distribution to all the components, and the Bulk Power Regulators
(BPR).
The BPAs support 2- or 4-line cords with a single universal type of 3-phase 200 - 480 V AC.
The power requirements depend on the number of CPC drawers (1 -4), number of PCIe I/O
drawers (0 - 10), and I/O features that are installed in the PCIe+ I/O drawers.
A schematic view of a maximum configured system with BPA power is shown in Figure 2-28.
Each installed BPA requires a customer-supplied power feed. The number of power cords that
are required depends on the system configuration. The total loss of one BPA does not affect
system operation.
Note: For initial installation, all power sources are required to run the system checkout
diagnostics successfully.
BPAs are installed in pairs (2 or 4), depending on the configuration. Consider the following
points:
BPA1 and BPA2 are always initially installed in Frame A.
BPA3 and BPA4 are installed to support more components for larger configurations (CPC
and I/O drawers).
BPA1/BPA2 and BPA3/BPA4 for paired redundancy.
From the rear of the system, the odd-numbered BPAs are above the even-numbered BPAs
in both frames.
The total loss of one BPA in a pair has no effect on system operation.
Note: Customer power sources should always maintain redundancy across BPA pairs; that
is, one power source or distribution panel supplies power for BPA1 and the separate power
source or distribution panel supplies power for BPA2.
As a best practice, connect the odd-numbered BPAs (BPA1 and BPA3) to one power
source or distribution panel, and the even-numbered BPAs (BPA2 and BPA4) to a separate
power source or distribution panel.
This tool estimates the power consumption for the specified configuration. The tool does not
verify that the specified configuration can be physically built.
Tip: The exact power consumption for your system varies. The object of the tool is to
estimate the power requirements to aid you in planning for your system installation. Actual
power consumption after installation can be confirmed by using the HMC Monitors
Dashboard task.
2.9.4 Cooling
The PU DCMs for IBM z16 A01 are cooled by a cold plate that is connected to the internal
water-cooling loop. In an air-cooled system, the radiator unit dissipates the heat from the
internal water loop with air. The radiator unit provides improved availability with N+ 1 pumps
and blowers.
For all IBM z16 A01 servers, the CPC drawer components (except for PU DCMs) and the
PCIe+ I/O drawers are air cooled by redundant fans. Airflow of the system is directed from
front (cool air) to the back of the system (hot air).
For more information, see 2.9.5, “Radiator Cooling Unit” on page 64.
Although the PU DCMs are cooled by water, the heat is exhausted into the room from the
radiator heat exchanger by forced air with blowers. At the system level, these IBM z16 A01
are still air-cooled systems.
The RCU discharges heat from the internal frame water loop to the customer’s data center.
Each RCU provides cooling to PU DCMs with closed loop water within the respective frame.
No connection to an external chilled water supply is required. For the IBM z16 A01 server, the
internal circulating water is conditioned water (BTA) that is added to the radiator unit during
system installation with the Fill and Drain Tool (FC 3393).
The new design uses a pump for filling the system and FRUs, and a compressor for draining
FRUs and discontinuance. The process is faster and more effective. The FDT is shown in
Figure 2-29.
The RCU contains two independent pump FRUs. Because the cooling capability is a
redundant N+1 design, a single working pump and blower can support the entire load. The
replacement of one pump or blower can be done concurrently and does not affect
performance.
The water pumps, manifold assembly, radiator assembly (which includes the heat exchanger),
and fan assemblies are the main components of the IBM z16 RCU, as shown in Figure 2-30.
The closed water loop in the radiator unit is shown in Figure 2-31. The warm water that is
exiting from the PU DCMs cold plates enters pumps through a common manifold and is
pumped through a heat exchanger where heat is extracted by the air flowing across the heat
exchanger fins. The cooled water is then recirculated back into the PU DCMs cold plates.
DCM
Standard SAPs 5 10 15 20 24
Number of IFP 2 2 2 2 2
Enabled Memorya sizes GB 512 - 9984 512 - 20224 512 - 30464 512 - 40704 512 - 40704
Flexible memorya sizes GB N/A 512 - 9984 512 - 20224 512 - 30464 512 - 30464
Clock frequency 5.2 GHz 5.2 GHz 5.2 GHz 5.2 GHz 5.2 GHz
Note: The IBM z16 Model A01, Machine Type (M/T) 3931 (M/T 3931), is further identified
in this document as IBM z16, unless otherwise specified.
For more information about the processor unit, see z/Architecture Principles of Operation,
SA22-7832.
IBM Z servers offer high levels of reliability, availability, serviceability (RAS), resilience, and
security. The IBM z16 fits into the IBM strategy in which mainframes play a central role in
creating an infrastructure for cloud, artificial intelligence, and analytics, which is underpinned
by security. The IBM z16 server is designed so that everything around it, such as operating
systems, middleware, storage, security, and network technologies that support open
standards, helps you achieve your business goals.
The IBM z16 extends the platform’s capabilities and adds value with breakthrough
technologies, such as the following examples:
On-chip Artificial Intelligence (AI) at speed and scale that is designed to leave no
transaction behind.
An industry-first system that uses quantum-safe technologies, cryptographic discovery
tools, and end-to-end data encryption to protect against future attacks now.
A continuous compliance solution to help keep up with changing regulations, which
reduces cost and risk.
A consistent cloud experience to enable accelerated modernization, rapid delivery of new
services, and end-to-end automation.
New options in flexible and responsible consumption to manage system resources across
geographical locations, with sustainability that is built in across its lifecycle.
The modular CPC drawer design aims to reduce (or in some cases even eliminate) planned
and unplanned outages. The design does so by offering concurrent repair, replace, and
upgrade functions for processors, memory, and I/O.
For more information about the IBM z16 RAS features, see Chapter 9, “Reliability, availability,
and serviceability” on page 379.
For more information about frames and configurations, see Chapter 2, “Central processor
complex hardware components” on page 15, and Appendix E, “Frame configurations Power
Distribution Units and Bulk Power Assembly-based systems” on page 529.
The modular CPC drawer design is flexible and expandable. It offers unprecedented capacity
and security features to meet consolidation needs.
IBM z16 servers CPC continues the line of mainframe processors that are compatible with an
earlier version. The IBM z16 brings the following processor design enhancements:
7 nm silicon lithography
Eight cores per PU chip design with 22.5 billion transistors per PU chip
Redesigned cache structure that is implemented in dense SRAM
Four PU Dual Chip Modules per CPC Drawer
Each PU chip features two PCIe Generation 4 ports (x16 @ 32 GBps)
Optimized pipeline
Improved SMT and SIMD
Improved branch prediction
Improved co-processor functions (CPACF)
IBM integrated accelerator for zEnterprise Data Compression (zEDC) (on-chip
compression accelerator)
IBM integrated accelerator for Z Sort (on-core sort accelerator)
IBM integrated accelerator for AI (on-chip AI accelerator)
Transparent memory encryption.
It uses 24-, 31-, and 64-bit addressing modes, multiple arithmetic formats, and multiple
address spaces for robust interprocess security.
The IBM z16 system design features the following main objectives:
Offer a data-centric approach to information (data) security that is simple, transparent, and
consumable (extensive data encryption from inception to archive, in-flight, and at-rest).
Offer a flexible infrastructure to concurrently accommodate a wide range of operating
systems and applications, from the traditional systems (for example, z/OS and z/VM) to
the world of Linux, cloud, analytics, and mobile computing.
Offer state-of-the-art integration capability for server consolidation by using virtualization
capabilities in a highly secure environment:
– Logical partitioning, which allows up to 85 independent logical servers.
– z/VM, which can virtualize hundreds to thousands of servers as independently running
virtual machines (guests).
1 Federal Information Processing Standard (FIPS) 140-2 Security Requirements for Cryptographic Modules.
The remaining sections in this chapter describe the IBM z16 system structure. It shows a
logical representation of the data flow from PUs, caches, memory cards, and various
interconnect capabilities.
Note: An IBM z16 server with Bulk Power Assembly (BPA) power option is limited to 5 TB
per processor drawer (up to 20 TB maximum in a full configuration with four CPC drawers).
The following types of CPC drawer configurations are available for IBM z16 A01:
One drawer: Max39
Two drawers: Max82
Three drawers: Max125
Four drawers: Max168
Four drawers: Max200
Note: Max168 and Max200 are factory-build only. It is not possible to upgrade in the field to
Max168 or Max200.
The IBM z16 A01 has up to 24 memory controller units (MCUs) for a Max200 feature (one
MCU per PU chip, and up to six MCUs populated per CPC drawer). The MCU configuration
uses eight channels Reed-Solomon redundant array of independent memory (RAIM).
The RAIM design is new compared to the IBM z15 T01, moving from a 4+1 DIMM structure
that is based on 5-channel RAIM design on IBM z15 to an 8-channel R-S RAIM design on the
IBM z16. The new memory architecture provides approximately 15% DRAM reduction for a
similar RAS, but a higher memory bandwidth at drawer level.
The DIMM sizes (32, 64, 128, or 256 GB) include RAIM overhead. An IBM z16 CPC drawer
can have up to 48 memory DDR4 DIMMs (populated with 32, 40, or 48 DIMMs).
The IBM z16 microprocessor chip (called IBM Telum) integrates a new cache hierarchy
design with only two levels of physical cache (L1 and L2). The cache hierarchy (L1, L2) is
implemented with dense static random access memory (SRAM).
Unlike the IBM z15, eDRAM is no longer used in the IBM Telum processor. On an IBM z16, L2
cache (32 MB) is semi-private with 16 MB dedicated to the associated core, and 16 MB
shared with the system (the 50/50 split is adjustable). Level 3 (L3) and Level 4 (L4) caches
are now virtual caches and are allocated on L2.
Two processor chips (up to eight active cores per PU chip) are combined in a Dual Chip
Module (DCM) and four DCMs are assembled in a CPC drawer (eight PU chips). An IBM z16
can have from one CPC drawer (Max39) up to four (Max200).
The new IBM z16 Dual Chip Module (DCM) is shown in Figure 3-2.
Concurrent maintenance allows dynamic central processing complex (CPAC) drawer add and
repair.2
IBM z16 processors use 7 nm extreme ultraviolet (EUV) lithography chip technology with
advanced low latency pipeline design, which creates high-speed yet power-efficient circuit
designs. The PU DCM has a dense packaging, which allows closed water loop cooling. The
heat from the closed loop is dissipated into the air by a radiator unit (RU).
On IBM z16, L1 and L2 caches are implemented on the PU chip, and L3 and L4 caches are
now implemented as virtual caches and dynamically allocated on the shared part of the L2
semi-private cache.
The cache structure of the IBM z16 features the following characteristics:
Large L1, L2 caches (more data closer to the core).
L1 cache is implemented by using SRAM technology and has the same size than on IBM
z15 (128 KB for instructions and 128 KB for data).
L2 cache (32 MB in total) uses SRAM technology, and is semi-private to each PU core
with 16 MB dedicated to the associated core, and 16 MB shared with the system (the
50/50 split is adjustable).
L3 cache (up to 256 LB) now becomes a virtual cache and can be allocated on any of the
share part of a L2 cache.
L4 cache (up to 2048 MB) is also a virtual cache and can be allocated on any of the share
part of a L2 cache.
Main storage has up to 10 TB addressable memory per CPC drawer, which uses up to 48
DDR4 DIMMs. A system with four CPC drawers can have up to 40 TB of main storage.
Considerations
Cache sizes are limited by ever-diminishing cycle times because they must respond quickly
without creating bottlenecks. Access to large caches costs more cycles. Instruction and data
cache (L1) sizes must be limited because larger distances must be traveled to reach long
cache lines. This L1 access time generally occurs in one cycle, which prevents increased
latency.
Also, the distance to remote caches as seen from the microprocessor becomes a significant
factor. For example, on an IBM z15 server, access to L4 physical cache (on the SC chip and
which might not even be in the same CPC drawer) requires several cycles to travel the
distance to the cache. On an IBM z16, having an L4 virtual, physically allocated on the shared
L2 requires fewer processor cycles in many instances.
Although large caches mean increased access latency, the new technology 7 nm EUV chip
lithography and the lower cycle time allows IBM z16 servers to increase the size of L2 cache
level within the PU chip.
To overcome the inherent delays of the SMP CPC drawer design and save cycles to access
the remote virtual L4 content, the system keeps instructions and data as close to the
processors as possible. This configuration can be managed by directing as much work of a
specific LPAR workload to the processors in the same CPC drawer as the L4 virtual cache.
Figure 3-5 IBM z16 and IBM z15 cache level comparison
Compared to IBM z15, the IBM z16 cache design has larger L2 cache size. Cache L3 and L4
are now virtual caches in the IBM z16. More affinity exists between the memory of a partition,
the L4 virtual cache in a drawer, and the cores in the PU chips. As in IBM z15, the IBM z16
cache level structure is focused on keeping more data closer to the PU. This design can
improve system performance on many production workloads.
HiperDispatch
To help avoid latency in a high-frequency processor design, PR/SM and the dispatcher must
be prevented from scheduling and dispatching a workload on any processor available, which
keeps the workload in as small a portion of the system as possible. The cooperation between
z/OS and PR/SM is bundled in a function that is called HiperDispatch. HiperDispatch uses
the IBM z16 cache topology, which features reduced cross-cluster “help” and better locality for
multi-task address spaces.
PR/SM can use dynamic PU reassignment to move processors (CPs, ZIIPs, IFLs, ICFs, and
spares) to a different chip and drawer to improve the reuse of shared caches by processors of
the same partition. It can use dynamic memory relocation (DMR) to move a running partition’s
memory to different physical memory to improve the affinity and reduce the distance between
the memory of a partition and the processors of the partition.
For more information about HiperDispatch, see 3.7, “Logical partitioning” on page 117.
The IBM z16 A01 inter-CPC drawer communication structure is shown in Figure 3-6.
Inter-CPC drawer communication occurs at the Level 4 virtual cache level, which is
implemented on the semi-private part of one of the Level 2 caches in a chip module. The
Level 4 cache function regulates coherent drawer-to-drawer traffic.
IBM z16 A01 core frequency is 5.2 GHz (same as IBM z15 T01), but with increased number
of processors that share larger caches to have shorter access times and improved capacity
and performance.
Through innovative processor design (significant architecture changes, new cache structure,
new Core-Nest interface, new branch prediction design that uses dense SRAM, and new
on-chip AI accelerator for inference), the IBM Z processor performance continues to evolve.
Enhancements were made on the processor unit design, including the following examples:
Cache structure
Branch prediction mechanism
Floating point unit
Divide engine scheduler
Load/Store Unit and Operand Store Compare (OSC)
Simultaneous multi-threading
Relative nest intensity (RNI) redesigns
For more information about RNI, see 12.4, “Relative Nest Intensity” on page 479.
The processing performance was enhanced through the following changes to the IBM z16
processor design:
Core optimization to enable performance and capacity growth.
New cache structure design, including a larger cache Level 2 (SRAM) and virtual Level 3
and Level 4 cache to reduce latency.
On-chip IBM Integrated Accelerator for zEnterprise Data Compression (Nest compression
accelerator, or NXU. For more information, see Figure 2-14 on page 31.)
Enhancement of nest-core staging.
On-chip IBM Integrated Accelerator for AI. For more information, see Chapter 2, “Central
processor complex hardware components” on page 15, and Appendix B, “IBM Z
Integrated Accelerator for AI” on page 503.
Because of these enhancements, the IBM z16 processor full speed z/OS single-thread
performance is on average 1.11 times faster than the IBM z15 at equal N-way. For more
information about performance, see Chapter 12, “Performance and capacity planning” on
page 473.
z13 servers introduced architectural extensions with instructions that reduce processor
quiesce effects, cache misses, and pipeline disruption, and increase parallelism with
instructions that process several operands in a single instruction (SIMD). The processor
architecture was further developed for IBM z14 and IBM z15.
the IBM z16 enhanced Instruction Set Architecture (ISA) includes a set of instructions that are
added to improve compiled code efficiency. These instructions optimize PUs to meet the
demands of various business and analytics workload types without compromising the
performance characteristics of traditional workloads.
An operating system with SMT support can be configured to dispatch work to a thread on a
zIIP (for eligible workloads in z/OS) or an IFL (for z/VM and Linux on IBM Z) core in single
thread or SMT mode so that HiperDispatch cache optimization can be considered. For more
information about operating system support, see Chapter 7, “Operating system support” on
page 247.
SMT technology allows instructions from more than one thread to run in any pipeline stage at
a time. SMT can handle up to four pending translations.
Each thread has its own unique state information, such as Program Status Word (PSW) and
registers. The simultaneous threads cannot necessarily run instructions instantly and must at
times compete to use certain core resources that are shared between the threads. In some
cases, threads can use shared resources that are not experiencing competition.
Figure 3-8 Two threads running simultaneously on the same processor core
The use of SMT provides more efficient use of the processors’ resources and helps address
memory latency, which results in overall throughput gains. The active thread shares core
resources in space, such as data and instruction caches, TLBs, branch history tables, and, in
time, pipeline slots, execution units, and address translators.
Although SMT increases the processing capacity, the performance in some cases might be
superior if a single thread is used. Enhanced hardware monitoring supports measurement
through CPUMF for thread usage and capacity.
For workloads that need maximum thread speed, the partition’s SMT mode can be turned off.
For workloads that need more throughput to decrease the dispatch queue size, the partition’s
SMT mode can be turned on.
SMT use is functionally transparent to middleware and applications, and no changes are
required to run them in an SMT-enabled partition.
The 32 vector registers feature 128 bits. The instructions include string operations, vector
integer, and vector floating point operations. Each register contains multiple data elements of
a fixed size. The following instructions code specifies which data format to use and the size of
the elements:
Byte (16 8-bit operands)
Halfword (eight 16-bit operands)
Word (four 32-bit operands)
Doubleword (two 64-bit operands)
Quadword (one 128-bit operand)
The collection of elements in a register is called a vector. A single instruction operates on all
of the elements in the register. Instructions include a nondestructive operand encoding that
allows the addition of the register vector A and register vector B and stores the result in the
register vector A (A = A + B).
A schematic representation of a SIMD instruction with 16-byte size elements in each vector
operand is shown in Figure 3-9.
For most operations, the condition code is not set. A summary condition code is used only for
a few instructions.
The IBM z16 processor includes pipeline enhancements that benefit Out-of-Order execution.
The processor design features advanced micro-architectural innovations that provide the
following benefits:
Maximized instruction-level parallelism (ILP) for a better cycles per instruction (CPI)
design.
Maximized performance per watt.
Enhanced instruction dispatch and grouping efficiency.
Increased Out-of-Order) resources, such as Global Completion Table entries, physical
GPR entries, and physical FPR entries.
Improved completion rate.
Reduced cache/TLB miss penalty.
Improved execution of D-Cache store and reload and new Fixed-point divide.
New Operand Store Compare (OSC) (load-hit-store conflict) avoidance scheme.
Enhanced branch prediction structure and sequential instruction fetching.
Program results
The Out-of-Order execution does not change any program results. Execution can occur out of
(program) order, but all program dependencies are accepted, and the same results are seen
as in-order (program) execution. The design was optimized by increasing the Global
Completion Table (GCT) from 48x3 to 60x3, which increased the issue queue size from 2x30
to 2x36 and designed a new Mapper.
This implementation requires special circuitry to make execution and memory accesses
display in order to the software. The logical diagram of an IBM z16 core is shown in
Figure 3-11 on page 83.
Memory address generation and memory accesses can occur out of (program) order. This
capability can provide a greater use of the IBM z16 superscalar core, and improve system
performance.
The IBM z16 A01 processor unit core is a superscalar, out-of-order, SMT processor with eight
execution units. Up to six instructions can be decoded per cycle, and up to 12 instructions or
operations can be started to run per clock cycle (0.192 ns). The execution of the instructions
can occur out of program order and memory address generation and memory accesses can
also occur out of program order. Each core includes special circuitry to display execution and
memory accesses in order to the software.
The IBM z16 superscalar PU core can have up to 10 instructions or operations that are
running per cycle. This technology results in shorter workload runtime.
Branch prediction is now implemented as a two level BTB, BTB1 (“small” and “fast”), and
BTB2 (large, dense-SRAM). Now, BTB1 and BTB2 feature dynamic (variable) capacity:
BTB1: First Level Branch Target Buffer, smaller than IBM z15, dynamic director, variable
capacity:
– Minimum total branches in all parents (all large branches) = 8 K
– Maximum total branches in all parents (all medium branches) = 12 K
BTB2: Second Level Branch Target Buffer, also variable capacity (variable directory), up to
260 k branches
IBM z16 is a superscalar processor. Each processor unit, or core, is a superscalar and
out-of-order processor that supports 10 concurrent issues to execution units in a single CPU
cycle:
Fixed-point unit (FXU): The FXU handles fixed-point arithmetic.
Load-store unit (LSU): The LSU contains the data cache. It is responsible for handling all
types of operand accesses of all lengths, modes, and formats as defined in the
z/Architecture.
COBOL enhancements
IBM z16 core implements new instructions for the compiler to accelerate numeric formatting,
and hardware support for new numeric conversion instructions (exponents and arithmetic
common in financial applications).
One Nest Accelerator Unit (NXU) is used per processor chip, which is shared by all cores on
the chip and features the following benefits:
Brand new concept of sharing and operating an accelerator function in the nest
Supports DEFLATE compliant compression/decompression and GZIP CRC/ZLIB Adler
Low latency
High bandwidth
Problem state execution
Hardware/Firmware interlocks to ensure system responsiveness
Designed instruction
Run in millicode
Moving the compression function from the I/O drawer to the processor chip means that
compression can operate directly on L2 cache and data does not need to be passed by using
I/O.
Data compression is running in one of the two execution modes available: Synchronous mode
or Asynchronous mode:
Synchronous execution occurs in problem states where the user application starts the
instruction in its virtual address space.
Asynchronous execution is optimized for Large Operations under z/OS for authorized
applications (for example, BSAM/QSAM) and issues I/O by using EADMF for
asynchronous execution.
Figure 3-12 shows the nest compression accelerator (NXU) for On-Chip Compression
acceleration.
Figure 3-12 Integrated Accelerator for zEDC (NXU) on the IBM z16 PU chip
For more information about sizing, migration considerations, and software support, see
Appendix C, “IBM Integrated Accelerator for zEnterprise Data Compression” on page 507.
The cryptography coprocessor is used for CPACF, which offers a set of symmetric
cryptographic functions for encrypting and decrypting of clear key operations.
For more information about these instructions, see z/Architecture Principles of Operation,
SA22-7832.
CPACF accelerator that is built into every core supports Pervasive Encryption by providing
fast synchronous cryptographic services:
Encryption (DES, TDES, and AES)
Hashing (SHA-1, SHA-2, SHA-3, and SHAKE)
Random Number Generation (PRNG, DRNG, and TRNG)
Elliptic Curve supported operations:
– ECDH[E]:
• P256, P384, and P521
• X25519 and X448
– ECDSA:
• Keygen, sign, and verify
• P256, P384, and P521
For more information about cryptographic functions on IBM z16 servers, see Chapter 6,
“Cryptographic features” on page 209.
Introduced on IBM z15 was the sort accelerator that is known as the IBM Integrated
Accelerator for Z Sort (see Figure 3-13 on page 87). The SORTL hardware instruction that is
implemented on each core is used by DFSORT and the Db2 utilities for z/OS Suite to allow
the use of a hardware-accelerated approach to sorting.
The IBM Integrated Accelerator for Z Sort feature termed as “ZSORT” helps to reduce the
CPU costs and improve the elapsed time for eligible workloads. One of the primary
requirements for ZSORT is providing enough virtual, real, and auxiliary storage.
Sort jobs that run in memory-constrained environments in which the amount of memory that
is available to be used by DFSORT jobs is restricted might not achieve optimal performance
results or might not be able to use ZSORT.
The 64-bit memory objects (above-the-bar-storage) can use the ZSORT accelerator for sort
workloads for optimal results. Because ZSORT is part of the CPU and memory latency is
much less than disk latency, sorting in memory is more efficient than sorting with memory and
disk workspace. By allowing ZSORT to process the input completely in memory, it can
achieve the best results in elapsed time and CPU time.
Because the goal of ZSORT is to reduce CPU time and elapsed time, it can require more
storage than a DFSORT application that does not use ZSORT.
Note: Not all sorts are eligible to use ZSORT. IBM’s zBNA tool provides modeling support
for identifying potential ZSORT-eligible candidate jobs and estimates the benefits of
ZSORT. The tool uses information in the SMF type 16 records.
The following restrictions disable ZSORT and revert to the use of traditional sorting technique:
SORTL facility is not enabled/unavailable on the processor
ZSORT is not enabled
OPTION COPY or SORT FIELDS=COPY is specified
Use of:
– INREC
– JOINKEYS
– MERGE FIELDS
– MODS(EXITS) statements
– OUTREC
– OUTFIL
– SUM FIELDS
Program started sorts
Memory objects cannot be created
The new IBM z16 microprocessor chip, also called the IBM Telum processor, integrates a
new AI accelerator. This innovation brings incredible value to applications and workloads that
are running on IBM Z platform.
Customers can benefit from the integrated AI accelerator by adding AI operations that are
used to perform fraud prevention and fraud detection, customer behavior predictions, and
supply chain operations. All of these operations are done in real time and fully integrated in
transactional workloads. As a result, valuable insights are gained from their data instantly.
The integrated accelerator for AI delivers AI inference in real time, at large scale, and high
throughput rate, with no transaction left behind. The AI capability applies directly to the
running transaction. It shifts the traditional paradigm of applying AI to the transactions that
were completed. This innovative technology also can be used for intelligent IT workloads
placement algorithms, which contributes to the better overall system performance.
The Telum processor also integrates powerful mechanisms of data prefetch, fast and high
capacity level 1 (L1) and level 2 (L2) caches, enhanced branch prediction, and other
improvements and innovations that streamlines the data processing by the AI accelerator.
The hardware, firmware, and software are vertically integrated to deliver the new AI for
inference functions seamless to the applications.
The AI accelerator is driven by the new Neural Networks Processing Assist (NNPA)
instruction.
Intelligent data movers and prefetchers are connected to the chip by way of ring interface for
high-speed, low-latency, read/write cache operations at 200+ GBps read/store bandwidth,
and 600+ GBps bandwidth between engines.
Compute Arrays consist of 128 processor tiles with 8-way FP-16 FMA SIMD, which are
optimized for matrix multiplication and convolution, and 32 processor tiles with 8-way
FP-16/FP-32 SIMD, which are optimized for activation functions and complex functions.
The integrated AI accelerator delivers more than 6 TFLOPs per chip and over 200 TFLOPs in
fully configured IBM z16 system with the 32 chips. The AI accelerator is shared by all cores
on the chip. The firmware, running on the cores and accelerator, orchestrates and
synchronizes the execution on the accelerator.
Acknowledging the diverse AI training frameworks, customers can train their models on
platforms of their choice, including IBM Z (on-premises and in hybrid cloud) and then, deploy
it efficiently on IBM Z in colocation with the transactional workloads. No other development
effort is needed to enable this strategy.
IBM has invested into Open Neural Network Exchange (ONNX), which is a standard format
for representing AI models that allows a data scientist to build and train a model in the
framework of choice without worrying about the downstream inference implications.
To enable deployment of ONNX models, IBM provides an ONNX model compiler that is
optimized for IBM Z. IBM also optimized key Open Source frameworks, such as TensorFlow
and TensorFlow Serving, for use on IBM Z platform.
IBM open-sourced zDNN library provides common APIs for the functions that allow to convert
tensor format to the accelerator required one. Customers can run zDNN under z/OS (in zCX)
and Linux on IBM Z.
A Deep Learning Compiler (DLC) for z/OS and for Linux on IBM Z provides the AI functions to
the applications.
Base 10 arithmetic is used for most business and financial computation. Floating point
computation that is used for work that is typically done in decimal arithmetic involves frequent
data conversions and approximation to represent decimal numbers. This process makes
floating point arithmetic complex and error-prone for programmers who use it for applications
in which the data is typically decimal.
IBM z16 servers have two DFP accelerator units per core, which improve the decimal floating
point execution bandwidth. The floating point instructions operate on newly designed vector
registers (32 new 128-bit registers).
IBM z16 servers include decimal floating point packed conversion facility support with the
following benefits:
Reduces code path length because extra instructions to format conversion are no longer
needed.
Packed data is operated in memory by all decimal instructions without general-purpose
registers, which were required only to prepare for decimal floating point packed conversion
instruction.
Converting from packed can now force the input packed value to positive instead of
requiring a separate OI, OILL, or load positive instruction.
Converting to packed can now force a positive zero result instead of requiring ZAP
instruction.
Software support
DFP is supported in the following programming languages and products:
Release 4 and later of the High Level Assembler
C/C++, which requires supported z/OS version
Enterprise PL/I Release 3.7 and Debug Tool Release 8.1 or later
Java Applications that use the BigDecimal Class Library
SQL support as of Db2 Version 9 and later
The IBM z16 core implements two other execution subunits for 2x throughput on BFP
(single/double precision) operations (see Figure 3-11 on page 83).
The key point is that Java and C/C++ applications tend to use IEEE BFP operations more
frequently than earlier applications. Therefore, the better the hardware implementation of this
set of instructions, the better the performance of applications.
Relocation under hardware control is possible because the R-unit has the full designed state
in its buffer. PU error detection and recovery are shown in Figure 3-17.
The BHT (Branch History Table) implementation on processors provides a large performance
improvement. Originally introduced on the IBM ES/9000 9021 in 1990, the BHT is
continuously improved.
It offers significant branch performance benefits. The BHT allows each PU to take instruction
branches that are based on a stored BHT, which improves processing times for calculation
routines.
In addition to the BHT, IBM z16 servers use the following techniques to improve the prediction
of the correct branch to be run:
BTB
PHT
CTB
The success rate of branch prediction contributes significantly to the superscalar aspects of
IBM z16 processor. This success is because the architecture rules prescribe that the correctly
predicted result of the branch is essential for successful parallel execution of an instruction
stream.
IBM z16 integrates a new branch prediction design that uses SRAM and supports the
following enhancements over IBM z15:
BTB1: 8 K - 12 K
BTB2: up to 260 K
TAGE PHT: 4 k x 2
TAGE CTB: 1 k x 2
With the wild branch hardware facility, the last address from which a successful branch
instruction was run is kept. z/OS uses this information with debugging aids, such as the SLIP
command, to determine from where a wild branch came.
It also can collect data from that storage location. This approach decreases the number of
debugging steps that are necessary when you want to know from where the branch came.
The size of the TLB is kept as small as possible because of its short access time
requirements and hardware space limitations. Because memory sizes recently increased
significantly as a result of the introduction of 64-bit addressing, a smaller working set is
represented by the TLB.
To increase the working set representation in the TLB without enlarging the TLB, large (1 MB)
page and giant page (2 GB) support is available and can be used when suitable. For more
information, see “Large page support” on page 114.
With the enhanced DAT-2 (EDAT-2) improvements, the IBM Z servers support 2 GB page
frames.
The new translation engine allows up to four translations pending concurrently. Each
translation step is ~2x faster, which helps second level guests.
Instruction fetching
Instruction fetching normally tries to get as far ahead of instruction decoding and execution as
possible because of the relatively large instruction buffers that are available. In the
microprocessor, smaller instruction buffers are used. The operation code is fetched from the
I-cache and put in instruction buffers that hold prefetched data that is awaiting decoding.
Instruction decoding
The processor can decode up to six instructions per cycle. The result of the decoding process
is queued and later used to form a group.
Instruction grouping
From the instruction queue, up to 12 instructions can be completed on every cycle. A
complete description of the rules is beyond the scope of this publication.
The compilers and JVMs are responsible for selecting instructions that best fit with the
superscalar microprocessor. They abide by the rules to create code that best uses the
superscalar implementation. All IBM Z compilers and JVMs are constantly updated to benefit
from new instructions and advances in microprocessor designs.
A new procedure is available to run against customer macros in assembler libraries to verify
that no conflicts exist between some older IBM Macro names and new IBM z16 (M/T 3931)
hardware instruction mnemonics.
The Transaction Execution Facility provides instructions, including declaring the beginning
and end of a transaction, and canceling the transaction. TX is expected to provide significant
performance benefits and scalability by avoiding most locks. This benefit is especially
important for heavily threaded applications, such as Java.
3.5.1 Overview
All PUs on an IBM z16 are physically identical. When the system is initialized, two integrated
firmware processors (IFP) are allocated from the pool of PUs that is available for the entire
system. The other PUs can be characterized to specific functions (CP, IFL, ICF, zIIP, or SAP).
The function that is assigned to a PU is set by the Licensed Internal Code (LIC). The LIC is
loaded when the system is initialized at power-on reset (POR) and the PUs are characterized.
This design brings outstanding flexibility to IBM z16 servers because any PU can assume any
available characterization. The design also plays an essential role in system availability
because PU characterization can be done dynamically, with no system outage.
For more information about software level support of functions and features, see Chapter 7,
“Operating system support” on page 247.
Concurrent upgrades
For all IBM z16 A01 features that have more processor units (PUs) installed
(non-characterized) than activated, concurrent upgrades can be done by using LIC activation.
This activation assigns a PU function to a previously non-characterized PU. No hardware
changes are required.
For more information about Capacity on Demand, see Chapter 8, “System upgrades” on
page 327.
PU sparing
If a PU failure occurs, the failed PU’s characterization is dynamically and transparently
reassigned to a spare PU. IBM z16 servers have two spare PUs. PUs that are not
characterized on a CPC configuration also can be used as extra spare PUs.
For more information about PU sparing, see 3.5.10, “Sparing rules” on page 111.
PU pools
PUs that are defined as CPs, IFLs, ICFs, and zIIPs are grouped in their own pools from where
they can be managed separately. This configuration significantly simplifies capacity planning
and management for LPARs. The separation also affects weight management because CP
and zIIP weights can be managed separately.
All assigned PUs are grouped in the PU pool. These PUs are dispatched to online logical
PUs. For example, consider an IBM z16 with 10 CPs, 2 IFLs, 5 zIIPs, and 1 ICF. This system
has a PU pool of 18 PUs, called the pool width. Subdivision defines the following pools:
A CP pool of 10 CPs
An ICF pool of one ICF
An IFL pool of two IFLs
A zIIP pool of five zIIPs
PUs are removed from their pools when a concurrent downgrade occurs as the result of the
removal of a CBU. They also are removed through the On/Off CoD process and the
conversion of a PU. When a dedicated LPAR is activated, its PUs are taken from the correct
pools. This process also is the case when an LPAR logically configures a PU as on, if the
width of the pool allows for it.
For an LPAR, logical PUs are dispatched from the supporting pool only. The logical CPs are
dispatched from the CP pool, logical zIIPs from the zIIP pool, logical IFLs from the IFL pool,
and the logical ICFs from the ICF pool.
PU weighting
Because CPs, zIIPs, IFLs, and ICFs have their own pools from where they are dispatched,
they can be given their own weights. For more information about PU pools and processing
weights, see the Processor Resource/Systems Manager Planning Guide, SB10-7178.
The IBM z16 can be initialized in LPAR (PR/SM) mode or in Dynamic Partition Manger (DPM)
mode.
CPs are defined as dedicated or shared. Reserved CPs can be defined to an LPAR to allow
for nondisruptive image upgrades. If the operating system in the LPAR supports the logical
processor add function, reserved processors are no longer needed. Regardless of the
installed model, an LPAR can have up to 200 logical CPs that are defined (the sum of active
and reserved logical CPs). In practice, define no more CPs than the operating system
supports.
All PUs that are characterized as CPs within a configuration are grouped into the CP pool.
The CP pool can be seen on the Hardware Management Console (HMC) workplace. Any
z/Architecture operating systems and CFCCs can run on CPs that are assigned from the CP
pool.
The IBM z16 A01 server recognizes four distinct capacity settings for CPs. Full-capacity CPs
are identified as CP7. In addition to full-capacity CPs, three subcapacity settings (CP6, CP5,
and CP4), each for up to 39 PUs, are offered.
Granular capacity adds 117 subcapacity settings to the 200 capacity settings that are
available with full capacity CPs (CP7). Each of the 117 subcapacity settings applies to up to
39 CPs only, independent of the model installed.
Note: Information about CPs in the remainder of this chapter applies to all CP capacity
settings, unless indicated otherwise. For more information about granular capacity, see
2.3.3, “PU characterization” on page 33.
Note: IFLs can be dedicated to a Linux, a z/VM, or an SSC LPAR, or can be shared by
multiple Linux guests, z/VM LPARs, or SSC that are running on the same IBM z16 server.
Only z/VM, Linux on IBM Z operating systems, SSC, and designated software products
can run on IFLs. IFLs are orderable by using FC 1959.
IFLs do not change the model capacity identifier of the IBM z16. Software product license
charges that are based on the model capacity identifier are not affected by the addition of
IFLs.
Unassigned IFLs
An IFL that is purchased but not activated is registered as an unassigned IFL (FC 1962).
When the system is later upgraded with another IFL, the system recognizes that an IFL was
purchased and is present.
The allowable number of IFLs and Unassigned IFLs per feature is listed in Table 3-1.
Unassigned ICFs
New on IBM z16, an ICF that is purchased but not activated is registered as an unassigned
ICF (FC 1974). When the system is later upgraded with another ICF, the system recognizes
that an ICF was purchased and is present.
The allowable number of ICFs and Unassigned ICFs for each model is listed in Table 3-2.
ICFs exclusively run CFCC. ICFs do not change the model capacity identifier of the IBM z16
system. Software product license charges that are based on the model capacity identifier are
not affected by the addition of ICFs.
All ICFs within a configuration are grouped into the ICF pool. The ICF pool can be seen on the
HMC workplace.
After the image is dispatched, “poll for work” logic in CFCC and z/OS can be used largely
as-is to locate and process the work. The new interrupt expedites the redispatching of the
partition.
LPAR presents these Coupling Thin Interrupts to the guest partition, so CFCC and z/OS both
require interrupt handler support that can deal with them. CFCC also changes to relinquish
control of the processor when all available pending work is exhausted, or when the LPAR
undispatches it off the shared processor, whichever comes first.
CF processor combinations
A CF image can have one of the following combinations that are defined in the image profile:
Dedicated ICFs
Shared ICFs
Dedicated CPs
Shared CPs
Shared ICFs add flexibility. However, running only with shared coupling facility PUs (ICFs or
CPs) is not a preferable production configuration. It is preferable for a production CF to
operate by using dedicated ICFs.
5 It is the only option for shared processors in a CF image (whether they be ICFs or CPs) on IBM z16.
The LPAR processing weights are used to define how much processor capacity each CF
image can include. The capped option also can be set for a test CF image to protect the
production environment.
Connections between these z/OS and CF images can use internal coupling links to avoid the
use of real (external) coupling links, and get the best link bandwidth available.
Dynamic CF dispatching
The dynamic coupling facility dispatching (DYNDISP) function features a dispatching
algorithm that you can use to define a backup CF in an LPAR on the system. When this LPAR
is in backup mode, it uses few processor resources.
DYNDISP allows more environments with multiple CF images to coexist in a server, and to
share CF engines with reasonable performance. DYNDISP THIN is the only option for CF
images that use shared processors on IBM z16. For more information, see 3.9.3, “Dynamic
CF dispatching” on page 139.
To improve CF processor scaling for the customer’s CF images and to make effective use of
more processors as the sysplex workload increases, CF work management and dispatcher
provide the following improvements IBM z16:
IBM z16 provides improved CF processor scalability for CF images.
Increased number of CF tasks
ICA SR latency improvements that improve coupling efficiency in a parallel sysplex.
A zIIP enables eligible z/OS workloads to have a portion of them directed for execution to a
processor that is characterized as a zIIP. Because the zIIPs do not increase the MSU value of
the processor, they do not affect the IBM software license changes.
IBM z16 is the fourth generation of IBM Z processors to support SMT. IBM z16 servers
implement two threads per core on IFLs and zIIPs. SMT must be enabled at the LPAR level
and supported by the z/OS operating system. SMT was enhanced for IBM z16 and it is
enabled for SAPs by default (no customer intervention required).
Introduced in z/OS V2R4, the z/OS Container Extensions7 allows deployment of Linux on IBM
Z software components, such as Docker Containers in a z/OS system, in direct support of
z/OS workloads without requiring a separately provisioned Linux server. It also maintains
overall solution operational control within z/OS and with z/OS qualities of service. Workload
deployed in z/OS Container Extensions is zIIP eligible.
This process reduces the CP time that is needed to run Java WebSphere applications, which
frees that capacity for other workloads.
6 IBM z Systems Application Assist Processors (zAAPs) are not available since IBM z14 servers. A zAAP workload is
dispatched to available zIIPs (zAAP on zIIP capability).
7
z/OS Container Extensions that are running on IBM z16 require IBM Container Hosting Foundation for z/OS
software product (5655-HZ1).
A zIIP runs IBM authorized code only. This IBM authorized code includes the z/OS JVM in
association with parts of system code, such as the z/OS dispatcher and supervisor services.
A zIIP cannot process I/O or clock comparator interruptions. It also does not support operator
controls, such as IPL.
Java application code can run on a CP or a zIIP. The installation can manage the use of CPs
so that Java application code runs only on CPs or zIIPs, or on both.
If zIIPs are defined to the LPAR but are not online, the zIIP-eligible work units are processed
by CPs in order of priority. The system ignores the IIPHONORPRIORITY parameter in this
case and handles the work as though it had no eligibility to zIIPs.
The following Db2 UDB for z/OS V8 or later workloads can run in Service Request Block
(SRB) mode:
Query processing of network-connected applications that access the Db2 database over a
TCP/IP connection by using IBM Distributed Relational Database Architecture (DRDA).
DRDA enables relational data to be distributed among multiple systems. It is native to Db2
for z/OS, which reduces the need for more gateway products that can affect performance
and availability. The application uses the DRDA requester or server to access a remote
database. IBM Db2 Connect is an example of a DRDA application requester.
Star schema query processing, which is mostly used in business intelligence work.
A star schema is a relational database schema for representing multidimensional data. It
stores data in a central fact table and is surrounded by more dimension tables that hold
information about each perspective of the data. For example, a star schema query joins
various dimensions of a star schema data set.
8 z/OS V2R2 and later (older z/OS versions are out of support)
The zIIP runs portions of eligible database workloads, which helps to free computer capacity
and lower software costs. Not all Db2 workloads are eligible for zIIP processing. Db2 UDB for
z/OS V8 and later gives z/OS the information to direct portions of the work to the zIIP. The
result is that in every user situation, different variables determine how much work is redirected
to the zIIP.
On an IBM z16, the following workloads also can benefit from zIIPs:
z/OS Communications Server uses the zIIP for eligible Internet Protocol Security (IPSec)
network encryption workloads. Portions of IPSec processing take advantage of the zIIPs,
specifically end-to-end encryption with IPSec. The IPSec function moves a portion of the
processing from the general-purpose processors to the zIIPs. In addition, to run the
encryption processing, the zIIP also handles the cryptographic validation of message
integrity and IPSec header processing.
z/OS Global Mirror, formerly known as Extended Remote Copy (XRC), also uses the zIIP.
Most z/OS Data Facility Storage Management Subsystem (DFSMS) system data mover
(SDM) processing that is associated with z/OS Global Mirror can run on the zIIP.
The first IBM user of z/OS XML system services is Db2 V9. For Db2 V9 before the z/OS
XML System Services enhancement, z/OS XML System Services non-validating parsing
was partially directed to zIIPs when used as part of a distributed Db2 request through
DRDA. This enhancement benefits Db2 by making all z/OS XML System Services
non-validating parsing eligible to zIIPs. This configuration is possible when processing is
used as part of any workload that is running in enclave SRB mode.
z/OS Communications Server also allows the HiperSockets Multiple Write operation for
outbound large messages (originating from z/OS) to be run by a zIIP. Application
workloads that are based on XML, HTTP, SOAP, and Java, and traditional file transfer can
benefit.
During the SRB boost period, ANY work in a boosting image is eligible to run on a zIIP
processor associated with the image (LPAR).
Many more workloads and software can use zIIP processors, such as the following examples:
IBM z/OS Container Extensions (zCX)
IBM z/OS CIM monitoring
IBM z/OS Management Facility (z/OSMF)
System Display and Search Facility (SDSF)
IBM z/OS Connect EE components
IBM Sterling™ Connect:Direct®
IBM Z System Automation:
Java components of IBM Z SMS and SAS
IBM Z NetView RESTful API server
IBM Z Workload Scheduler & Dynamic Workload Console (under WebSphere Liberty)
IMS workloads (DRDA, SOAP, MSC, ISC)
Db2 for z/OS Data Gate
Db2 Sort for z/OS
Db2 Analytics Accelerator Loader for z/OS
Db2 Utilities Suite for z/OS
Db2 Log Analysis Tool for z/OS
Data Virtualization Manager for z/OS (DVM)
IzODA (Apache Spark workloads)
Watson Machine Learning for z/OS (WMLz) for Mleap and Spark workloads
For more information about zIIP and eligible workloads, see the IBM zIIP web page.
zIIP installation
One CP must be installed with or before any zIIP is installed. In IBM z16 A01, the zIIP-to-CP
ratio is 2:19, which means that up to 132 zIIPs on feature Max200 can be characterized.
Unassigned zIIPs
New on IBM z16, an ICF that is purchased but not activated is registered as an unassigned
ICF (FC 1974). When the system is later upgraded with another ICF, the system recognizes
that an ICF was purchased and is present.
The allowable number of zIIPs for each model is listed in Table 3-3.
zIIPs are orderable by using FC 1961. Up to two zIIPs can be ordered for each CP or marked
CP configured in the system. If the installed CPC drawer has no remaining unassigned PUs,
the assignment of the next zIIP might require the installation of another CPC drawer.
PUs that are characterized as zIIPs within a configuration are grouped into the zIIP pool. This
configuration allows zIIPs to have their own processing weights, independent of the weight of
parent CPs. The zIIP pool can be seen on the hardware console.
The number of permanent zIIPs plus temporary zIIPs cannot exceed twice the number of
purchased CPs plus temporary CPs. Also, the number of temporary zIIPs cannot exceed the
number of permanent zIIPs.
LPAR: In an LPAR, as many zIIPs as are available can be defined together with at least
one CP.
9
The 2:1 ratio can be exceeded (during boost periods) if System Recovery Boost Upgrade (FC 9930 and FC 6802)
is used for activating temporary zIIP capacity.
Standard SAPs 5 10 15 20 24
SAP configuration
A standard SAP configuration provides a well-balanced system for most environments.
However, some application environments feature high I/O rates, typically Transaction
Processing Facility (TPF) environments. In this case, more SAPs can be ordered.
Assigning more SAPs can increase the capability of the channel subsystem to run I/O
operations. In IBM z16 systems, the number of SAPs plus the number of optional SAPs
cannot exceed firmware limit of 128 threads.
By using reserved processors, you can define more logical processors than the number of
available CPs, IFLs, ICFs, and zIIPs in the configuration to an LPAR. This process makes it
possible to nondisruptively configure online more logical processors after more CPs, IFLs,
ICFs, and zIIPs are made available concurrently. They can be made available with one of the
capacity on-demand options.
Do not define more active and reserved processors than the operating system for the LPAR
can support. For more information about logical processors and reserved processors and
their definitions, see 3.7, “Logical partitioning” on page 117.
The two PUs that are characterized as IFP are dedicated to supporting firmware functions
that are implemented in Licensed Internal Code (LIC); for example, the resource groups
(RGs) that are used for managing the following native Peripheral Component Interconnect
Express (PCIe) features:
10GbE and 25GbE RoCE Express3 Short Reach (SR) and Long Reach (LR)
10GbE and 25GbE RoCE Express2
10GbE and 25GbE RoCE Express2.1
Coupling Express2 Long Reach
IFPs also are initialized at POR. They support Resource Group (RG) LIC11 to provide native
PCIe I/O feature management and virtualization functions.
The IBM z16 A01 PU assignment is based on CPC drawer plug order (not “ordering”).
Feature upgrade provides more processor (CPC) drawers. Max125 cannot be upgraded
because the supposed targeted features (Max168 and Max200) are factory built only.
The CPC drawers are populated from the bottom up. This process defines the low-order and
the high-order CPC drawers:
CPC drawer 1 (CPC 0 at position A10): Plug order 1 (low-order CPC drawer)
CPC drawer 2 (CPC 1 at position A15): Plug order 2
CPC drawer 3 (CPC 2 at position A20): Plug order 3 (high-order CPC drawer)
11 IBM zHyperLink Express1.1 and IBM zHyperLink Express are not managed by Resource Groups LIC.
These rules are intended to isolate processors that are used by different operating systems
as much as possible on different CPC drawers and even on different PU chips. This
configuration ensures that different operating systems do not use the same shared caches.
For example, CPs and zIIPs are all used by z/OS, and can benefit by using the same shared
caches. However, IFLs are used by z/VM and Linux, and ICFs are used by CFCC.
This initial PU assignment, which is done at POR, can be dynamically rearranged by an LPAR
by swapping an active core to a core in a different PU chip in a different CPC drawer to
improve system performance. For more information, see “LPAR dynamic PU reassignment”
on page 122.
When a CPC drawer is added concurrently after POR and new LPARs are activated, or
processor capacity for active partitions is dynamically expanded, the extra PU capacity can
be assigned from the new CPC drawer. The processor unit assignment rules consider the
newly installed CPC drawer dynamically.
Systems with a failed PU for which no spare is available call home for a replacement. A
system with a failed PU that is spared and requires a DCM to be replaced (referred to as a
pending repair) can still be upgraded when sufficient PUs are available.
With transparent sparing, the status of the application that was running on the failed
processor is preserved. The application continues processing on a newly assigned CP, IFL,
ICF, zIIP, SAP, or IFP (allocated to one of the spare PUs) without client intervention.
Application preservation
If no spare PU is available, application preservation (z/OS only) is started. The state of the
failing processor is passed to another active processor that is used by the operating system.
Through operating system recovery services, the task is resumed successfully (in most
cases, without client intervention).
12 For a layout of CPC drawers’ locations, refer to 3.5.11, “CPC drawer numbering” on page 112
3.6.1 Overview
the IBM z16 A01 memory design also provides flexibility, high availability, and the following
upgrades:
Concurrent memory upgrades if the physically installed capacity is not yet reached
IBM z16 servers can have more physically installed memory than the initial available
capacity. Memory upgrades within the physically installed capacity can be done
concurrently by LIC, and no hardware changes are required. However, memory upgrades
cannot be done through CBU or On/Off CoD.
Concurrent memory upgrades if the physically installed capacity is reached
Physical memory upgrades require a processor drawer to be removed and reinstalled after
replacing the memory cards in the processor drawer. Except for the feature Max39, the
combination of enhanced drawer availability and the flexible memory option allows you to
concurrently add memory to the system. For more information, see 2.5.5, “Drawer
replacement and memory” on page 45, and 2.5.7, “Flexible Memory Option” on page 46.
When the total capacity that is installed has more usable memory than required for a
configuration, the Licensed Internal Code Configuration Control (LICCC) determines how
much memory is used from each processor drawer. The sum of the LICCC provided memory
from each CPC drawer is the amount that is available for use in the system.
Memory allocation
When the system is activated by using a POR, PR/SM determines the total installed memory
and the customer enabled memory. Later in the process, during LPAR activation, PR/SM
assigns and allocates each partition memory according to their image profile.
PR/SM controls all physical memory, and can make physical memory available to the
configuration when a CPC drawer is added.
In older IBM Z processors, memory allocation was striped across the available CPC drawers
because relatively fast connectivity (that is, relatively fast to the processor clock frequency)
existed between the drawers. Splitting the work between all of the memory controllers allowed
a smooth performance variability.
The memory allocation algorithm changed starting with IBM z13®. For IBM z16, PR/SM tries
to allocate memory into a single CPC drawer. If memory does not fit into a single drawer,
PR/SM tries to allocate the memory into the CPC drawer with the most processor entitlement.
The PR/SM memory and logical processor resources allocation goal is to place all partition
resources on a single CPC drawer, if possible. The resources, such as memory and logical
processors, are assigned to the logical partitions at the time of their activation. Later on, when
all partitions are activated, PR/SM can move memory between CPC drawers to benefit the
performance of each LPAR, without operating system knowledge. This process was done on
the previous families of IBM Z servers only for PUs that use PR/SM dynamic PU reallocation.
PR/SM schedules a global reoptimization of the resources in use. It does so by reviewing all
the partitions that are active and prioritizing them based on their processing entitlement and
weights, which creates a high- and low-priority rank. Then, the resources, such as logical
processors and memory, can be moved from one CPC drawer to another to address the
priority ranks that were created.
When partitions are activated, PR/SM tries to find a home assignment CPC drawer, home
assignment node, and home assignment chip for the logical processors that are defined to
them. The PR/SM goal is to allocate all the partition logical processors and memory to a
single CPC drawer (the home drawer for that partition).
If all logical processors can be assigned to a home drawer and the partition-defined memory
is greater than what is available in that drawer, the exceeding memory amount is allocated on
another CPC drawer. If all the logical processors cannot fit in one CPC drawer, the remaining
logical processors spill to another CPC drawer. When that overlap occurs, PR/SM stripes the
memory (if possible) across the CPC drawers where the logical processors are assigned.
The process of reallocating memory is based on the memory copy/reassign function, which is
used to allow enhanced drawer availability (EDA) and concurrent drawer replacement
(CDR)13. This process was enhanced starting with z13 and IBM z13s® to provide more
efficiency and speed to the process without affecting system performance.
IBM z16 A01 implements a faster dynamic memory reallocation mechanism, which is
especially useful during service operations (EDA and CDR). PR/SM controls the
reassignment of the content of a specific physical memory array in one CPC drawer to a
physical memory array in another CPC drawer. To accomplish this task, PR/SM uses all the
available physical memory in the system. This memory includes the memory that is not in use
by the system that is available but not purchased by the client, and the planned memory
options, if installed.
Because of the memory allocation algorithm, systems that undergo many miscellaneous
equipment specification (MES) upgrades for memory can have different memory mixes and
quantities in all processor drawers of the system. If the memory fails, it is technically feasible
to run a POR of the system with the remaining working memory resources. After the POR
completes, the memory distribution across the processor drawers is different, as is the total
amount of available memory.
The TLB reduces the amount of time that is required to translate a virtual address to a real
address. This translation is done by dynamic address translation (DAT) when it must find the
correct page for the correct address space.
13
In previous IBM Z generations (before z13), these service operations were known as enhanced book availability
(EBA) and concurrent book repair (CBR).
The worst-case translation time occurs when a TLB miss occurs and the segment table
(which is needed to find the page table) and the page table (which is needed to find the entry
for the particular page in question) are not in cache. This case involves two complete real
memory access delays plus the address translation delay. The duration of a processor cycle
is much shorter than the duration of a memory cycle, so a TLB miss is relatively costly.
It is preferable to have addresses in the TLB. With 4 K pages, holding all of the addresses for
1 MB of storage takes 256 TLB lines. When 1 MB pages are used, it takes only one TLB line.
Therefore, large page size users have a much smaller TLB footprint.
Large pages allow the TLB to better represent a large working set and suffer fewer TLB
misses by allowing a single TLB entry to cover more address translations.
Users of large pages are better represented in the TLB and are expected to see performance
improvements in elapsed time and processor usage. These improvements are because DAT
and memory operations are part of processor busy time, even though the processor waits for
memory operations to complete without processing anything else in the meantime.
To overcome the processor usage that is associated with creating a 1 MB page, a process
must run for some time. It also must maintain frequent memory access to keep the pertinent
addresses in the TLB.
Short-running work does not overcome the processor usage. Short processes with small
working sets are expected to receive little or no improvement. Long-running work with high
memory-access frequency is the best candidate to benefit from large pages.
Long-running work with low memory-access frequency is less likely to maintain its entries in
the TLB. However, when it does run, few address translations are required to resolve all of the
memory it needs. Therefore, a long-running process can benefit even without frequent
memory access.
Weigh the benefits of whether something in this category must use large pages as a result of
the system-level costs of tying up real storage. A balance exists between the performance of
a process that uses large pages and the performance of the remaining work on the system.
On IBM z16 server, 1 MB large pages become pageable if Virtual Flash Memory14 is
available and enabled. They are available only for 64-bit virtual private storage, such as virtual
memory that is above 2 GB.
It is easy to assume that increasing the TLB size is a feasible option to deal with TLB-miss
situations. However, this process is not as simple as it seems. As the size of the TLB
increases, so does the processor usage that is involved in managing the TLB’s contents.
Correct sizing of the TLB is subject to complex statistical modeling to find the optimal tradeoff
between size and performance.
14 Virtual Flash Memory replaced IBM zFlash Express. No carry forward of zFlash Express exists.
Main storage can be accessed by all processors, but cannot be shared between LPARs. Any
system image (LPAR) must include a defined main storage size. This defined main storage is
allocated exclusively to the LPAR during partition activation.
The fixed size of the HSA eliminates planning for future expansion of the HSA because the
hardware configuration definition (HCD)/input/output configuration program (IOCP) always
reserves space for the following items:
Six channel subsystems (CSSs)
A total of 15 LPARs in CSSs 1 through 5, and 10 LPARs for the sixth CSS for a total of 85
LPARs
Subchannel set 0 with 63.75-K devices in each CSS
Subchannel set 1 with 64-K devices in each CSS
Subchannel set 2 with 64-K devices in each CSS
Subchannel set 3 with 64-K devices in each CSS
The HSA features sufficient reserved space to allow for dynamic I/O reconfiguration changes
to the maximum capability of the processor.
For IBM z16 A01, IBM VFM provides up to 6.0 TB of virtual flash memory in 512 GB
increments. The minimum is 0, while the maximum is 12 features. The number of VFM
features ordered reduces the maximum orderable memory for the IBM z16.
3.7.1 Overview
Logical partitioning is a function that is implemented by the PR/SM on IBM z16. IBM z16 can
run in LPAR mode, or in DPM mode. DPM provides a GUI for PR/SM to manage I/O
resources dynamically.
PR/SM is aware of the processor drawer structure on IBM z16 servers. However, LPARs do
not feature this awareness. LPARs feature resources that are allocated to them from various
physical resources. From a systems standpoint, LPARs have no control over these physical
resources, but the PR/SM functions do have this control.
PR/SM manages and optimizes allocation and the dispatching of work on the physical
topology. Most physical topology that was handled by the operating systems is the
responsibility of PR/SM.
As described in 3.5.9, “Processor unit assignment” on page 110, the initial PU assignment is
done during POR by using rules to optimize cache usage. This step is the “physical” step,
where CPs, zIIPs, IFLs, ICFs, and SAPs are allocated on the processor drawers.
When an LPAR is activated, PR/SM builds logical processors and allocates memory for the
LPAR.
PR/SM assigns all logical processors to one CPC drawer that are packed into chips of that
drawer and cooperates with operating system use of HiperDispatch.
All processor types of an IBM z16 can be dynamically reassigned, except IFPs.
Memory allocation changed from the previous IBM Z servers. Partition memory is now
allocated based on processor drawer affinity. For more information, see “Memory allocation”
on page 113.
Processor drawers assignment is more important because they optimize virtual L4 cache
usage. Therefore, logical processors from a specific LPAR are packed into a processor
drawer as much as possible.
PR/SM optimizes chip assignments within the assigned processor drawer (or drawers) to
maximize virtual L3 cache efficiency. Logical processors from an LPAR are dispatched on
physical processors on the same PU chip as much as possible.
PR/SM also tries to redispatch a logical processor on the same physical processor to
optimize private cache (L1 and L2) usage.
HiperDispatch
PR/SM and z/OS work in tandem to use processor resources more efficiently. HiperDispatch
is a function that combines the dispatcher actions and the knowledge that PR/SM has about
the topology of the system.
The nested topology is returned to z/OS by the Store System Information (STSI) instruction.
HiperDispatch uses the information to concentrate logical processors around shared caches
(virtual L3 and virtual L4 caches at drawer level), and dynamically optimizes the assignment
of logical processors and units of work.
z/OS dispatcher manages multiple queues, called affinity queues, with a target number of
eight processors per queue, which fits well onto a single PU chip. These queues are used to
assign work to as few logical processors as are needed for an LPAR workload. Therefore,
even if the LPAR is defined with many logical processors, HiperDispatch optimizes this
number of processors to be near the required capacity. The optimal number of processors to
be used is kept within a processor drawer boundary, when possible.
Logical partitions
PR/SM enables IBM z16 A01 systems to be initialized for a logically partitioned operation,
supporting up to 85 LPARs. Each LPAR can run its own operating system image in any image
mode, independently from the other LPARs.
An LPAR can be added, removed, activated, or deactivated at any time. Changing the number
of LPARs is not disruptive and does not require a POR. Certain facilities might not be
available to all operating systems because the facilities might have software corequisites.
Each LPAR has the following resources that are the same as a real CPC:
Processors
Called logical processors, they can be defined as CPs, IFLs, ICFs, or zIIPs. They can be
dedicated to an LPAR or shared among LPARs. When shared, a processor weight can be
defined to provide the required level of processor resources to an LPAR. Also, the capping
option can be turned on, which prevents an LPAR from acquiring more than its defined
weight and limits its processor consumption.
LPARs for z/OS can have CP and zIIP logical processors. The logical processor types can
be defined as all dedicated or all shared. The zIIP support is available in z/OS.
The weight and number of online logical processors of an LPAR can be dynamically
managed by the LPAR CPU Management function of the Intelligent Resource Director
(IRD). These functions can be used to achieve the defined goals of this specific partition
and of the overall system. The provisioning architecture of IBM z16 systems adds a
dimension to the dynamic management of LPARs, as described in Chapter 8, “System
upgrades” on page 327.
PR/SM supports an option to limit the amount of physical processor capacity that is used
by an individual LPAR when a PU is defined as a general-purpose processor (CP) or an
IFL that is shared across a set of LPARs.
This capability is designed to provide a physical capacity limit that is enforced as an
absolute (versus relative) limit. It is not affected by changes to the logical or physical
configuration of the system. This physical capacity limit can be specified in units of CPs or
IFLs. The Change LPAR Controls and Customize Activation Profiles tasks on the HMC
were enhanced to support this new function.
Memory
Memory (main storage) must be dedicated to an LPAR. The defined storage must be
available during the LPAR activation; otherwise, the LPAR activation fails.
Reserved storage can be defined to an LPAR, which enables nondisruptive memory
addition to and removal from an LPAR by using the LPAR dynamic storage reconfiguration
(z/OS and z/VM). For more information, see 3.7.5, “LPAR dynamic storage
reconfiguration” on page 127.
Channels
Channels can be shared between LPARs by including the partition name in the partition
list of a channel-path identifier (CHPID). I/O configurations are defined by the IOCP or the
HCD with the CHPID mapping tool (CMT). The CHPID Mapping Tool (CMT) is an optional
tool that is used to map CHPIDs onto physical channel IDs (PCHIDs). PCHIDs represent
the physical location of a port on a card in a PCIe I/O drawer.
IOCP is available on the z/OS, z/VM, and z/VSE operating systems, and as a stand-alone
program on the hardware console. For more information, see Input/Output Configuration
Program User’s Guide for ICP IOCP, SB10-7177. HCD is available on the z/OS and z/VM
operating systems. Review the suitable 3931DEVICE Preventive Service Planning (PSP)
buckets before implementation.
Modes of operation
The modes of operation are listed in Table 3-6. All available mode combinations, including
their operating modes and processor types, operating systems, and addressing modes, also
are listed. Only the currently supported versions of operating systems are considered.
CP z/VSE 64-bit
Linux on IBM Z
z/TPF
The 64-bit z/Architecture mode has no special operating mode because the architecture
mode is not an attribute of the definable images operating mode. The 64-bit operating
systems are in 31-bit mode at IPL and change to 64-bit mode during their initialization. The
operating system is responsible for taking advantage of the addressing capabilities that are
provided by the architectural mode.
For information about operating system support, see Chapter 7, “Operating system support”
on page 247.
General mode also is used to run the z/TPF operating system on dedicated or shared CPs
CF mode, by loading the CFCC code into the LPAR that is defined as one of the following
types:
– Dedicated or shared CPs
– Dedicated or shared ICFs
Linux only mode to run the following systems:
– A Linux on IBM Z operating system, on either of the following types:
• Dedicated or shared IFLs
• Dedicated or shared CPs
– A z/VM operating system, on either of the following types:
• Dedicated or shared IFLs
• Dedicated or shared CPs
z/VM mode to run z/VM on dedicated or shared CPs or IFLs, plus zIIPs and ICFs
IBM SSC mode LPAR can run on dedicated or shared:
– CPs
– IFLs
All LPAR modes, required characterized PUs, operating systems, and the PU
characterizations that can be configured to an LPAR image are listed in Table 3-7. The
available combinations of dedicated (DED) and shared (SHR) processors are also included.
For all combinations, an LPAR also can include reserved processors that are defined, which
allows for nondisruptive LPAR upgrades.
Linux only IFLs or CPs Linux on IBM Z IFLs DED or IFLs SHR
z/VM or
CPs DED or CPs SHR
z/VM CPs, IFLs, z/VM (V7R1 and later) All PUs must be SHR or DED
zIIPs, or
ICFs
The extra channel subsystem and multiple image facility (MIF) image ID pairs (CSSID/MIFID)
can be later assigned to an LPAR for use (or later removed). This process can be done
through dynamic I/O commands by using the HCD. At the same time, required channels must
be defined for the new LPAR.
Partition profile: Cryptographic coprocessors are not tied to partition numbers or MIF IDs.
They are set up with Adjunct Processor (AP) numbers and domain indexes. These
numbers are assigned to a partition profile of a specific name. The client assigns these AP
numbers and domains to the partitions and continues to have the responsibility to clear
them out when their profiles change.
Logical processors also can be concurrently added to a logical partition dynamically by using
the Support Element (SE) “Logical Processor Add” function under the CPC Operational
Customization task. This SE function allows the initial and reserved processor values to be
dynamically changed. The operating system must support the dynamic addition15 of these
resources.
For more information, see 3.5.9, “Processor unit assignment” on page 110.
15
In z/OS, this support is available since Version 1 Release 10 (z/OS V1R10), while z/VM supports this addition
since z/VM V5R4, and z/VSE since V4R3. However, IBM z16 supports z/OS V2R2 and later, z/VSE V6R2 and
z/VM V7R1 and later.
LPAR dynamic PU reassignment can swap client processors of different types between
processor drawers. For example, reassignment can swap an IFL on processor drawer 1 with a
CP on processor drawer 2. Swaps can also occur between PU chips within a processor
drawer or a node and can include spare PUs. The goals are to pack the LPAR on fewer
processor drawers and also on fewer PU chips, based on the IBM z16 processor drawers’
topology. The effect of this process is evident in dedicated and shared LPARs that use
HiperDispatch.
PR/SM and WLM work together to enforce the capacity that is defined for the group and the
capacity that is optionally defined for each individual LPAR.
Unlike traditional LPAR capping, absolute capping is designed to provide a physical capacity
limit that is enforced as an absolute (versus relative) value that is not affected by changes to
the virtual or physical configuration of the system.
Absolute capping provides an optional maximum capacity setting for logical partitions that is
specified in the absolute processors capacity (for example, 5.00 CPs or 2.75 IFLs). This
setting is specified independently by processor type (namely CPs, zIIPs, and IFLs) and
provides an enforceable upper limit on the amount of the specified processor type that can be
used in a partition.
Absolute capping is ideal for processor types and operating systems that the z/OS WLM
cannot control. Absolute capping is not intended as a replacement for defined capacity or
group capacity for z/OS, which are managed by WLM.
Absolute capping can be used with any z/OS, z/VM, or Linux on IBM Z LPAR (that is running
on an IBM Z server). If specified for a z/OS LPAR, absolute capping can be used concurrently
with defined capacity or group capacity management for z/OS. When used concurrently, the
absolute capacity limit becomes effective before other capping controls.
DPM provides facilities to define and run virtualized computing systems by using a
firmware-managed environment that coordinate the physical system resources that are
shared by the partitions. The partitions’ resources include processors, memory, network,
storage, crypto, and accelerators.
DPM provides a new mode of operation for IBM Z servers that provide the following services:
Facilitates defining, configuring, and operating PR/SM LPARs in a similar way to how
these tasks are performed on another platform.
Lays the foundation for a general IBM Z new user experience.
DPM is not another hypervisor for IBM Z servers. DPM uses the PR/SM hypervisor
infrastructure and provides an intelligent interface on top of it that allows customers to define,
use, and operate the platform virtualization without IBM Z experience or skills.
For more information about operating system main storage support, see the PR/SM Planning
Guide, SB10-7178.
For more information, see 3.7.5, “LPAR dynamic storage reconfiguration” on page 127.
Operating systems that run as guests of z/VM can use the z/VM capability of implementing
virtual memory to guest virtual machines. The z/VM dedicated real storage can be shared
between guest operating systems.
Important: The memory allocation and usage depends on the operating system
architecture and tested (documented for each operating system) limits.
While the maximum supported memory per LPAR for IBM z16 is 32 TB, each operating
system has its own support specifications.
For more information about general guidelines, see the PR/SM Planning Guide,
SB10-7178, which is available at the IBM Resource Link website (log in required).
Note: For information about the (kernel) supported amount of memory, check the Linux
Distribution specific documentation.
z/VM
In z/VM mode, specific types of processor units can be defined within one LPAR. This
feature increases flexibility and simplifies systems management by allowing z/VM to run
the following tasks in the same z/VM LPAR:
– Manage guests to operate Linux on IBM Z on IFLs
– Operate z/VSE and z/OS on CPs
– Offload z/OS system software processor usage, such as Db2 workloads on zIIPs
– Provide an economical Java execution environment under z/OS on zIIPs
IBM SSC
In IBM SSC mode, storage addressing is 64-bit for an embedded product. The amount of
usable main storage by the appliance code that is deployed in the SSC LPAR is
documented by the appliance code supplier.
Without the reserved storage definition, an LPAR storage upgrade is a disruptive process that
requires the following steps:
1. Partition deactivation.
2. An initial storage size definition change.
3. Partition activation.
The extra storage capacity for an LPAR upgrade can come from the following sources:
Any unused available storage
Another partition that features released storage
A memory upgrade
A concurrent LPAR storage upgrade uses DSR. z/OS uses the reconfigurable storage unit
(RSU) definition to add or remove storage units in a nondisruptive way.
z/VM V7R1 and later releases support the dynamic addition of memory to a running LPAR by
using reserved storage. It also virtualizes this support to its guests. In z/VM 7.1, removing
storage from the z/VM LPAR is disruptive. Removing memory from a z/VM guest is not
disruptive to the z/VM LPAR.
z/VM V7R2 and later also support Dynamic Memory Downgrade (DMD), which allows the
removal of up to 50% of the real storage from a running z/VM system.
LPAR storage granularity information is required for LPAR image setup and for z/OS RSU
definition. On IBM z16 A01, LPARs are limited to a maximum of 16 TB of main storage.
However, the maximum amount of memory that is supported by z/OS V2R2, V2R3, and V2R4
is 4 TB. z/OS V2R5 supports up to 16 TB. For z/VM V7R1 the limit is 2 TB, and z/VM V7R2
increases the limit to 4 TB.
With dynamic storage reconfiguration, the unused storage does not have to be continuous.
When an operating system that is running on an LPAR assigns a storage increment to its
configuration, PR/SM determines whether any free storage increments are available. PR/SM
then dynamically brings the storage online.
PR/SM dynamically takes offline a storage increment and makes it available to other
partitions when an operating system running on an LPAR releases a storage increment.
16 When defining an LPAR on the HMC, the 2 G boundary must be followed in PR/SM.
An LPAR cluster is shown in Figure 3-21. It contains three z/OS images and one Linux image
that is managed by the cluster. Included as part of the entire Parallel Sysplex is another z/OS
image and a CF image. In this example, the scope over which IRD has control is the defined
LPAR cluster.
For more information about implementing LPAR processor management under IRD, see z/OS
Intelligent Resource Director, SG24-5952.
Figure 3-22 shows an IBM z16 model A01system that contains multiple z/OS sysplex
partitions. It contains an internal CF, an IBM z15 model T02 system that contains a
stand-alone CF, and an IBM z14 model M03 that contains multiple z/OS sysplex partitions.
STP over coupling links provides time synchronization to all systems. Selecting the suitable
CF link technology Coupling Express2 Long Reach (CE2 LR) or Integrate Coupling Adapter
Short Reach (ICA SR and SR1.1) depends on the system configuration and how distant they
are physically.
For more information about link technologies, see “Coupling links” on page 187.
Parallel Sysplex is an enabling technology that allows highly reliable, redundant, and robust
IBM Z technology to achieve near-continuous availability. A Parallel Sysplex consists of one or
more (z/OS) operating system images that are coupled through one or more Coupling Facility
LPARs.
Note: Parallel sysplex coupling and timing links connectivity for IBM z16 (M/T 3931) is
supported to N-2 generation CPCs (IBM z16, IBM z15, and IBM z14).
Through state-of-the-art cluster technology, the power of multiple images can be harnessed
to work in concert on common workloads. The IBM Z Parallel Sysplex cluster takes the
commercial strengths of the platform to improved levels of system management, competitive
price performance, scalable growth, and continuous availability.
Consideration: IBM z16, IBM z15, and IBM z14 servers cannot coexist in the same
sysplex with z13/z13s or earlier generation systems.
Note: In IBM z16, DYNDISP=THIN option is the only available behavior for
shared-engine CF dispatching.
Specifying OFF or ON in CF commands and the CF configuration file are preserved for
compatibility; however, a warning message is issued to indicate that these options are no
longer supported and that DYNDISP=THIN behavior is to be used.
CFCC Level 24
CFCC level 24 is delivered on the IBM z15 (M/T 8561 and 8562) with driver level 41. CFCC
level 24 adds the following features:
CFCC Fair Latch Manager
This feature is an enhancement to the internals of the Coupling Facility (CFCC) dispatcher
to provide CF work management efficiency and processor scalability improvements, and
improve the “fairness” of arbitration for internal CF resource latches across tasks.
The tasks that are waiting for CF latches are not placed on the global suspend queue at
all; instead, they are placed on latch-specific waiter queues for the exact instance of the
latch they are requesting, and in the exact order in which they requested the latch. As a
result, the global suspend queue is much less heavily used, and thus is much less a
source of global contention or cache misses in the CF.
Also, when a latch is released, the specific latch’s latch waiter queue is used to transfer
ownership of the latch directly to the next request in line (or multiple requests, in the case
of a shared latch), and make that task (or tasks) ready to run, with the transferred latch
already held. No possibility exists of any unfairness or “cutters” in line between the time
that the latch is released versus when is obtained again.
For managing latches correctly for structures that are System-Managed (SM)
synchronous duplexing, it is now important for the CF to understand which of the duplexed
pair of requests operates as the “master” versus “slave” from a latching perspective, which
requires more SM duplexing setup information from z/OS.
z/OS XCF/XES toleration APAR support is required to provide this enhancement.
CFCC Level 23
CFCC level 23 is delivered on the z14 (M/T 3906 and 3907) with driver level 36. CFCC Level
23 introduces the following enhancements:
Asynchronous cross-invalidate (XI) of CF cache structures. Requires PTF support for
z/OS and explicit data manager support (Db2 V12 with PTFs). Consider the following
points:
– Instead of performing XI signals synchronously on every cache update request that
causes them, data managers can “opt in” for the CF to perform these XIs
asynchronously (and then sync them up with the CF at or before transaction
completion). Data integrity is maintained if all XI signals complete by the time
transaction locks are released.
– Results in faster completion of cache update CF requests, especially with cross-site
distance involved.
– Provides improved cache structure service times and coupling efficiency.
Coupling Facility hang detect enhancements provide a significant reduction in failure
scope and client disruption (CF-level to structure-level), with no loss of FFDC collection
capability:
– When a hang is detected, in most cases the CF confines the scope of the failure to
“structure damage” for the single CF structure the hung command was processing
against, capture diagnostics with a nondisruptive CF memory dump, and continue
operating without ending or restarting the CF image.
– Provides a significant reduction in failure scope and client disruption (CF-level to
structure-level), with no loss of FFDC collection capability.
Coupling Facility ECR granular latching:
– With this support, most CF list and lock structure ECR processing no longer use
structure-wide latching. It serializes its execution by using the normal structure object
latches that all mainline commands use.
– Eliminates the performance degradation that is caused by structure-wide latching.
– A few “edge conditions” in ECR processing still require structure-wide latching to be
used to serialize them.
z14 servers with CFCC Level 23 require z/OS V1R13 or later, and z/VM V6R4 or later for
virtual guest coupling.
CFCC Level 22
CFCC level 22 is delivered on the IBM z14 (M/T 3906) with driver level D32. CFCC Level 22
introduces the following enhancements:
CF Enhancements:
– CF structure encryption
CF Structure encryption is transparent to middleware and applications that use CFs,
while CF users are unaware of and not involved in the encryption. All data and adjunct
data that flows between z/OS and the CF is encrypted. The intent is to encrypt all data
that might be sensitive.
Internal control information and related request metadata is not encrypted, including
locks and lock structures.
z/OS generates the required structure-related encryption keys and does much of the
key management automatically by using CFRM that uses secure, protected keys
(never clear keys). Secure keys maintained in CFRM couple data set.
CF Asynchronous Duplexing for Lock Structures:
– New asynchronous duplexing protocol for lock structures:
• z/OS sends command to primary CF only
• Primary CF processes command and returns result
• Primary CF forwards description of required updates to secondary CF
• Secondary CF updates secondary structure instance asynchronously
– Provided for lock structures only:
• z/OS V2R2 SPE with PTFs for APAR OA47796
• Db2 V12 with PTFs
• Most performance-sensitive structures for duplexing
– Benefit and value:
• Db2 locking receives performance similar to simplex operations
• Reduces CPU and CF link overhead
• Avoids the overhead of synchronous protocol handshakes on every update
• Duplexing failover much faster than log-based recovery
– Targeted at multi-site clients who run split workloads at distance to make duplexing
lock structures at distance practical.
CF Processor Scalability:
– CF work management and dispatcher changes to allow improved efficiency as
processors are added to scale up the capacity of a CF image.
– Functionally specialized ICF processors that operate for CF images having more than
a threshold number of dedicated processors defined for them:
• One functionally specialized processor for inspecting suspended commands.
• One functionally specialized processor for pulling in new commands.
• The remaining processors are nonspecialized for general CF request processing.
– Avoids many inter-processor contentions that were associated with CF dispatching
To support an upgrade from one CFCC level to the next, different levels of CFCC can be run
concurrently while the CF LPARs are running on different servers. CF LPARs that run on the
same server share the CFCC level.
IBM z16 servers (CFCC level 25) can coexist in a sysplex with CFCC levels 24, 23, and 22;
nevertheless, the latest Coupling Facility Control Code and MCLs levels are always
recommended for best performance and availability:
On IBM z15 (M/T 8561 and 8562): CFCC 24 - Service level 00.22, Bundle S48 released in
August 2021.
On IBM z14 (M/T 3906): CFCC 23 - Service Level 00.20, Bundle S64 released in August
2021.
For a CF LPAR with dedicated processors, the CFCC is implemented by using the active wait
technique. This technique means that the CFCC is always running (processing or searching
for service) and never enters a wait state. Therefore, the CF Control Code uses all the
processor capacity that is available for the CF LPAR.
If the LPAR that is running the CFCC includes only dedicated processors (CPs or ICFs), the
use of all processor capacity (cycles) is not an issue. However, this configuration can be an
issue if the LPAR that is running the CFCC includes shared processors. On IBM z16, Thin
Interrupts is the only valid option for shared engines in a CF LPAR (Thin Interrupts is the
default option on IBM z15).
CF structure sizing changes are expected when moving to CFCC Level 25. Always review the
CF structure size by using the CFSizer tool when changing CFCC levels.
For more information about the recommended CFCC levels, see the current exception letter
that is published on IBM Resource Link®.
The interrupt causes a shared logical processor CF partition to be dispatched by PR/SM (if it
is not already dispatched), which allows the request or signal to be processed in a more
timely manner. The CF relinquishes control when work is exhausted or when PR/SM takes
the physical processor away from the logical processor.
With IBM z16, DYNDISP=THIN is the only mode of operation for CF images that use shared
processors.
This capability allows ICF engines to be shared by several CF images. In this environment, it
provides faster and far more consistent CF service times. It can also provide performance that
is reasonably close to dedicated-engine CF performance.
The use of Thin Interrupts allows a CF to run by using a shared processor while maintaining
good performance. The shared engine is allowed to be undispatched when no more work
exists, as in the past. The Thin Interrupt gets the shared processor that is dispatched when a
command or duplexing signal is presented to the shared engine.
This function saves processor cycles and is an excellent option to be used by a production
backup CF or a testing environment CF. This function is activated by default when a CF
processor is shared.
The CPs can run z/OS operating system images and CF images. For software charging
reasons, generally use only ICF processors to run CF images.
The “storage class memory” that is provided by Flash Express adapters is replaced with
memory that is allocated from main memory (VFM).
VFM helps improve availability and handling of paging workload spikes when running z/OS.
With this support, z/OS is designed to help improve system availability and responsiveness
by using VFM across transitional workload events, such as market openings and diagnostic
data collection.
z/OS also helps improve processor performance by supporting middleware use of pageable
large (1 MB) pages, and eliminates delays that can occur when collecting diagnostic data
during failures.
VFM also can be used in CF images to provide extended capacity and availability for
workloads that use IBM WebSphere MQ Shared Queues structures.
Ordered VFM memory reduces the maximum orderable memory for the model.
VFM provides physical memory DIMMs that are needed to support activation of all customer
purchased memory and HSA on a multiple drawer IBM z16 A01 with one drawer that is done
for the following features:
Scheduled concurrent drawer upgrade, such as memory adds
Scheduled concurrent drawer maintenance, such N+1 repair
Concurrent repair of an out-of-service CPC drawer “fenced” during Activation (POR)
Note: All of these features can be done without VFM. However, all customer-purchased
memory is not available for use in most cases. Some work might need to be shut down or
not restarted.
The information is relocated during CDR in a manner that is identical to the process that was
used for expanded storage. VFM is much simpler to manage (HMC task) and no hardware
repair and verify (no cables and no adapters) are needed. Also, because this feature is part of
internal memory, VFM is protected by RAIM and ECC and can provide better performance
because no I/O to an attached adapter occurs.
Note: Use cases for Flash did not change (for example, z/OS paging and CF shared queue
overflow). Instead, they transparently benefit from the changes in the hardware
implementation.
No option is available for VFM plan ahead. The only option is to always include VFM plan
ahead when Flexible Memory option is selected.
The IBM Secure Service Container (SSC) is a container technology through which you can
more quickly and securely deploy software appliances on IBM z16.
An IBM SSC partition is a specialized container for installing and running specific appliances.
An appliance is an integration of operating system, middleware, and software components
that work autonomously and provide core services and infrastructures that focus on usability
and security.
IBM SSC hosts most sensitive client workloads and applications. It acts as a highly protected
and secured digital vault, enforcing security by encrypting the entire stack: memory, network,
and data (both in-flight and at-rest). Applications that are running inside IBM SSC are isolated
and protected from outsider and insider threats.
IBM SSC combines hardware, software, and middleware and is unique to IBM Z platform.
Though it is called a container, it should not be confused with purely software open source
containers (such as Kubernetes or Docker).
IBM SSC is a part of the Pervasive Encryption concept that was introduced with IBM z14,
which is aimed at delivering best IBM Security hardware and software enhancements,
services, and practices for 360-degree infrastructure protection.
IBM z16 technology provides built-in data encryption with excellent vertical scalability and
performance that protects against data breach threats and data manipulation by privileged
users. IBM SSC is a powerful IBM technology for providing the extra protection of the most
sensitive workloads.
The following IBM solutions and offerings, and more to come, can be deployed in an IBM SSC
environment:
IBM Hyper Protect Virtual Servers (HPVS) solution is available for running Linux-based
virtual servers with sensitive data and applications delivering a confidential computing
environment to address your top security concerns.
For more information, see this IBM Cloud® web page.
IBM Db2 Analytics Accelerator (IDAA) is a high-performance component that is tightly
integrated with Db2 for z/OS. It delivers high-speed processing for complex Db2 queries to
support business-critical reporting and analytic workloads. The accelerator transforms the
mainframe into a hybrid transaction and analytic processing (HTAP) environment.
For more information, see this IBM web page.
IBM Cloud Hyper Protect Data Base as a Service (DBaaS) for PostgreSQL or MongoDB
offers enterprise cloud database environments with high availability for sensitive data
workloads.
For more information, see this IBM Cloud web page.
IBM Cloud Hyper Protect Crypto Services is a key management service and cloud
hardware security module (HSM) that supports industry standards such as PKCS #11.
For more information, see this IBM Cloud web page.
I/O cage, I/O drawer, and PCIe I/O drawer are not supported.
Note: Throughout this chapter, the terms adapter and card refer to a PCIe I/O feature that
is installed in a PCIe+ I/O drawer.
The PCIe I/O infrastructure in IBM z16 A01 consists of the following components:
PCIe+ Gen3 dual port fan-outs that support 16 GBps I/O bus for CPC drawer connectivity
to the PCIe+ I/O drawers. It connects to the PCIe Interconnect Gen3 in the PCIe+ I/O
drawers.
Integrated Coupling Adapter Short Reach (ICA SR and ICA SR1.1), which are PCIe Gen3
features that support short distance coupling links. The ICA SR and ICA SR1.1 features
have two ports, each port supporting 8 GBps.
The 8U, 16-slot, and 2-domain PCIe+ I/O drawer for PCIe I/O features.
The PCIe standard uses a low-voltage differential serial bus. Two wires are used for signal
transmission, and a total of four wires (two for transmit and two for receive) form a lane of a
PCIe link, which is full-duplex. Multiple lanes can be aggregated into a larger link width. PCIe
supports link widths of 1, 2, 4, 8, 12, 16, and 32 lanes (x1, x2, x4, x8, x12, x16, and x32).
The data transmission rate of a PCIe link is determined by the link width (numbers of lanes),
the signaling rate of each lane, and the signal encoding rule. The signaling rate of one PCIe
Generation 3 lane is eight gigatransfers per second (GTps), which means that nearly
8 gigabits are transmitted per second (Gbps).
Note: I/O infrastructure for IBM z16, as for IBM z15, is implemented as PCIe Generation 3.
The PU chip PCIe interface is PCIe Generation 4 (x16 @32 GBps), but the CPC I/O
fan-out infrastructure provides external connectivity as PCIe Generation 3 @16GBps
A PCIe Gen3 x16 link features the following data transmission rates:
The maximum theoretical data transmission rate per lane:
8 Gbps * 128/130 bit (encoding) = 7.87 Gbps=984.6 MBps
The maximum theoretical data transmission rate per link:
984.6 MBps * 16 (lanes) = 15.75 GBps
Link performance: The link speeds do not represent the performance of the link. The
performance depends on many factors, including latency through the adapters, cable
lengths, and the type of workload.
PCIe Gen3 x16 links are used in IBM z16 servers for driving the PCIe+ I/O drawers, and for
coupling links for CPC to CPC communications.
Note: Unless specified otherwise, PCIe refers to PCIe Generation 3 in remaining sections
of this chapter.
4.2.1 Characteristics
The IBM z16 A01 I/O subsystem provides great flexibility, high availability, and the following
excellent performance characteristics:
High bandwidth
IBM z16 servers use PCIe Gen3 protocol to drive PCIe+ I/O drawers and CPC to CPC
(coupling) connections. The I/O bus infrastructure data rate of up to 128 GBps1 per
system (12 PCIe+ Gen3 fan-out slots).
For more information about coupling link connectivity, see 4.6.4, “Parallel Sysplex
connectivity” on page 187.
Connectivity options:
– IBM z16 servers can be connected to an extensive range of interfaces, such as
FICON/FCP for SAN connectivity, OSA features for LAN connectivity, and zHyperLink
Express for storage connectivity (low latency compared to FICON).
– For CPC to CPC connections, IBM z16 servers use Integrated Coupling Adapter (ICA
SR and ICA SR 1.1) and the Coupling Express2 Long Reach (CE2 LR).
The Parallel Sysplex InfiniBand is not supported.
– The 25 GbE and 10 GbE RoCE Express3, 25 GbE and 10 GbE RoCE Express2.1, and
25 GbE and 10 GbE RoCE Express2 provide high-speed memory-to-memory data
exchange to a remote CPC by using the Shared Memory Communications over RDMA
(SMC-R) protocol for TCP (socket-based) communications.
The RoCE Express3 features also can provide local area network (LAN) connectivity
for Linux on IBM Z, and comply with IEEE standards. In addition, RoCE Express
features assume several functions of the TCP/IP stack that normally are performed by
the PU, which allows significant performance benefits by offloading processing from
the operating system.
1
The link speeds do not represent the performance of the link. The performance depends on many factors, including
latency through the adapters, cable lengths, and the type of workload.
2
OSA-Express 1000BASE-T features do not have optics (copper only, RJ45 connectors).
3
Multifiber Termination Push-On.
4 For more information, see https://cw.infinibandta.org/document/dl/7157.
PCIe switch application-specific integrated circuits (ASICs) are used to fan out the host bus
from the CPC drawer through the PCIe+ I/O drawer to the individual I/O features. Maximum
16 PCIe I/O features (up to 32 channels) per PCIe+ I/O drawer are supported.
The PCIe+ I/O drawer is a one-sided drawer (all I/O cards on one side, in the rear of the
drawer) that is 8U high. The PCIe+ I/O drawer contains the 16 I/O slots for PCIe features, two
switch cards, and two power supply units (PSUs) to provide redundant power, as shown in
Figure 4-1.
Figure 4-3 IBM z16 I/O connectivity(Max39 feature with two PCIe+ I/O drawers represented
The PCIe switch card provides the fan-out from the high-speed x16 PCIe host bus to eight
individual card slots. The PCIe switch card is connected to the CPC drawer through a single
x16 PCIe Gen3 bus from a PCIe fan-out card (PCIe+ fan-out cards on IBM z16 have two PCIe
Gen3 x16 ports/busses/links).
The PCIe slots in a PCIe+ I/O drawer are organized into two I/O domains. Each I/O domain
supports up to eight features and is driven through a PCIe switch card. Two PCIe switch cards
always provide a backup path for each other through the passive connection in the PCIe+ I/O
drawer backplane.
During a PCIe fan-out card or cable failure, 16 I/O cards in two domains can be driven through
a single PCIe switch card. It is not possible to drive 16 I/O cards after one of the PCIe switch
cards is removed.
The two switch cards are interconnected through the PCIe+ I/O drawer board (Redundant I/O
Interconnect, or RII). In addition, switch cards in same PCIe+ I/O drawer are connected to
PCIe fan-outs across clusters in CPC drawer for higher availability.
The RII design provides a failover capability during a PCIe fan-out card failure. Both domains
in one of these PCIe+ I/O drawers are activated with two fan-outs. The Base Management
Cards (BMCs) are used for system control.
The domains and their related I/O slots are shown in Figure 4-2 on page 152.
Each I/O domain supports up to eight features (FICON, OSA, Crypto, and so on.) All I/O
cards connect to the PCIe switch card through the backplane board. The I/O domains and
slots are listed in Table 4-1.
Only the following PCIe features can be carried forward for an upgrade to IBM z16 A01
servers:
FICON Express16SA
FICON Express16S+
zHyperLink Express (all features)
OSA-Express7S 25 GbE SR1.1
OSA-Express7S (all features)
OSA-Express6S (all features)
25 GbE RoCE Express2.1
25 GbE RoCE Express2
10 GbE RoCE Express2.1
10 GbE RoCE Express2
Crypto Express7S(one or two ports/HSMs)
Crypto Express6S
IBM z16 A01 server supports the following PCIe I/O new features that are hosted in the
PCIe+ I/O drawers:
FICON Express32S
OSA-Express7S 1.2 25 GbE
OSA-Express7S 1.2 10 GbE
OSA-Express7S 1.2 GbE
OSA-Express7S 1.2 1000BASE-T
25 GbE RoCE Express3
10 GbE RoCE Express3
Crypto Express8S (one or two HSMs)
Coupling Express2 Long Reach (CE2 LR)
zHyperLink Express1.1
The IBM z16 CPC drawer I/O infrastructure consists of the following features:
The PCIe+ Generation 3 fan-out cards: Two ports per card (feature) that connect to PCIe+
I/O drawers.
ICA SR (ICA SR and ICA SR1.1) fan-out cards: Two ports per card (feature) that connect
to other (external) CPCs.
Note: IBM z16 does not support Parallel Sysplex InfiniBand (PSIFB) links.
Also, if and IBM z16 is part of a Parallel Sysplex or Coordinated Timing Network, InfiniBand
links cannot be used on older servers, even if installed.
Unless otherwise noted, ICA SR is used for ICA SR and ICA SR1.1 for the rest of the
chapter.
The PCIe fan-outs cards are installed in the rear of the CPC drawers. Each CPC drawer
features 12 PCIe+ Gen3 fan-out slots.
The PCIe fan-outs and ICA SR fan-outs are installed in locations LG01 - LG12 at the rear in
the CPC drawers (see Figure 2-8 on page 25).
On the CPC drawer, two BMC/OSC cards are available, each being a combination of a Base
Management (BMC) card and an Oscillator Card (OSC) card. Each BMC/OSC card features
one PPS port and one ETS port (RJ45 Ethernet) for both PTP and NTP.
A 16x PCIe copper cable of 1.5 meters (4.92 feet) - 4.0 meters (13.1 feet) is used for
connection to the PCIe switch card in the PCIe+ I/O drawer. PCIe fan-out cards are always
plugged in pairs and provide redundancy for I/O domains within the PCIe+ I/O drawer.
Note: The PCIe fan-out is used exclusively for I/O and cannot be shared for any other
purpose.
The ICA SR (FC 0172) and ICA SR1.1 (FC 0176) use PCIe Gen3 technology, with x16 lanes
that are bifurcated into x8 lanes for coupling.
Both cards are designed to drive distances up to 150 meters (492 feet) with a link data rate of
8 GBps. ICA SR supports up to four channel-path identifiers (CHPIDs) per port and eight
subchannels (devices) per CHPID.
The coupling links can be defined as shared between images (z/OS) within a CSS. They also
can be spanned across multiple CSSs in a CPC. For ICA SR features, a maximum four
CHPIDs per port can be defined.
When STP5 (FC 1021) is available, ICA SR coupling links can be defined as timing-only links
to other IBM z16, IBM z15, IBM z14 ZR1, or IBM z14 M0x systems. The ICA SR cannot be
connected to InfiniBand coupling fan-outs.
These two fan-out features are housed in the PCIe+ Gen3 I/O fan-out slot on the IBM z16
CPC drawers. Up 48 ICA SR and ICA SR1.1 features (up to 96 ports) are supported on an
IBM z16 A01 system.
OM3 fiber optic can be used for distances up to 100 meters (328 feet). OM4 fiber optic cables
can be used for distances up to 150 meters (492 feet). For more information, see the following
publications:
Planning for Fiber Optic Links, GA23-1409
3931 Installation Manual for Physical Planning, GC28-7015
Table 4-2 Fan-out locations and their AIDs for the CPC drawer (IBM z16 A01)
Fan-out CPC0 CPC1 CPC2 CPC3
locations Location A10 Location A15 Location A20 Location B10
AID (Hex) AID (Hex) AID (Hex) AID (Hex)
LG01 00 0C 18 24
LG02 01 0D 19 25
LG03 02 0E 1A 26
LG04 03 0F 1B 27
LG05 04 10 1C 28
LG06 05 11 1D 29
LG07 06 12 1E 2A
LG08 07 13 1F 2B
LG09 08 14 20 2C
LG10 09 15 21 2D
LG11 0A 16 22 2E
LG12 0B 17 23 2F
Fan-out slots
The fan-out slots are numbered LG01 - LG12 (from left to right), as listed in Table 4-2. All
fan-out locations and their AIDs for the CPC drawer are shown for reference only.
Important: The AID numbers that are listed in Table 4-2 are valid for a new build system
only. If a fan-out is moved, the AID follows the fan-out to its new physical location.
The AID assignment is listed in the PCHID REPORT that is provided for each new server or
for an MES upgrade on servers. Part of a PCHID REPORT for an IBM z16 is shown in
Example 4-1 on page 157. In this example, four fan-out cards are installed at locations
A10/LG06, A15/LG06, A15/LG12, and A15/LG12 with AIDs 05, 11, and PCHIDs 100, 101,
and 10C, respectively.
Fan-out features that are supported by the IBM z16 are listed in Table 4-3, which includes the
feature type, feature code, and information about the link that is supported by the fan-out
feature.
PCIe+ Gen3 0175 PCIe I/O Copper N/A 4 m (13.1 ft.) 16 GBps
fan-out drawer conn.
ICA SR 0172 Coupling link OM4 MTP 150 m (492 ft.) 8 Gbps
ICA SR1.1 0176 Coupling link OM4 MTP 150 m (492 ft.) 8 Gbps
6 Certain I/O features do not have external ports, such as Crypto Express.
A PCHID REPORT is created for each new build server and for upgrades on servers. The
report lists all I/O features that are installed, the physical slot location, and the assigned
PCHID. A portion of a sample PCHID REPORT is shown in Example 4-2.
The PCHID REPORT that is shown in Example 4-2 includes the following components:
Feature Code 0176 (Integrated Coupling Adapters (ICA SR1.1) is installed in the CPC
drawer (location A10, slot LG06, and A15 LG06), and has AIDs 05 and 11 assigned.
Feature 0461 (FICON Express32SA LX) is installed in PCIe+ I/O drawer 1: Location Z01B,
slot 02 with PCHIDs 100/D1 and 101/D2 assigned.
Two feature codes 0457 (OSA-Express7S 1.2 10 GbE SR) installed in PCIe+ I/O drawer 1
in slots 05 and 07, with PCHIDs 10C/D1 and 110/D1, respectively.
A resource group (RG) parameter also is shown in the PCHID REPORT for native PCIe
features. A balanced plugging of native PCIe features exists between four resource groups
(RG1, RG2, RG3, and RG4).
The preassigned PCHID number of each I/O port relates directly to its physical location (jack
location in a specific slot).
For more information about connectivity to external I/O subsystems (for example, disks), see
4.6.2, “Storage connectivity” on page 165.
For more information about communication to LANs, see 4.6.3, “Network connectivity” on
page 173.
Storage access
OSA-Express featuresd
Coupling features
At least one I/O feature (FICON) or one coupling link feature (ICA SR or CE LR) must be
present in the minimum configuration.
Cables: All fiber optic cables, cable planning, labeling, and installation are client
responsibilities for new IBM z16 installations and upgrades. Fiber optic conversion kits and
mode conditioning patch cables are not orderable as features on IBM z16 servers. All other
cables must be sourced separately.
The required connector and cable type for each I/O feature on IBM z16 servers are listed in
Table 4-6.
FICON channels
IBM z16 supports the following FICON features:
FICON Express32S LX and SX (FC 0461/0462)
FICON Express16SA LX and SX (FC 0436/0437)
FICON Express16S+ LX and SX (FC 0427/0428)
The ICON Express32S, FICON Express16SA, and FICON Express16S+ features conform to
the following architectures:
Fibre Connection (FICON)
High Performance FICON on Z (zHPF)
Fibre Channel Protocol (FCP)
The FICON features provide connectivity between any combination of servers, directors,
switches, and devices (control units, disks, tapes, and printers) in a SAN.
Each FICON Express feature occupies one I/O slot in the PCIe+ I/O drawer. Each feature
includes two ports, each supporting an LC Duplex connector, with one PCHID and one
CHPID that is associated with each port.
Each FICON Express feature uses SFP (SFP+ for FICON Express16SA and FICON
Express32S) optics that allow for concurrent repairing or replacement for each SFP. The data
flow on the unaffected channels on the same feature can continue. A problem with one
FICON Express port does not require replacement of a complete feature.
Each FICON Express feature also supports cascading, which is the connection of two FICON
Directors in succession. This configuration minimizes the number of cross-site connections
and helps reduce implementation costs for disaster recovery applications, IBM
Geographically Dispersed Parallel Sysplex (GDPS), and remote copy.
IBM z16 servers support 32 K devices per FICON channel for all FICON features.
Each FICON Express channel can be defined independently for connectivity to servers,
switches, directors, disks, tapes, and printers, by using the following CHPID types:
CHPID type FC: The FICON, zHPF, and FCTC protocols are supported simultaneously.
CHPID type FCP: Fibre Channel Protocol that supports attachment to SCSI devices
directly or through Fibre Channel switches or directors.
FICON channels (CHPID type FC or FCP) can be shared among LPARs and defined as
spanned. All ports on a FICON feature must be of the same type (LX or SX). The features are
connected to a FICON capable control unit (point-to-point or switched point-to-point) through
a Fibre Channel switch.
The following types of FICON Express32S optical transceivers are supported (no mix on
same card):
FICON Express32S LX feature, FC 0461, with two ports per feature, LC Duplex
connectors
FICON Express32S SX feature, FC 0462, with two ports per feature, LC Duplex
connectors
For more information about supported distances, see Table 4-6 on page 163.
FICON Express16SA
The FICON Express16SA feature is installed in the PCIe+ I/O drawer. Each of the two
independent ports is capable of 8 or 16 Gbps. The link speed depends on the capability of the
attached switch or device. The link speed is auto-negotiated, point-to-point, and is transparent
to users and applications.
The following types of FICON Express16SA optical transceivers are supported (no mix on
same card):
FICON Express16SA LX feature, FC 0436, with two ports per feature, LC Duplex
connectors
FICON Express16SA SX feature, FC 0437, with two ports per feature, LC Duplex
connectors
For more information about supported distances, see Table 4-6 on page 163.
FICON Express16S+
The FICON Express16S+ feature is installed in the PCIe+ I/O drawer. Each of the two
independent ports is capable of 4, 8, or 16 Gbps. The link speed depends on the capability of
the attached switch or device. The link speed is auto-negotiated, point-to-point, and is
transparent to users and applications.
The following types of FICON Express16S+ optical transceivers are supported (no mix on
same card):
FICON Express16S+ LX feature, FC 0427, with two ports per feature, supporting LC
Duplex connectors
FICON Express16S+ SX feature, FC 0428, with two ports per feature, supporting LC
Duplex connectors
The following prerequisites must be met for this optionally priced feature:
FICON Express32S and FICON Express16SA for both link encryption and endpoint
authentication
The FICON Express32S, FICON Express16SA, and FICON Express16S+ support FEC
coding on top of its 64b/66b data encoding for 16 and 32 Gbps connections. This design can
correct up to 11 bit errors per 2112 bits transmitted.
Therefore, while connected to devices that support FEC at 16 Gbps connections, the FEC
design allows FICON Express16SE, FICON Express16S+, and FICON Express16S channels
to operate at higher speeds, over longer distances, with reduced power and higher
throughput. At the same time, the same reliability and robustness for which FICON channels
are traditionally known are maintained.
With the IBM DS8870 or newer, IBM z16 servers can extend the use of FEC to the fabric
N_Ports for a completed end-to-end coverage of 32 Gbps FC links.
A static SAN routing policy normally assigns the ISL routes according to the incoming port
and its destination domain (port-based routing), or the source and destination ports pairing
(device-based routing).
The port-based routing (PBR) assigns the ISL routes statically that is based on “first come,
first served” when a port starts a fabric login (FLOGI) to a destination domain. The ISL is
round-robin that is selected for assignment. Therefore, I/O flow from same incoming port to
same destination domain always is assigned the same ISL route, regardless of the
destination port of each I/O.
This setup can result in some ISLs becoming overloaded while some are under-used. The
ISL routing table is changed whenever IBM Z server undergoes a power-on-reset (POR), so
the ISL assignment is unpredictable.
Device-based routing (DBR) assigns the ISL routes statically that is based on a hash of the
source and destination port. That I/O flow from same incoming port to same destination is
assigned to same ISL route. Compared to PBR, the DBR is more capable of spreading the
load across ISLs for I/O flow from the same incoming port to different destination ports within
a destination domain.
When a static SAN routing policy is used, the FICON director features limited capability to
assign ISL routes that are based on workload. This limitation can result in unbalanced use of
ISLs (some might be overloaded, while others are under-used).
The dynamic routing ISL routes are dynamically changed based on the Fibre Channel
exchange ID, which is unique for each I/O operation. ISL is assigned at I/O request time;
therefore, different I/Os from same incoming port to same destination port are assigned
different ISLs.
With FIDR, IBM z16 servers feature the following advantages for performance and
management in configurations with ISL and cascaded FICON directors:
Support sharing of ISLs between FICON and FCP (PPRC or distributed)
I/O traffic is better balanced between all available ISLs
Improved use of FICON director and ISL
Easier to manage with a predicable and repeatable I/O performance
FICON dynamic routing can be enabled by defining dynamic routing-capable switches and
control units in HCD. Also, z/OS implemented a health check function for FICON dynamic
routing.
IBM z14 and newer servers can read this extra diagnostic data for all the ports that are
accessed in the I/O configuration and make the data available to an LPAR. For z/OS LPARs
that use FICON channels, z/OS displays the data with a new message and display command.
For Linux on IBM Z, z/VM, and z/VSE, and LPARs that use FCP channels, this diagnostic
data is available in a new window in the SAN Explorer tool.
N_Port ID Virtualization
N_Port ID Virtualization (NPIV) allows multiple system images (in LPARs or z/VM guests) to
use a single FCP channel as though each were the sole user of the channel. First introduced
with IBM z9® EC, this feature can be used with earlier FICON features that were carried
forward from earlier servers.
By using the FICON Express16S (or newer) as an FCP channel with NPIV enabled, the
maximum numbers of the following aspects for one FCP physical channel are doubled:
Maximum number of NPIV hosts defined: Increased from 32 to 64
Maximum number of remote N_Ports communicated: Increased from 512 to 1024
Maximum number of addressable LUNs: Increased from 4096 to 8192
Concurrent I/O operations: Increased from 764 to 1528
For more information about operating systems that support NPIV, see “N_Port ID
Virtualization” on page 296.
IBM z14 and newer servers allow for the modification of these default assignments, which
also allows FCP channels to keep previously assigned WWPNs, even after being moved to a
different slot position. This capability can eliminate the need for reconfiguration of the SAN in
many situations, and is especially helpful during a system upgrade (FC 0099 - WWPN
Persistence).
IBM z14 and newer servers support up to three hops in a cascaded FICON SAN
environment. This support allows clients to more easily configure a three- or four-site disaster
recovery solution.
For more information about the FICON multi-hop, see the FICON Multihop: Requirements
and Configurations white paper at the IBM Techdocs Library website.
The zHyperLink Express1.1 feature (FC 0451) provides a low latency direct connection
between IBM z16 and DS8000 storage system.
The zHyperLink Express1.1 is the result of new business requirements that demand fast and
consistent application response times. It dramatically reduces latency by interconnecting the
IBM z16 directly to I/O Bay of the DS8k by using PCIe Gen3 x 8 physical link (up to
150 meters [492 feet]). A new transport protocol is defined for reading and writing IBM CKD
data records9, as documented in the zHyperLink interface specification.
On IBM z16, zHyperLink Express1.1 card is a PCIe Gen3 adapter, which installed in the
PCIe+ I/O drawer. HCD definition support was added for new PCIe function type with PORT
attributes.
The zHyperLink Express1.1 is virtualized as a native PCIe adapter and can be shared by
multiple LPARs. Each port can support up to 127 Virtual Functions (VFs), with one or more
VFs/PFIDs being assigned to each LPAR. This configuration gives a maximum of 254 VFs
per adapter.
9 CKD data records are handled by using IBM Enhanced Count Key Data (ECKD) command set.
The zHyperLink Express feature (FC 0431) provides a low latency direct connection between
IBM z16 and DS8000 I/O Port.
The zHyperLink Express is the result of new business requirements that demand fast and
consistent application response times. It dramatically reduces latency by interconnecting the
IBM z16 directly to I/O Bay of the DS8880 by using PCIe Gen3 x 8 physical link (up to
150 meters [492 feet]). A new transport protocol is defined for reading and writing IBM CKD
data records10, as documented in the zHyperLink interface specification.
On IBM z16, the zHyperLink Express card is a carry forward PCIe adapter that installed in the
PCIe+ I/O drawer. HCD definition support was added for new PCIe function type with PORT
attributes.
Requirements of zHyperLink
The zHyperLink Express feature is available on IBM z16 servers, and includes the following
requirements:
z/OS 2.2 or later
Supported DS8000 (see Getting Started with IBM zHyperLink for z/OS, REDP-5493)
zHyperLink Express adapter (FC 0431) installed
FICON channel as a driver
Only ECKD supported
z/VM is not supported
The zHyperLink Express is virtualized as a native PCIe adapter and can be shared by
multiple LPARs. Each port can support up to 127 Virtual Functions (VFs), with one or more
VFs/PFIDs being assigned to each LPAR. This configuration gives a maximum of 254 VFs
per adapter.
The zHyperLink Express uses optical cable with MTP connector. Maximum supported cable
length is 150 meters (492 feet).
10 CKD data records are handled by using IBM Enhanced Count Key Data (ECKD) command set.
OSA-Express7S 1.2 25 Gigabit Ethernet Short Reach (SR) feature includes one PCIe Gen3
adapter and one port per feature. The port supports CHPID types OSD.
The OSA-Express7S 1.2 25 GbE SR feature supports the use of an industry standard small
form factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device
includes an SR transceiver. The sending and receiving transceivers must be the same (SR to
SR).
The OSA-Express7S 1.2 25 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 µm multimode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express7S 1.2 25 GbE LR feature supports the use of an industry standard small
form factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device
includes an LR transceiver. The transceivers at both ends must be the same (LR to LR).
The OSA-Express7S 1.2 25 GbE LR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
The OSA-Express7S 1.2 10 GbE LR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
The OSA-Express7S 1.2 10 GbE SR feature supports the use of an industry standard small
form factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device
includes an SR transceiver. The sending and receiving transceivers must be the same (SR to
SR).
The OSA-Express7S 1.2 10 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 or a 62.5 µm multimode fiber optic cable that ends with an LC Duplex connector is
required for connecting each port on this feature to the selected device.
The OSA-Express7S 1.2 GbE LX feature supports the use of an LC Duplex connector.
Ensure that the attaching or downstream device has an LX transceiver. The sending and
receiving transceivers must be the same (LX to LX).
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device. If multimode fiber optic cables are
being reused, a pair of Mode Conditioning Patch cables is required, with one cable for each
end of the link.
The OSA-Express7S 1.2 GbE SX feature supports the use of an LC Duplex connector.
Ensure that the attaching or downstream device has an SX transceiver. The sending and
receiving transceivers must be the same (SX to SX).
A multi-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express7S 1.2 1000BASE-T Ethernet feature does not support auto-negotiation. It
supports links at 1000 Mbps in full duplex mode only.
The OSA-Express7S 1.2 1000BASE-T Ethernet feature can be configured as CHPID type
OSC, OSD, OSE. Non-QDIO operation mode requires CHPID type OSE.
Note: CHPID types OSM, OSN, and OSX are not supported on IBM z16.
OSA-Express7S 25 Gigabit Ethernet Short Reach1.1 (SR1.1) feature includes one PCIe
Gen3 adapter and one port per feature. The port supports CHPID types OSD.
The OSA-Express7S 25 GbE SR1.1 feature supports the use of an industry standard small
form factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device
includes an SR transceiver. The sending and receiving transceivers must be the same (SR to
SR).
The OSA-Express7S 25 GbE SR1.1 feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 µm multimode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express7S 25 GbE Short Reach (SR) feature includes one PCIe adapter and one
port per feature. The port supports CHPID types OSD. The OSA-Express7S 25 GbE feature
is designed to support attachment to a multimode fiber 25 Gbps Ethernet LAN or Ethernet
switch that is capable of 25 Gbps. The port can be defined as a spanned channel and shared
among LPARs within and across logical channel subsystems.
The OSA-Express7S 25 GbE SR feature supports the use of an industry standard small form
factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device has an
SR transceiver. The sending and receiving transceivers must be the same (SR to SR).
The OSA-Express7S 25 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
The following other OSA-Express7S features can be installed on IBM z16 servers (when
carried forward):
OSA-Express7S 10 Gigabit Ethernet LR, FC 0444
OSA-Express7S 10 Gigabit Ethernet SR, FC 0445
OSA-Express7S Gigabit Ethernet LX, FC 0442
OSA-Express7S Gigabit Ethernet SX, FC 0443
OSA-Express7S 1000BASE-T Ethernet, FC 0446
The supported OSA-Express7S features are listed in Table 4-5 on page 161.
The OSA-Express7S 10 GbE LR feature supports the use of an industry standard small form
factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device includes
an LR transceiver. The transceivers at both ends must be the same (LR to LR).
The OSA-Express7S 10 GbE LR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
The OSA-Express7S 10 GbE SR feature supports the use of an industry standard small form
factor (SFP+) LC Duplex connector. Ensure that the attaching or downstream device has an
SR transceiver. The sending and receiving transceivers must be the same (SR to SR).
The OSA-Express7S 10 GbE SR feature does not support auto-negotiation to any other
speed and runs in full duplex mode only.
A 50 or a 62.5 µm multimode fiber optic cable that ends with an LC Duplex connector is
required for connecting each port on this feature to the selected device.
The OSA-Express7S GbE LX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device includes an LX transceiver. The sending and
receiving transceivers must be the same (LX to LX).
The OSA-Express7S GbE SX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device has an SX transceiver. The sending and receiving
transceivers must be the same (SX to SX).
A multi-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
The OSA-Express7S 1000BASE-T Ethernet feature can be configured as CHPID type OSC,
OSD, OSE, or OSM. Non-QDIO operation mode requires CHPID type OSE.
Note: CHPID types OSM, OSN, and OSX are not supported on IBM z16.
OSA-Express6S
The OSA-Express6S feature is installed in the PCIe+ I/O drawer. The following
OSA-Express6S features can be installed on IBM z16 servers (when carried forward):
OSA-Express6S 10 Gigabit Ethernet LR, FC 0424
OSA-Express6S 10 Gigabit Ethernet SR, FC 0425
OSA-Express6S Gigabit Ethernet LX, FC 0422
OSA-Express6S Gigabit Ethernet SX, FC 0423
OSA-Express6S 1000BASE-T Ethernet, FC 0426
The supported OSA-Express6S features are listed in Table 4-8 on page 179.
The OSA-Express6S 10 GbE LR feature supports the use of an industry standard small form
factor LC Duplex connector. Ensure that the attaching or downstream device includes an LR
transceiver. The transceivers at both ends must be the same (LR to LR).
A 9 µm single-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting this feature to the selected device.
For more information about supported distances, see Table 4-8 on page 179.
The OSA-Express6S GbE SX feature supports the use of an LC Duplex connector. Ensure
that the attaching or downstream device includes an SX transceiver. The sending and
receiving transceivers must be the same (SX to SX).
A multi-mode fiber optic cable that ends with an LC Duplex connector is required for
connecting each port on this feature to the selected device.
For more information about supported distances, see Table 4-8 on page 179.
The OSA-Express6S 1000BASE-T Ethernet feature can be configured as CHPID type OSC,
OSD, OSE, or OSM. Non-QDIO operation mode requires CHPID type OSE.
Note: CHPID types OSM, OSN, and OSX are not supported on IBM z16
The following settings are supported on the OSA-Express6S 1000BASE-T Ethernet feature
port:
Auto-negotiate
100 Mbps half-duplex or full-duplex
1000 Mbps full-duplex
If auto-negotiate is not used, the OSA-Express port attempts to join the LAN at the specified
speed and duplex mode. If this specified speed and duplex mode do not match the speed and
duplex mode of the signal on the cable, the OSA-Express port does not connect.
For more information about supported distances, see Table 4-8 on page 179.
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 25 GbE switch or to another 25 GbE RoCE Express3 SR feature:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 25 GbE RoCE Express3 feature uses SR optics and
supports the use of a multimode fiber optic cable that ends with an LC Duplex connector.
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 25 GbE RoCE Express3 feature uses LR optics and
supports the use of a single mode fiber optic cable that ends with an LC Duplex connector.
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 10 GbE switch or to another 10 GbE RoCE Express3 SR feature:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 10 GbE RoCE Express3 feature uses SR optics and
supports the use of a multimode fiber optic cable that ends with an LC Duplex connector.
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 10 GbE RoCE Express3 feature uses LR optics and
supports the use of a single mode fiber optic cable that ends with an LC Duplex connector.
Switch configuration for 25 GbE RoCE Express2.1: The switches must meet the
following requirements:
Global Pause function is enabled
Priority flow control (PFC) is disabled
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 10 GbE switch or to another 10 GbE RoCE Express3 SR feature:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 25 GbE RoCE Express2.1 feature uses SR optics and
supports the use of a multimode fiber optic cable that ends with an LC Duplex connector.
Switch configuration for 10 GbE RoCE Express2.1: The 10 GbE switches must meet
the following requirements:
Global Pause function is enabled
Priority flow control (PFC) is disabled
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 10 GbE switch or to another 10 GbE RoCE Express3 SR feature:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
Switch configuration for RoCE Express2: If the IBM 25 GbE RoCE Express2 features
are connected to 25 GbE switches, the switches must meet the following requirements:
Global Pause function is enabled
Priority flow control (PFC) is disabled
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 10 GbE switch or to another 10 GbE RoCE Express3 SR feature:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 25 GbE RoCE Express2.1 feature uses SR optics and
supports the use of a multimode fiber optic cable that ends with an LC Duplex connector.
The maximum supported unrepeated distance, point-to-point, is 100 meters (328 feet). A
client-supplied cable is required. Two types of cables can be used for connecting the port to
the selected 10 GbE switch or to another 10 GbE RoCE Express3 SR feature:
OM3 50-micron multimode fiber optic cable that is rated at 2000 MHz-km that ends with an
LC Duplex connector, which supports 70 meters (229 feet)
OM4 50-micron multimode fiber optic cable that is rated at 4700 MHz-km that ends with an
LC Duplex connector, which supports 100 meters (328 feet)
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID). The 10 GbE RoCE Express2 feature uses SR optics and
supports the use of a multimode fiber optic cable that ends with an LC Duplex connector.
ISM is defined by the FUNCTION statement with a virtual CHPID (VCHID) in HCD/IOCDS.
Identified by the PNETID parameter, each ISM VCHID defines an isolated, internal virtual
network for SMC-D communication, without any hardware component required. Virtual
adapters are defined by virtual function (VF) statements. Multiple LPARs can access the
same virtual network for SMC-D data exchange by associating their VF with same VCHID.
Applications that use HiperSockets can realize network latency and CPU reduction benefits
and performance improvement by using the SMC-D over ISM.
IBM z16 servers support up to 32 ISM VCHIDs per CPC. Each VCHID supports up to 255
VFs, with a total maximum of 8,000 VFs.
The initial version of SMC was limited to TCP/IP connections over the same layer 2 network;
therefore, it was not routable across multiple IP subnets. The associated TCP/IP connection
was limited to hosts within a single IP subnet that requires the hosts to have direct access to
the same physical layer 2 network (that is, the same Ethernet LAN over a single VLAN ID).
The scope of eligible TCP/IP connections for SMC was limited to and defined by the single IP
subnet.
SMC Version 2 (SMCv2) provides support for SMC over multiple IP subnets for SMC-D and
SMC-R and is referred to as SMC-Dv2 and SMC-Rv2. SMCv2 requires updates to the
underlying network technology. SMC-Dv2 requires ISMv2 and SMC-Rv2 requires RoCEv2.
The SMCv2 protocol is downward compatible, which allows SMCv2 hosts to continue to
communicate with SMCv1 down-level hosts.
Although SMCv2 changes the SMC connection protocol enabling multiple IP subnet support,
SMCv2 does not change how user TCP socket data is transferred, which preserves the
benefits of SMC to TCP workloads.
TCP/IP connections that require IPsec are not eligible for any form of SMC.
HiperSockets
The HiperSockets function of IBM z16 servers provides up to 32 high-speed virtual LAN
attachments.
HiperSockets eliminates the need to use I/O subsystem operations and traverse an external
network connection to communicate between LPARs in the same IBM z16 CPC.
Traffic can be IPv4 or IPv6, or non-IP, such as AppleTalk, DECnet, IPX, NetBIOS, or SNA.
Layer 2 support helps facilitate server consolidation, and can reduce complexity and simplify
network configuration. It also allows LAN administrators to maintain the mainframe network
environment similarly to nonmainframe environments.
Packet forwarding decisions are based on Layer 2 information instead of Layer 3. The
HiperSockets device can run automatic MAC address generation to create uniqueness within
and across LPARs and servers. The use of Group MAC addresses for multicast is supported,
and broadcasts to all other Layer 2 devices on the same HiperSockets networks.
Datagrams are delivered only between HiperSockets devices that use the same transport
mode. A Layer 2 device cannot communicate directly to a Layer 3 device in another LPAR
network. A HiperSockets device can filter inbound datagrams by VLAN identification, the
destination MAC address, or both.
HiperSockets Layer 2 is supported by Linux on IBM Z, and by z/VM for Linux guest use.
IBM z16 supports the HiperSockets Completion Queue function that is designed to allow
HiperSockets to transfer data synchronously (if possible) and asynchronously, if necessary.
This feature combines ultra-low latency with more tolerance for traffic peaks.
With the asynchronous support, data can be temporarily held until the receiver has buffers
that are available in its inbound queue during high volume situations. The HiperSockets
Completion Queue function requires the following minimum applications11:
z/OS V2.2 with PTFs
Linux on IBM Z distributions:
– Red Hat Enterprise Linux (RHEL) 6.2
– SUSE Linux Enterprise Server (SLES) 11 SP2
– Ubuntu server 16.04 LTS
z/VSE V6.2
z/VM V6.412 with maintenance
11
Minimum OS support for IBM z16 can differ. For more information, see Chapter 7, “Operating system support” on
page 247.
12 z/VM V6 is not supported on IBM z16. z/VM V7.1 or newer is needed.
This section describes coupling link features supported in a Parallel Sysplex in which an IBM
z16 can participate.
Coupling links
The type of coupling link that is used to connect a CF to an operating system LPAR is
important. The link performance significantly affects response times and coupling processor
usage. For configurations that extend over large distances, the time that is spent on the link
can be the largest part of the response time.
Note: Parallel Sysplex supports connectivity between systems that differ by up to two
generations (n-2). For example, an IBM z16 can participate in an IBM Parallel Sysplex
cluster with IBM z15, and IBM z14 systems.
Only Integrated Coupling Adapter Short Reach (ICA SR) and Coupling Express2 Long
Reach (CE2 LR) features are supported on IBM z16.
Figure 4-4 shows the supported Coupling Link connections for the IBM z16. Only ICA SR and
CE LR links are supported on IBM z16, IBM z15, and IBM z14 ZR1 systems.
13
IBM z14 M0x (M/T 3906) also supports InfiniBand coupling links; however, these links are not supported on IBM
z16, IBM z15, and IBM z14 ZR1. Careful connectivity planning is needed if InfiniBand links are present in the
supported systems.
Table 4-10 Coupling link options that are supported on IBM z16
Type Description Feature Link rate Max unrepeated Maximum number of supported
Code distance links
The maximum number of combined external coupling links (active CE LR, ICA SR links) is
160 per IBM z16 A01 system. IBM z16 systems support up to 384 coupling CHPIDs per CPC.
An IBM z16 coupling link support summary is shown in Table 4-10.
For more information about distance support for coupling links, see IBM Z End-to-End
Extended Distance Guide, SG24-8047.
An IC link is a fast coupling link that uses memory-to-memory data transfers. IC links do not
have PCHID numbers, but do require CHPIDs.
IC links are enabled by defining CHPID type ICP. A maximum of 64 IC links can be defined on
an IBM z16.
ICA SR and ICA SR1.1 are two-port, short-distance coupling features that allow the
supported IBM Z to connect to each other. ICA SR and ICA SR1.1 use coupling channel type:
CS5.
The ICA SR uses PCIe Gen3 technology, with x16 lanes that are bifurcated into x8 lanes for
coupling. ICA SR1.1 uses PCIe Gen4 technology, with x16 lanes that are bifurcated into x8
lanes for coupling.
The ICA SR and SR1.1 are designed to drive distances up to 150 m (492 feet) and supports a
link data rate of 8 GBps. It is designed to support up to four CHPIDs per port and eight
subchannels (devices) per CHPID.
For more information, see IBM Z Planning for Fiber Optic Links (FICON/FCP, Coupling Links,
and Open System Adapters), GA23-1409. This publication is available in the Library section
of Resource Link (log-in required).
The Coupling Express2 LR uses 10 GbE RoCE technology and is designed to drive distances
up to 10 km (6.21 miles) unrepeated and support a link data rate of 10 Gigabits per second
(Gbps). For distance requirements greater than 10 km (6.21 miles), clients must use a
Wavelength Division Multiplexer (WDM). The WDM vendor must be qualified by IBM Z.
Coupling Express2 LR is designed to support up to four CHPIDs per port, 32 buffers (that is,
32 subchannels) per CHPID. The Coupling Express2 LR feature is in the PCIe+ I/O drawer on
IBM z16.
For more information, see IBM Z Planning for Fiber Optic Links (FICON/FCP, Coupling Links,
Open Systems Adapters, and zHyperLink Express), GA23-1409. This publication is available
in the Library section of Resource Link (log-in required).
14 PCIe+ I/O drawer (FC 4023 on IBM z16, FC 4021 on IBM z15, and FC 4001 on IBM z14 ZR1) is installed in a
19-inch rack. PCIe+ I/O Drawers contains and can host up to 16 PCIe I/O features (adapters). They are not
supported on IBM z14 M0x or older servers. Also, the PCIe I/O drawer cannot be carried forward during and MES
upgrade to IBM z14 ZR1 or newer. IBM z16, IBM z15, and IBM z14 ZR1 support PCIe+ I/O drawers only.
Migration considerations
Upgrading from previous generations of IBM Z systems in a Parallel Sysplex to IBM z16
servers in that same Parallel Sysplex requires proper planning for coupling connectivity.
Planning is important because of the change in the supported type of coupling link adapters
and the number of available fan-out slots of the IBM z16 CPC drawers.
The ICA SR fan-out features provide short-distance connectivity to another IBM z16, IBM
z15, or IBM z14 server.
The CE LR adapter provides long-distance connectivity to another IBM z16, IBM z15, or
IBM z14 server.
The IBM z16 fan-out slots in the CPC drawer provide coupling link connectivity through the
ICA SR fan-out cards. In addition to coupling links for Parallel Sysplex, the fan-out cards
provide connectivity for the PCIe+ I/O drawer (PCIe+ Gen3 fan-out).
To migrate from an older generation machine to an IBM z16 without disruption in a Parallel
Sysplex environment requires that the older machines are no more than n-2 generation
(namely, at least IBM z14) and that they carry enough coupling links to connect to the existing
systems while also connecting to the new machine. This requirement is necessary to allow
individual components (z/OS LPARs and CFs) to be shut down and moved to the target
machine and continue to be connected to the remaining systems.
It is beyond the scope of this book to describe all possible migration scenarios. Always
consult with subject matter experts to help you to develop your migration strategy.
The use of the coupling links to exchange STP messages has the following advantages:
By using the same links to exchange STP messages and CF messages in a Parallel
Sysplex, STP can scale with distance. Servers that are exchanging messages over short
distances (ICA SR links), can meet more stringent synchronization requirements than
servers that exchange messages over long distance (CE2 LR links), with distances up to
100 kilometers (62 miles)15. This advantage is an enhancement over the IBM Sysplex
Timer implementation, which does not scale with distance.
Coupling links provide the connectivity that is necessary in a Parallel Sysplex. Therefore, a
potential benefit can be realized of minimizing the number of cross-site links that is
required in a multi-site Parallel Sysplex.
15 10 km (6.2 miles) without DWDM extenders; 100 km (62 miles) with certified DWDM equipment.
These coprocessors are Hardware Security Modules (HSMs) that provide high-security
cryptographic processing as required by banking and other industries. This feature provides a
secure programming and hardware environment wherein crypto processes are performed.
The Crypto Express8S (2 HSM) includes two IBM PCIe Cryptographic Co-processors
(PCIeCC); the Crypto Express8S (1 HSM) includes one PCIeCC per feature. For availability
reasons, a minimum of two features is required. Up to 30 Crypto Express8S (2 HSM) features
are supported. The maximum number of the 1 HSM features is 16. The Crypto Express8S
feature occupies one I/O slot in a PCIe+ I/O drawer.
Each adapter can be configured as a Secure IBM CCA coprocessor, a Secure IBM Enterprise
PKCS #11 (EP11) coprocessor, or as an accelerator.
TKE feature: The Trusted Key Entry (TKE) Workstation feature is required for
supporting the administration of the Crypto Express6S when configured as an
Enterprise PKCS #11 coprocessor or managing the CCA mode PCI-HSM.
When the Crypto Express8S PCI Express adapter is configured as a secure IBM CCA
co-processor, it still provides accelerator functions. However, up to 3x better performance for
those functions can be achieved if the Crypto Express8S PCI Express adapter is configured
as an accelerator.
CCA enhancements include the ability to use triple-length (192-bit) Triple-DES (TDES) keys
for operations, such as data encryption, IBM PIN processing, and key wrapping to strengthen
security. CCA also extended the support for the cryptographic requirements of the German
Banking Industry Committee Deutsche Kreditwirtschaft (DK).
Several features that support the use of the AES algorithm in banking applications also were
added to CCA. These features include the addition of AES-related key management features
and the AES ISO Format 4 (ISO-4) PIN blocks as defined in the ISO 9564-1 standard. PIN
block translation is supported and use of AES PIN blocks in other CCA callable services.
IBM continues to add enhancements as AES finance industry standards are released.
4.7.3 Crypto Express7S feature (FC 0898 and FC 0899) as carry forward only
The Crypto Express7S are supported on IBM z16. These coprocessors are HSMs that
provide high-security cryptographic processing as required by banking and other industries.
This feature provides a secure programming and hardware environment wherein crypto
processes are performed.
The Crypto Express7S (1 port), FC 0899 includes one IBM PCIe Cryptographic
Coprocessors (PCIeCC) per feature. For availability reasons, a minimum of two features is
required for the one port feature.
Up to 30 Crypto Express7S (2 port) features are supported on IBM z15 T01. The maximum
number of the one-port features is 16. The total number of HSMs supported on IBM z16 A01
is 60 in a combination of Crypto Express8S (2 HSM), Crypto Express8S (1 HSM), Crypto
Express7S (2 port), Crypto Express7S (1 port), or Crypto Express6S.
The Crypto Express7S feature occupies one I/O slot in a PCIe+ I/O drawer.
Each adapter can be configured as a Secure IBM CCA coprocessor, Secure IBM Enterprise
PKCS #11 (EP11) coprocessor, or accelerator.
The accelerator function is designed for maximum-speed Secure Sockets Layer and
Transport Layer Security (SSL/TLS) acceleration, rather than for specialized financial
applications for secure, long-term storage of keys or secrets.
The Crypto Express7S can also be configured as one of the following configurations:
The Secure IBM CCA coprocessor includes secure key functions with an emphasis on the
specialized functions that are required for banking and payment card systems. It is
optionally programmable to add custom functions and algorithms by using UDX.
A new mode (called PCI-HSM) is available exclusively for Crypto Express6S in CCA
mode. PCI-HSM mode simplifies compliance with PCI requirements for hardware security
modules.
The Secure IBM Enterprise PKCS #11 (EP11) coprocessor implements an
industry-standardized set of services that adheres to the PKCS #11 specification v2.20
and more recent amendments. It was designed for extended FIPS and Common Criteria
evaluations to meet industry requirements.
This cryptographic coprocessor mode introduced the PKCS #11 secure key function.
TKE feature: The TKE Workstation feature is required for supporting the administration
of the Crypto Express8S when configured as an Enterprise PKCS #11 coprocessor or
managing the CCA mode PCI-HSM.
When the Crypto Express7S PCI Express adapter is configured as a secure IBM CCA
co-processor, it still provides accelerator functions. However, up to 3x better performance for
those functions can be achieved if the Crypto Express7S PCI Express adapter is configured
as an accelerator.
CCA enhancements include the ability to use triple-length (192-bit) Triple-DES (TDES) keys
for operations, such as data encryption, PIN processing, and key wrapping to strengthen
security. CCA also extended the support for the cryptographic requirements of the German
Banking Industry Committee Deutsche Kreditwirtschaft (DK).
IBM continues to add enhancements as AES finance industry standards are released.
Each Crypto Express6S feature holds one PCI Express cryptographic adapter. Each adapter
can be configured by the installation as a Secure IBM Common Cryptographic Architecture
(CCA) coprocessor, as a Secure IBM Enterprise Public Key Cryptography Standards (PKCS)
#11 (EP11) coprocessor, or as an accelerator.
The tamper-resistant hardware security module, which is contained on the Crypto Express6S
feature, conforms to the Federal Information Processing Standard (FIPS) 140-2 Level 4
Certification. It supports UDX services to implement cryptographic functions and algorithms
(when defined as an IBM CCA coprocessor).
The following EP11 compliance levels are available (Crypto Express6S and Crypto
Express5S):
FIPS 2009 (default)
FIPS 2011
BSI 2009
BSI 2011
Each Crypto Express6S feature occupies one I/O slot in the PCIe I/O drawer, and features no
CHPID assigned. However, it includes one PCHID.
All native PCIe features should be ordered in pairs for redundancy. The features are assigned
to one of the four resource groups (RGs) that are running on the IFP according to their
physical location in the PCIe+ I/O drawer, which provides management functions and
virtualization functions.
If two features of the same type are installed, one always is managed by resource group 1
(RG 1) or resource group 3 (RG3); the other feature is managed by resource group 2 (RG 2)
or resource group 4 (RG 4). This configuration provides redundancy if one of the features or
resource groups needs maintenance or fails.
The IFP and RGs support the following infrastructure management functions:
Firmware update of adapters and resource groups
Error recovery and failure data collection
Diagnostic and maintenance tasks
The channel subsystem directs the flow of information between I/O devices and main storage.
It allows data processing to proceeded concurrently with I/O processing, which relieves data
processors (central processor (CP) and Integrated Facility for Linux [IFL]) of the task of
communicating directly with I/O devices.
The channel subsystem includes subchannels, I/O devices that are attached through control
units, and channel paths between the subsystem and control units. For more information
about the channel subsystem, see 5.1.1, “Multiple logical channel subsystems”.
The design of IBM Z servers offers considerable processing power, memory size, and I/O
connectivity. In support of the larger I/O capability, the CSS structure is scaled up by
introducing the multiple logical channel subsystem (LCSS) since IBM z990, and multiple
subchannel sets (MSS) since IBM z9.
An overview of the channel subsystem for IBM z16 servers is shown in Figure 5-1. IBM z16
A01 systems are designed to support up to six logical channel subsystems, each with four
subchannel sets and up to 256 channels.
All channel subsystems are defined within a single configuration, which is called I/O
configuration data set (IOCDS). The IOCDS is loaded into the hardware system area (HSA)
during a central processor complex (CPC) power-on reset (POR) to start all of the channel
subsystems.
On IBM z16 A01 systems, the HSA is pre-allocated in memory with a fixed size of 256 GB,
which is in addition to the customer purchased memory. This fixed size memory for HSA
eliminates the requirement for more planning of the initial I/O configuration and planning for
future I/O expansions.
CPC drawer repair: The HSA can be moved from one CPC drawer to a different drawer in
an enhanced availability configuration as part of a concurrent CPC drawer repair (CDR)
action.
The following objects are always reserved in the IBM z16 A01 HSA during POR, whether they
are defined in the IOCDS for use:
The introduction of multiple LCSSs enabled an IBM Z server to have more than one channel
subsystems logically, while each logical channel subsystem maintains the same manner of
I/O processing. Also, a logical partition (LPAR) is now attached to a specific logical channel
subsystem, which makes the extension of multiple logical channel subsystems not apparent
to the operating systems and applications. The multiple image facility (MIF) in the structure
enables resource sharing across LPARs within a single LCSS or across the LCSSs.
The multiple LCSS structure extended the IBM Z servers’ total number of I/O connectivity to
support a balanced configuration for the growth of processor and I/O capabilities.
Note: The phrase channel subsystem has same meaning as logical channel subsystem in
this section, unless otherwise stated.
Subchannels
A subchannel provides the logical appearance of a device to the program and contains the
information that is required for sustaining a single I/O operation. Each device is accessible by
using one subchannel in a channel subsystem to which it is assigned according to the active
IOCDS of the IBM Z server.
In z/Architecture, the first subchannel set of an LCSS can have 63.75 K subchannels (with
0.25 K reserved), with a subchannel set ID (SSID) of 0. By enabling the multiple subchannel
sets, extra subchannel sets are available to increase the device addressability of a channel
subsystem.
For more information about multiple subchannel sets, see 5.1.2, “Multiple subchannel sets”
on page 200.
Each channel path in a channel subsystem features a unique 2-digit hexadecimal identifier
that is known as a channel-path identifier (CHPID), which ranges 00 - FF. Therefore, a total of
256 CHPIDs are supported by a CSS, and a maximum of 1536 CHPIDs are available on an
IBM z16 with six logical channel subsystems.
A port on an I/O feature card features a unique physical channel identifier (PCHID) according
to the physical location of this I/O feature adapter, and the sequence of this port on the
adapter.
In addition, a port on a fan-out adapter has a unique adapter identifier (AID), according to the
physical location of this fan-out adapter, and the sequence of this port on the adapter.
A CHPID is assigned to a physical port by defining the corresponding PCHID or AID in the I/O
configuration definitions.
Control units
A control unit provides the logical capabilities that are necessary to operate and control an
I/O device. It adapts the characteristics of each device so that it can respond to the standard
form of control that is provided by the CSS.
A control unit can be housed separately or can be physically and logically integrated with the
I/O device, channel subsystem, or within the IBM Z server.
I/O devices
An I/O device provides external storage, a means of communication between
data-processing systems, or a means of communication between a system and its
environment. In the simplest case, an I/O device is attached to one control unit and is
accessible through one or more channel paths that are connected to the control unit.
Each subchannel has a unique four-digit hexadecimal number 0x0000 - 0xFFFF. Therefore, a
single subchannel set can address and access up to 64 K I/O devices.
The IBM z16 A01 systems support four subchannel sets for each logical channel subsystem.
The IBM z16 A01 system can access a maximum of 255.74 K devices for a logical channel
subsystem and a logical partition and the programs that are running on it.
Subchannel number
The subchannel number is a four-digit hexadecimal number 0x0000 - 0xFFFF, which is
assigned to a subchannel within a subchannel set of a channel subsystem. Subchannels in
each subchannel set are always assigned subchannel numbers within a single range of
contiguous numbers.
With the subchannel numbers, a program that is running on an LPAR (for example, an
operating system) can specify all I/O functions relative to a specific I/O device by designating
a subchannel that is assigned to the I/O devices.
Normally, subchannel numbers are used only in communication between the programs and
the channel subsystem.
Device number
A device number is an arbitrary number 0x0000 - 0xFFFF, which is defined by a system
programmer in an I/O configuration for naming an I/O device. The device number must be
unique within a subchannel set of a channel subsystem. It is assigned to the corresponding
subchannel by channel subsystem when an I/O configuration is activated. Therefore, a
subchannel in a subchannel set of a channel subsystem includes a device number together
with a subchannel number for designating an I/O operation.
The device number provides a means to identify a device that is independent of any
limitations that are imposed by the system model, configuration, or channel-path protocols.
A device number also can be used to designate an I/O function to a specific I/O device.
Because it is an arbitrary number, it can easily be fit into any configuration management and
operating management scenarios. For example, a system administrator can set all of the
z/OS systems in an environment to device number 1000 for their system RES volumes.
With multiple subchannel sets, a subchannel is assigned to a specific I/O device by the
channel subsystem with an automatically assigned subchannel number and a device number
that is defined by user. An I/O device can always be identified by an SSID with a subchannel
number or a device number. For example, a device with device number AB00 of subchannel
set 1 can be designated as 1AB00.
For the extra subchannel sets enabled by the MSS facility, each has 65535 subchannels
(64 K minus one) for specific types of devices. These extra subchannel sets are referred as
alternative subchannel sets in z/OS.
Also, a device that is defined in an alternative subchannel set is considered a special device,
which normally features a special device type in the I/O configuration.
Currently, an IBM z16 A01 system that is running z/OS defines the following types of devices
in another subchannel set, with proper APAR or PTF installed:
Alias devices of the parallel access volumes (PAV).
Secondary devices of GDPS Metro Mirror Copy Service (formerly Peer-to-Peer Remote
Copy [PPRC]).
IBM FlashCopy® SOURCE and TARGET devices with program temporary fix (PTF)
OA46900.
Db2 data backup volumes with PTF OA24142.
The use of another subchannel set for these special devices helps reduce the number of
devices in the subchannel set 0, which increases the growth capability for accessing more
devices.
This configuration allows the users of Metro Mirror (formerly PPRC) secondary devices that
are defined by using the same device number and a new device type in an alternative
subchannel set to be used for IPL, an I/O definition file (IODF), and stand-alone memory
dump volumes, when needed.
D IOS,CONFIG(ALL)
IOS506I 11.14.46 I/O CONFIG DATA 606
ACTIVE IODF DATA SET = SYS9.IODF81
CONFIGURATION ID = ITSO EDT ID = 01
TOKEN: PROCESSOR DATE TIME DESCRIPTION
SOURCE: PAVO 22-03-04 09:12:20 SYS9 IODF81
ACTIVE CSS: 3 SUBCHANNEL SETS CONFIGURED: 0, 1, 2, 3
CHANNEL MEASUREMENT BLOCK FACILITY IS ACTIVE
SUBCHANNEL SET FOR PPRC PRIMARY: INITIAL = 0 ACTIVE = 0
HYPERSWAP FAILOVER HAS OCCURRED: NO
LOCAL SYSTEM NAME (LSYSTEM): PAVO
HARDWARE SYSTEM AREA AVAILABLE FOR CONFIGURATION CHANGES
PHYSICAL CONTROL UNITS 8061
CSS 0 - LOGICAL CONTROL UNITS 4027
SS 0 SUBCHANNELS 59510
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 1 - LOGICAL CONTROL UNITS 4005
SS 0 SUBCHANNELS 59218
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 2 - LOGICAL CONTROL UNITS 4025
SS 0 SUBCHANNELS 59410
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 3 - LOGICAL CONTROL UNITS 4026
SS 0 SUBCHANNELS 60906
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 4 - LOGICAL CONTROL UNITS 4043
SS 0 SUBCHANNELS 61266
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
CSS 5 - LOGICAL CONTROL UNITS 4088
SS 0 SUBCHANNELS 65280
SS 1 SUBCHANNELS 65535
SS 2 SUBCHANNELS 65535
SS 3 SUBCHANNELS 65535
ELIGIBLE DEVICE TABLE LATCH COUNTS
0 OUTSTANDING BINDS ON PRIMARY EDT
Although a shared channel path can be shared by LPARs within a same LCSS, a spanned
channel path can be shared by LPARs within and across LCSSs.
By assigning the same CHPID from different LCSSs to the same channel path (for example, a
PCHID), the channel path can be accessed by any LPARs from these LCSSs at the same
time. The CHPID is spanned across those LCSSs. The use of spanned channels paths
decreases the number of channels that are needed in an installation of IBM Z servers.
A sample of channel paths that are defined as dedicated, shared, and spanned is shown in
Figure 5-3.
Channel spanning is supported for internal links (HiperSockets and IC links) and for specific
types of external links. External links that are supported on IBM z16 A01 systems include
FICON Express32S, FICON Express16SA, FICON Express16S+, OSA-Express7S 1.2,
OSA-Express7S, OSA-Express6S, and Coupling Links.
LPAR name
The LPAR name is defined as partition name parameter in the RESOURCE statement of an
I/O configuration. The LPAR name must be unique across the server.
MIF image ID
The MIF image ID is defined as a parameter for each LPAR in the RESOURCE statement of
an I/O configuration. It ranges 1 - F, and must be unique within an LCSS. However, duplicates
are allowed in different LCSSs.
If a MIF image ID is not defined, an arbitrary ID is assigned when the I/O configuration
activated. The IBM z16 server supports a maximum of six LCSSs, with a total of 85 LPARs
that can be defined.
Each LCSS of an IBM z16 A01 system can support the following numbers of LPARs:
LCSS0 to LCSS4 support 15 LPARs each, and the MIF image ID is 1 - F.
LCSS5 supports 10 LPARs, and the MIF image IDs are 1 - A.
LPAR ID
The LPAR ID is defined by a user in an image activation profile for each LPAR. It is a 2-digit
hexadecimal number 00 - 7F. The LPAR ID must be unique across the server.
Although it is arbitrarily defined by the user, an LPAR ID often is the CSS ID concatenated to
its MIF image ID, which makes the value more meaningful for the system administrator. For
example, an LPAR with LPAR ID 1A defined in that manner means that the LPAR is defined in
LCSS1, with the MIF image ID A.
Note: Specific functions might require specific levels of an operating system, PTFs, or
both.
This chapter also introduces the principles of cryptography and describes the implementation
of cryptography in the hardware and software architecture of IBM Z. It also describes the
features that IBM z16 offers.
The IBM z16 uses quantum-safe technologies to help protect your business-critical
infrastructure and data from potential attacks.
The IBM z16 delivers a transparent and consumable approach that enables extensive
(pervasive) encryption of data in flight and at rest, with the goal of substantially simplifying
data security and reducing the costs that are associated with protecting data while achieving
compliance mandates.
Naming: The IBM z16, Machine Type 3931 (M/T 3931), Model A01 is further identified as
IBM z16, unless otherwise specified.
IBM z16 introduced the new PCI Crypto Express8S feature (that can be managed by a new
Trusted Key Entry (TKE) workstation) together with a further improved CPACF Coprocessor.
In addition, the IBM Common Cryptographic Architecture (CCA) and the IBM Enterprise
PKCS #11 (EP11) Licensed Internal Code (LIC) were enhanced.
The new features support new standards and meet the following compliance requirements:
Payment Card Industry (PCI) Hardware Security Module (HSM) certification to strength
the cryptographic standards for attack resistance in the payment card systems area.
PCI HSM certification is available for Crypto Express8S, Crypto Express7S, and Crypto
Express6S.
National Institute of Standards and Technology (NIST) through the Federal Information
Processing Standard (FIPS) standard to implement guidance requirements.
Common Criteria EP11 EAL4.
German Banking Industry Commission (GBIC).
Visa Format Preserving Encryption (VFPE) for credit card numbers.
Enhanced public key Elliptic Curve Cryptography (ECC) for users such as Chrome,
Firefox, and Apple’s iMessage.
Accredited Standards Committee X9 Inc Technical Report-34 (ASC X9 TR-34)
IBM z16 includes standard cryptographic hardware and optional cryptographic features for
flexibility and growth capability. IBM has a long history of providing hardware cryptographic
solutions. This history stretches from the development of the Data Encryption Standard (DES)
in the 1970s to the Crypto Express tamper-sensing and tamper-responding programmable
features.
Crypto Express meets the US Government’s highest security rating of FIPS 140-3 Level 4
certification1. It also meets several other security ratings, such as the Common Criteria for
Information Technology Security Evaluation, the PCI HSM criteria, and the criteria for German
Banking Industry Commission (formerly known as Deutsche Kreditwirtschaft evaluation).
1 FIPS 140-3 certification - Security Requirements for Cryptographic Modules.
Also, it is necessary to ensure that a message cannot be corrupted (message integrity), while
ensuring that the sender and the receiver really are the persons who they claim to be. Over
time, several methods were used to achieve these objectives, with more or less success.
Many procedures and algorithms for encrypting and decrypting data were developed that are
increasingly complicated and time-consuming.
These goals should all be possible without unacceptable overhead to the communication. The
goal is to keep the system secure, manageable, and productive.
The basic data protection method is achieved by encrypting and decrypting the data, while
hash algorithms, message authentication codes (MACs), digital signatures, and certificates
are used for authentication, data integrity, and nonrepudiation.
In other words, the security of a cryptographic system should depend on the security of the
key, so the key must be kept secret. Therefore, the secure management of keys is the primal
task of modern cryptographic systems.
6.2.3 Keys
The keys that are used for the cryptographic algorithms often are sequences of numbers and
characters, but can also be any other sequence of bits. The length of a key influences the
security (strength) of the cryptographic method. The longer the used key, the more difficult it
is to compromise a cryptographic algorithm.
For example, the DES (symmetric key) algorithm uses keys with a length of 56 bits,
Triple-DES (TDES) uses keys with a length of 112 bits, and Advanced Encryption Standard
(AES) uses keys of 128, 192, or 256 bits. The asymmetric key RSA algorithm (named after its
inventors Rivest, Shamir, and Adleman) uses keys with a length of 1024, 2048, or 4096 bits.
secure
key
protected
key
clear
key
speed
Figure 6-1 Three levels of protection with three levels of speed
The most prominent one-way algorithms are the Secure Hash Algorithms (SHA).
The cryptographic hardware that is supported on IBM z16 is shown in Figure 6-2. These
features are described in this chapter.
Implemented in every processor unit (PU) or core in a central processor complex (CPC) is a
cryptographic coprocessor that can be used2 for cryptographic algorithms that uses clear
keys or protected keys. For more information, see 6.4, “CP Assist for Cryptographic
Functions” on page 219.
The Crypto Express8S adapter is an HSM that is placed in the PCIe+ I/O drawer of IBM z16.
It also supports cryptographic algorithms by using secret keys. For more information, see 6.5,
“Crypto Express8S” on page 224.
Finally, a TKE workstation is required for entering keys in a secure way into the Crypto
Express8S HSM, which often also is equipped with smart card readers. For more information,
see 6.6, “Trusted Key Entry workstation” on page 239.
This feature is a prerequisite to use CPACF (except for SHA-1, SHA-224, SHA-256,
SHA-384, and SHA-512) and the PCIe Crypto Express features.
These features are optional. The 2-port feature contains two IBM 4770 PCIe
Cryptographic Coprocessors (HSMs), which can be independently defined as
Coprocessor or Accelerator. New feature. Not supported on previous generations
IBM Z systems.
A TKE, smart card Reader and latest available level smart cards are required to operate
the Crypto adapter card in EP11 mode.
These features are optional. The 2-port feature contains two IBM 4770 PCIe
Cryptographic Coprocessors (HSMs), which can be independently defined as
Coprocessor or Accelerator. New feature. Not supported on previous generations
IBM Z systems.
A TKE, Smart Card Reader and latest available level smart cards are required to
operate the Crypto adapter card in EP11 mode.
Carry forward from IBM z15. This feature contains two IBM 4769 PCIe Cryptographic
Coprocessors (HSMs), which can be independently defined as Coprocessor or
Accelerator. Supported on IBM z15 and IBM z16.
Carry forward from IBM z15. This feature contains one IBM 4769 PCIe Cryptographic
Coprocessor (HSM), which can be defined as Coprocessor or Accelerator. Supported
on IBM z15 and IBM z16.
This feature is available as a carry forward MES from IBM z14. This feature is optional.
Each feature one IBM 4768 PCIe Cryptographic Coprocessor (HSM). This feature is
supported in IBM z16, IBM z15, and IBM z14.
Included with the TKE tower workstation FC 0058 and the TKE rack-mounted
workstation FC 0057 for IBM z16. Earlier versions of TKE features (feature codes: 0087,
0088, 0085, and 0086) also can be upgraded to TKE 10.0 LIC, adding FC 0851
(IBM 4770 PCIeCC) if the TKE is assigned to an IBM z16 and manages Crypto
Express8S.
The stand-alone crypto adapter is required for TKE upgrade from FC 0085 and FC 0086
TKE tower, or FC 0087 and FC 0088 TKE Rack Mount when carry forward these
features to IBM z16.
TKE Tower FC 0088 can be carried forward to IBM z16. It requires IBM 4770 PCIeCC
(FC 0851) for compatibility with TKE LIC 10.0 (FC 0882) and for managing Crypto
Express8S (FC 0144 = FC 0088 + FC 0851 + FC 0882).
TKE Rack Mount FC 0087 can be carried forward to IBM z16. It requires IBM 4770
PCIeCC (FC 0851) for compatibility with TKE LIC 10.0 (FC 0882) and for managing
Crypto Express8S (FC 0145 = FC 0087 + FC 0851 + FC 0882).
TKE Rack Mount FC 0085 can be carried forward to IBM z16. It requires IBM 4770
PCIeCC (FC 0851) for compatibility with TKE LIC 10.0 (FC 0882) and for managing
Crypto Express8S (FC 0233 = FC 0085 + FC 0851 + FC 0882).
TKE Tower FC 0086 can be carried forward to IBM z16. It requires IBM 4770 PCIeCC
(FC 0851) for compatibility with TKE LIC 10.0 (FC 0882) and for managing Crypto
Express8S (FC 0234 = FC 0086 + FC 0851 + FC 0882).
Access to information in the smart card is protected by a PIN. One feature code
includes two smart card readers, two cables to connect to the TKE workstation, and 20
smart cards. Existing smart cards can be used on TKE 9.2 with these readers.
Access to information in the smart card is protected by a PIN. Carry forward with
existing cards (non-FIPS). Existing smart cards can be used on TKE 9.2 with these
readers.
This card allows the TKE to support zones with EC 521 key strength (EC 521 strength
for Logon Keys, Authority Signature Keys, and EP11 signature keys).
Carry forward only to IBM z16. Ten smart cards are included.
a. The maximum number of combined features of all types cannot exceed 60 HSMs on an IBM
z16 A01 (any combination of single and dual HSM Crypto Express features). Therefore, the
maximum number for Feature Code 0908 is 30; for all other (single HSM) types, it is 16 for an
IBM z16 system.
A TKE includes support for AES encryption algorithm with 256-bit master keys and key
management functions to load or generate master keys to the cryptographic coprocessor.
If the TKE workstation is chosen to operate the Crypto Express8S adapter in an IBM z16,
TKE workstation with the TKE 10.0 LIC is required. For more information, see 6.6, “Trusted
Key Entry workstation” on page 239.
Important: Products that include any of the cryptographic feature codes contain
cryptographic functions that are subject to special export licensing requirements by the US
Department of Commerce. It is your responsibility to understand and adhere to these
regulations when you are moving, selling, or transferring these products.
To access and use the cryptographic hardware devices that are provided by IBM z16, the
application must use an application programming interface (API) that is provided by the
operating system. In z/OS, the Integrated Cryptographic Service Facility (ICSF) provides the
APIs and is managing the access to the cryptographic devices, as shown in Figure 6-3.
ICSF transparently routes application requests for cryptographic services to one of the
integrated cryptographic engines (CPACF or a Crypto Express feature), depending on
performance or requested cryptographic function. ICSF is also the means by which the
secure Crypto Express features are loaded with master key values, which allows the
hardware features to be used by applications.
The cryptographic hardware that is installed in IBM z16 determines the cryptographic
features and services that are available to the applications.
The users of the cryptographic services call the ICSF API. Some functions are performed by
the ICSF software without starting the cryptographic hardware features. Other functions result
in ICSF going into routines that contain proprietary IBM Z crypto instructions. These
instructions are run by a CPU engine and result in a work request that is generated for a
cryptographic hardware feature.
This cryptographic coprocessor, which is known as the CPACF, is not qualified as an HSM;
therefore, it is not suitable for handling algorithms that use secret keys. However, the
coprocessor can be used for cryptographic algorithms that use clear keys or protected keys.
The CPACF works synchronously with the PU, which means that the owning processor is
busy when its coprocessor is busy. This setup provides a fast device for cryptographic
services.
The CPACF offers a set of symmetric cryptographic functions that enhances the encryption
and decryption performance of clear key operations. These functions are for SSL, virtual
private network (VPN), and data-storing applications that do not require FIPS 140-2 Level 4
security.
CPACF is designed to facilitate the privacy of cryptographic key material when used for data
encryption through key wrapping implementation. It ensures that key material is not visible to
applications or operating systems during encryption operations. For more information, see
6.4.2, “CPACF protected key” on page 222.
The CPACF feature provides hardware acceleration for the following cryptographic services:
Symmetric ciphers:
– DES
– Triple-DES
– AES-128
– AES-192
– AES-256 (all for clear and protected keys)
Elliptic curves cryptography (ECC):
– ECDSA, ECDH, support for the NIST P256, NIST P386, NIST P521
– EdDSA for Ed25519, Ed448 Curves
– ECDH for X25519, X448 Curves
– Key generation for NIST, Ed, and X curves
Hashes/MACs:
– SHA-1
– SHA-224 (SHA-2 or SHA-3 standard)
– SHA-256 (SHA-2 or SHA-3 standard)
– SHA-384 (SHA-2 or SHA-3 standard)
– SHA-512 (SHA-2 or SHA-3 standard)
– SHAKE-128
– SHAKE-256
– GHASH
Random number generator:
– PRNG (3DES based)
– DRNG (NIST SP-800-90A SHA-512 based)
– TRNG (true random number generator
These functions are provided as problem-state z/Architecture instructions that are directly
available to application programs. These instructions are known as Message-Security Assist
(MSA). When enabled, the CPACF runs at processor speed for every CP, IFL, and zIIP.
For more information about MSA instructions, see z/Architecture Principles of Operation,
SA22-7832.
For activating these functions, the CPACF must be enabled by using Feature Code (FC) 3863,
which is available at no charge. Support for hashing algorithms SHA-1, SHA-256, SHA-384,
and SHA-512 always is enabled.
The following tools might benefit from the throughput improvements for IBM z16 CPACF:
Db2/IMS encryption tool
Db2 built-in encryption
z/OS Communication Server: IPsec/IKE/AT-TLS
z/OS System SSL
z/OS Network Authentication Service (Kerberos)
DFDSS Volume encryption
z/OS Java SDK
z/OS Encryption Facility
Linux on IBM Z: Kernel, openSSL, openCryptoki, and GSKIT
The IBM z16 hardware includes the implementation of algorithms as hardware synchronous
operations. This configuration holds the PU processing of the instruction flow until the
operation completes.
For the SHA hashing algorithms and the random number generation algorithms, only clear
keys are used. For the symmetric encryption and decryption DES and AES algorithms and
clear keys, protected keys also can be used. On IBM z16, protected keys require a Crypto
Express adapter that is running in CCA mode.
For more information, see 6.5.2, “Crypto Express8S as a CCA coprocessor” on page 228.
The hashing algorithms SHA-1, SHA-2, and SHA-3 support for SHA-224, SHA-256,
SHA-384, and SHA-512, are enabled on all systems and do not require the CPACF
enablement feature. For all other algorithms, the no-charge CPACF enablement feature
(FC 3863) is required.
The CPACF functions are implemented as processor instructions and require operating
system support for use. Operating systems that use the CPACF instructions include z/OS,
z/VM, z/VSE, z/TPF, and Linux on IBM Z.
Clear keys process faster than secure keys because the process is done synchronously on
the CPACF. Protected keys blend the security of Crypto Express8S, Crypto Express7S, or
Crypto Express6S coprocessors and the performance characteristics of the CPACF. This
process allows it to run closer to the speed of clear keys.
CPACF facilitates the continued privacy of cryptographic key material when used for data
encryption. In Crypto Express8S, Crypto Express7S, or Express6S coprocessors, a secure
key is encrypted under a master key. However, a protected key is encrypted under a wrapping
key that is unique to each LPAR.
Because the wrapping key is unique to each LPAR, a protected key cannot be shared with
another LPAR. By using key wrapping, CPACF ensures that key material is not visible to
applications or operating systems during encryption operations.
CPACF code generates the wrapping key and stores it in the protected area of the hardware
system area (HSA). The wrapping key is accessible only by firmware. It cannot be accessed
by operating systems or applications.
DES/T-DES and AES algorithms are implemented in CPACF code with the support of
hardware assist functions. Two variations of wrapping keys are generated: one for
DES/T-DES keys and another for AES keys.
3 PCIeCC: IBM PCIe Cryptographic Coprocessor is the Hardware Security Module (HSM).
Figure 6-5 CPACF key wrapping for Crypto Express8S, Crypto Express7S, and Crypto Express6S
The key wrapping for Crypto Express8S is similar to Crypto Express7S or Crypto Express6S.
The CPACF Wrapping Key and the Transport Key for use with Crypto Express7S or Crypto
Express6S are in a protected area of the HSA that is not visible to operating systems or
applications.
A new segment in the profiles of the CSFKEYS class in IBM RACF® restricts which secure
keys can be used as protected keys. By default, all secure keys are considered not eligible to
be used as protected keys. The process that is shown in Figure 6-5 considers a secure key as
the source of a protected key.
The source key in this case is stored in the ICSF Cryptographic Key Data Set (CKDS) as a
secure key, which was encrypted under the master key. This secure key is sent to CEX8C,
CEX7C, or CEX6C to be deciphered and then, sent to the CPACF in clear text.
At the CPACF, the key is wrapped under the LPAR wrapping key, and is then returned to ICSF.
After the key is wrapped, ICSF can keep the protected value in memory. It then passes it to
the CPACF, where the key is unwrapped for each encryption or decryption operation.
Each Crypto Express8S PCI Express adapter (HSM) is available in one of the following
configurations:
Secure IBM CCA coprocessor (CEX8C) for FIPS 140-2 Level 4 certification. This
configuration includes secure key functions. It is optionally programmable to deploy more
functions and algorithms by using UDX.
For more information, see 6.5.2, “Crypto Express8S as a CCA coprocessor” on page 228.
A TKE workstation is required to support the administration of the Crypto Express8S when
it is configured in CCA mode when in full Payment Card Industry (PCI)-compliant mode for
the necessary certificate management in this mode. The TKE is optional in all other use
cases for CCA.
Secure IBM Enterprise PKCS #11 (EP11) coprocessor (CEX8P) implements an
industry-standardized set of services that adheres to the PKCS #11 specification V2.20
and more recent amendments. It was designed for extended FIPS and Common Criteria
evaluations to meet public sector requirements. This new cryptographic coprocessor
mode introduced the PKCS #11 secure key function.
For more information, see 6.5.3, “Crypto Express8S as an EP11 coprocessor” on
page 234.
A TKE workstation is always required to support the administration of the Crypto
Express7S when it is configured in EP11 mode.
Accelerator (CEX8A) for acceleration of public key and private key cryptographic
operations that are used with SSL/TLS processing.
For more information, see 6.5.4, “Crypto Express8S as an accelerator” on page 235.
These modes can be configured by using the SE. The PCIe adapter must be configured
offline to change the mode.
Attention: Switching between configuration modes erases all adapter secrets. The
exception is when you are switching from Secure CCA to accelerator, and vice versa.
The Crypto Express8S feature does not include external ports and does not use optical fiber
or other cables. It does not use channel path identifiers (CHPIDs), but requires one slot in the
PCIe I/O drawer and one physical channel ID (PCHID) for each PCIe cryptographic adapter.
Removal of the feature or adapter zeroizes its content. Access to the PCIe cryptographic
adapter is controlled through the setup in the image profiles on the SE.
Adapter: Although PCIe cryptographic adapters include no CHPID type and are not
identified as external channels, all logical partitions (LPARs) in all channel subsystems can
access the adapter. In IBM z16, up to 85 LPARs are supported per adapter (HSM).
Accessing the adapter requires a setup in the image profile for each partition. The adapter
must be in the candidate list.
Each IBM z16 A01 supports up to 60 HSMs in total (combination of Crypto Express8S (1 or 2
HSMs), Crypto Express7S (1 or 2 ports, and Crypto Express6S). Crypto Express7S (1 or 2
ports) and Crypto Express6S features (single HSM) are not orderable for a new build IBM z16
A01 system, but can be carried forward from an IBM z14 or IBM z15 by using an MES.
Configuration information for Crypto Express8S is listed in Table 6-2.
The concept of dedicated processor does not apply to the PCIe cryptographic adapter.
Whether configured as a coprocessor or an accelerator, the PCIe cryptographic adapter is
made available to an LPAR. It is made available as directed by the domain assignment and
the candidate list in the LPAR image profile. This availability is not changed by the shared or
dedicated status that is given to the PUs in the partition.
The definition of domain indexes and PCIe cryptographic adapter numbers in the candidate
list for each LPAR must be planned to allow for nondisruptive changes. Consider the following
points:
Operational changes can be made by using the Change LPAR Cryptographic Controls
task from the SE, which reflects the cryptographic definitions in the image profile for the
partition. With this function, adding and removing the cryptographic feature without
stopping a running operating system can be done dynamically.
The same usage domain index can be defined more than once across multiple LPARs.
However, the PCIe cryptographic adapter number that is coupled with the usage domain
index that is specified must be unique across all active LPARs.
The same PCIe cryptographic adapter number and usage domain index combination can be
defined for more than one LPAR (up to 85 for IBM z16). For example, you might define a
configuration for backup situations. However, only one of the LPARs can be active at a time.
For more information, see 6.5.5, “Managing Crypto Express8S” on page 236.
4
SHA-3 was standardized by NIST in 2015. SHA-2 is still acceptable and no indication exists that SHA-2 is
vulnerable or that SHA-3 is more or less vulnerable than SHA-2.
Several of these algorithms require a secure key and must run on an HSM. Some of these
algorithms can also run with a clear key on the CPACF. Many standards are supported only
when Crypto Express8S is running in CCA mode. Others are supported only when the
adapter is running in EP11 mode.
The three modes for Crypto Express8S are described next. For more information, see 6.7,
“Cryptographic functions comparison” on page 242.
UDX is supported under a special contract through an IBM or approved third-party service
offering. The Crypto Cards website directs your request to an IBM representative for your
geographic location. A special contract is negotiated between IBM and you for the
development of the UDX code by IBM according to your specifications and an agreed-upon
level of the UDX.
A UDX toolkit for IBM Z is tied to specific versions of the CCA code and the related host code.
UDX is available for the Crypto Express8S, Crypto Express7S, and Crypto Express6S
(Secure IBM CCA coprocessor mode only) features. An UDX migration is no more disruptive
than a normal Microcode Change Level (MCL) or ICSF release migration.
Consideration: CCA features a new code level starting with z13 systems, and the UDX
clients require a new UDX.
On IBM z16 A01, Crypto Express8S is delivered with CCA Level 8.0 firmware. A new set of
cryptographic functions and callable services is provided by the IBM CCA LIC to enhance the
functions that secure financial transactions and keys. The Crypto Express8S includes the
following features:
Greater than 16 domains support up to 85 LPARs on IBM z16 A01.
PCI PIN Transaction Security (PTS) HSM Certification that is available to IBM z16 in
combination with CEX8S, CEX7S, or CEX6S features, to z15 in combination with CEX7S
or CEX6S features, and to IBM z14 with CEX6S features.
VFPE support, which was introduced with z13/z13s systems.
AES PIN support for the German banking industry.
PKA Translate UDX function into CCA.
Verb Algorithm Currency.
CCA improvements
The following CCA improvements were made:
CCA Quantum Safe Algorithm enhancements:
– Updated support for Dilithium signatures:
• Round 2: Level 2 (6 5) and 3 (8 7)
• Round 3: Level 3 (6 5) and 5 (8 7)
– Added support for Kyber key encapsulation in Round 2: Level 5 (1024)
Quantum Safe protected key support for CCA
Host Firmware and CCA now employ a hybrid scheme combining ECDH and Kyber to
accomplish a quantum safe transport key exchange for protected key import.
CCA 8.0
The following CCA 8.0 improvements were made:
Performance enhancement for mixed workloads: better performance when one partition
focuses on RSA/ECC and another partition focuses on AES/DES/TDES or financial
operations
Hardware accelerated key unwrap for AES wrapped keys
– Trusted Key Entry workstation (TKE) controlled selection of WRAPENH3 as the default
TDES key token wrapping method for easier management.
CCA 7.1
The following CCA 7.1 improvements were made:
Supported curves:
– NIST Prime Curves: P192, P224, P256, P384, and P521
– Brainpool Curves: 160, 192, 224, 256, 320, 384, and 512
Support in the CCA coprocessor for Edwards curves ED25519 (128-bit security strength)
and ED448 (224-bit security strength)
Although ED25519 is faster, ED448 is more secure. Practically though, 128-bit security
strength is very secure.
Edwards curves are used for digitally signing documents and verifying those signatures.
Edwards curves are less susceptible to side channel attacks when compared to Prime and
Brainpool curves.
ECC Protected Keys
Crypto Express8S and Crypto Express7S provide support in CCA coprocessors to take
advantage of fast DES and AES data encryption speeds in CPACF while maintaining high
levels of security for the secure key material. The key remains encrypted and the key
encrypting key never appears in host storage.
When CCA ECC services are used, ICSF can now take advantage of ECC support in
CPACF (protected key support) for the following curves:
– Prime: P256, P384, P521
– Edwards: ED25519, ED448
CPACF can achieve much faster crypto speeds compared to the coprocessor.
The translation to protected key happens automatically after the attribute is set in the key
token. No application change is required.
Starting with IBM z14, the IBM Z crypto architecture can support up to 256 domains in an
adjunct processor (AP) with the AP extended addressing (APXA) facility that is installed. As
such, the Crypto Express adapters are enhanced to handle 256 domains.
The IBM Z firmware provides up to 85 domains for IBM z16 to customers (to match the
current LPAR maximum). Customers can map individual LPARs to unique crypto domains or
continue to share crypto domains across LPARs.
Compliance with the PCI-HSM standard is valuable for customers, particularly those
customers who are in the banking and finance industry. This certification is important to
clients for the following fundamental reasons:
Compliance is increasingly becoming mandatory
The requirements in PCI-HSM make the system more secure.
If you are a bank, acquirer, processor, or other participant in the payment card systems, the
card brands can impose requirements on you if you want to process their cards. One set of
requirements they are increasingly enforcing is the PCI standards.
The card brands work with PCI in developing these standards, and they focused first on the
standards they considered most important, particularly the PCI Data Security Standard
(PCI-DSS). Some of the other standards were written or required later, and PCI-HSM is one
of the last standards to be developed. In addition, the standards themselves were increasing
the strength of their requirements over time. Some requirements that were optional in earlier
versions of the standards are now mandatory.
In general, the trend is for the card brands to enforce more of the PCI standards and to
enforce them more rigorously. The trend in the standards is to impose more and stricter
requirements in each successive version. The net result is that companies subject to these
requirements can expect that they eventually must comply with all of the requirements.
The result of these requirements is that applications and procedures often must be updated
because they used some of the things that are now prohibited. Although this issue is
inconvenient and imposes some costs, it does increase the resistance of the systems to
attacks of various kinds. Updating a system to use PCI-HSM compliant HSMs is expected to
reduce the risk of loss for the institution and its clients.
One of the classic examples is a 16-digit credit card number. After VFPE is used to encrypt a
credit card number, the resulting cipher text is another 16-digit number. This process helps
older databases contain encrypted data of sensitive fields without having to restructure the
database or applications.
VFPE allows customers to add encryption to their applications in such a way that the
encrypted data can flow through their systems without requiring a massive redesign of their
application. In our example, if the credit card number is VFPE-encrypted at the point of entry,
the cipher text still behaves as a credit card number. It can flow through business logic until it
meets a back-end transaction server that can VFPE-decrypt it to get the original credit card
number to process the transaction.
Note: VFPE technology forms part of Visa, Inc.’s, Data Secure Platform (DSP). The use of
this function requires a service agreement with Visa. You must maintain a valid service
agreement with Visa when you use DSP/FPE.
This support includes PIN method APIs, PIN administration APIs, new key management
verbs, and new access control points support that is needed for DK-defined functions.
7 Always check the latest information about security certification status for your specific model.
UDX is integrated into the base CCA code to support translating an external RSA CRT key
into new formats. These formats use tags to identify key components. Depending on which
new rule array keyword is used with the PKA Key Translate callable service, the service TDES
encrypts those components in CBC or ECB mode. In addition, AES CMAC support is
delivered.
In EP11, keys can now be generated and securely wrapped under the EP11 Master Key. The
secure keys never leave the secure coprocessor boundary decrypted.
The secure IBM Enterprise PKCS #11 (EP11) coprocessor runs the following tasks:
Encrypt and decrypt (AES, DES, TDES, and RSA)
Sign and verify (DSA, RSA, and ECDSA)
Generate keys and key pairs (DES, AES, DSA, ECC, and RSA)
HMAC (SHA1, SHA2 or SHA3 [SHA224, SHA256, SHA384, and SHA512])
Digest (SHA1, SHS2 or SHA3 [SHA224, SHA256, SHA384, and SHA512])
Wrap and unwrap keys
Random number generation
Get mechanism list and information
Attribute values
Key Agreement (Diffie-Hellman)
Note: The function extension capability through UDX is not available to the EP11.
When defined in EP11 mode, the TKE workstation is required to manage the Crypto Express
features.
FIPS 140-2 certification is not relevant to the accelerator because it operates with clear keys
only. The function extension capability through UDX is not available to the accelerator.
The RSA encryption and decryption functions support key lengths of 512 - 4,096 bits in the
Modulus-Exponent (ME) and Chinese Remainder Theorem (CRT) formats.
Each Crypto Express8S FC 0908 includes two IBM 4770 PCIe Cryptographic Coprocessors
(PCIeCC), which is a hardware security module (HSM); FC 0909 includes one IBM 4770
PCIeCC. The adapters are available in the following configurations:
IBM Enterprise Common Cryptographic Architecture (CCA) Coprocessor (CEX8C)
IBM Enterprise Public Key Cryptography Standards #11 (PKCS) Coprocessor (CEX8P)
IBM Crypto Express7S Accelerator (CEX8A)
During the feature installation, the PCI-X adapter is configured by default as the CCA
coprocessor.
The configuration of the Crypto Express8S adapter as EP11 coprocessor requires a TKE
workstation (FC 0057/0058) with TKE 10.0 (FC 0882) LIC. The same requirement applies to
CCA mode for a full PCI-compliant environment.
The Crypto Express8S feature does not use CHPIDs from the channel subsystem pool.
However, the Crypto Express8S feature requires one slot in a PCIe I/O drawer, and one
PCHID for each PCIe cryptographic adapter.
For enabling an LPAR to use a Crypto Express8S adapter, the following cryptographic
resources in the image profile must be defined for each partition:
Usage domain index
Control domain index
PCI Cryptographic Coprocessor Candidate List
PCI Cryptographic Coprocessor Online List
Important: After this definition is modified, any change to the image profile requires a
DEACTIVATE and ACTIVATE of the logical partition for the change to take effect.
Therefore, this cryptographic definition is disruptive to a running system.
After they are activated, the active partition cryptographic definitions can be viewed from the
HMC. Select the CPC, and click View LPAR Cryptographic Controls in the CPC
Operational Customization window. The resulting window displays the definition of Usage and
Control domain indexes, and PCI Cryptographic candidate and online lists, as shown in
Figure 6-7. (Information is provided for active logical partitions only.)
Operational changes can be made by using the Change LPAR Cryptographic Controls task,
which reflects the cryptographic definitions in the image profile for the partition. With this
function, the cryptographic feature can be added and removed dynamically, without stopping
a running operating system.
For more information about the management of Crypto Express8S, see IBM z16
Configuration Setup, SG24-8960.
The TKE contains a combination of hardware and software. A mouse, keyboard, flat panel
display, PCIe adapter, and a writable USB media to install the TKE Licensed Internal Code
(LIC) are included with the system unit. The TKE workstation requires an IBM 4770 crypto
adapter.
A TKE workstation is part of a customized solution for the use of the Integrated Cryptographic
Service Facility for z/OS (ICSF for z/OS) or Linux for IBM Z. This program provides a basic
key management system for the cryptographic keys of an IBM z16 system that has Crypto
Express features installed.
The TKE provides a secure, remote, and flexible method of providing Master Key Part Entry,
and to remotely manage PCIe cryptographic coprocessors. The cryptographic functions on
the TKE run by one PCIe cryptographic coprocessor. The TKE workstation communicates
with the IBM Z system through a TCP/IP connection. The TKE workstation is available with
Ethernet LAN connectivity only. Up to 10 TKE workstations can be ordered.
TKE FCs 0057 and 0058 can be used to control any supported Crypto Express feature on
IBM z16, IBM z15, IBM z14 systems, and the Crypto adapters on older, still supported
systems.
The TKE 10.0 LIC (FC 0882) feature requires a 4770 HSM. The following features are
supported:
Managing the Crypto Express8S HSMs (CCA normal mode, CCA PCI mode, and EP11)
Quantum Safe Cryptography (QSC) used when:
– TKE authenticates Crypto Express8S HSMs
– Deriving a Transport Key between TKE’s HSM and target Crypto Express8S HSM
– On-demand HSM dual validation check.
Domain groups limitations All HSM in group must:
– Support QSC (can include Crypto Express8S HSMs only)
– Not support QSC (cannot include Crypto Express8S HSMs)
Configuration migration tasks support:
– Can collect and apply data to a Crypto Express8S HSM
– Can apply data from a pre-Crypto Express8S HSM.
New default wrapping method for the Crypto Express8S HSM
New AES DUKPT key attribute on AES DKYGENKY parts
Tip: For more information about handling a TKE, see the TKE Introduction video.
Each LPAR in the same system that uses a domain that is managed through a TKE
workstation connection is a TKE host or TKE target. An LPAR with a TCP/IP connection to the
TKE is referred to as the TKE host; all other partitions are TKE targets.
The cryptographic controls that are set for an LPAR through the SE determine whether the
workstation is a TKE host or a TKE target.
Smart card readers from FC 0885 or FC 0891 also can be carried forward. Smart cards can
be used on TKE 10.0 with these readers. Access to and use of confidential data on the smart
card are protected by a user-defined PIN. Up to 990 other smart cards can be ordered for
backup (the extra smart card feature code is FC 0900). When one feature code is ordered, 10
smart cards are included. The order increment is 1 - 99 (10 - 990 blank smart cards).
If smart cards with applets that are not supported by the new smart card reader are reused,
new smart cards on TKE 8.1 or later must be created and the content from the old smart
cards to the new smart cards must be copied. The new smart cards can be created and
copied on a TKE 8.1 system. If the copies are done on TKE 9.0, the source smart card must
be placed in an older smart card reader from feature code 0885 or 0891.
A new smart card for the Trusted Key Entry (TKE) allows stronger Elliptic Curve Cryptography
(ECC) levels. More TKE Smart Cards (FC 0900, packs of 10, FIPS certified blanks) require
TKE 9.1 LIC or up.
Note: Several options for ordering the TKE with or without ordering Keyboard, Mouse, and
Display are available. Ask your IBM Representative for more information about which
option is the best option for you.
The Omnikey Cardman 3821 smart card readers can be carried forward to any TKE 10.0
workstation. Smart cards 45D3398, 74Y0551, and 00JA710 can be used on TKE 10.0.
When performing a MES upgrade from TKE 8.x, or TKE 9.x to a TKE 10.0 installation, the
following steps must be completed:
1. Save Upgrade Data on an old TKE to USB memory to save client data.
2. Replace the 4767 crypto adapter with the 4768 crypto adapter.
3. Upgrade the firmware to TKE 10.0.
4. Install the Frame Roll to apply Save Upgrade Data (client data) to the TKE 10.0 system.
5. Run the TKE Workstation Setup wizard.
Note: If your IBM z16 includes Crypto Express7S or Crypto Express6S, you can use TKE
V9.2, which requires the 4768 cryptographic adapter.
For more information about TKE hardware support, see Table 6-3. For some functions,
requirements must be considered; for example, the characterization of a Crypto Express
adapter in EP 11 mode always requires the use of a TKE.
Attention: The TKE is unaware of the CPC type where the host crypto module is installed.
That is, the TKE does not consider whether a Crypto Express is running on IBM z16, IBM
15, or IBM z14 system. Therefore, the LIC can support any CPC where the coprocessor is
supported, but the TKE LIC must support the specific crypto module.
Offers UDX - X - -
RSA functions - X X X
For more information about the minimum that is required and recommended distribution
levels, see the IBM Z website.
For more information about the software support levels for cryptographic functions, see
Chapter 7, “Operating system support” on page 247.
Because this information is subject to continuous updating, for the most current information
check the hardware fix categories for IBM z16 A01 IBM.Device.Server.IBM z16-3931.*.
Support of IBM z16 functions depends on the operating system and version and release.
End of service operating systems: Operating system levels that are no longer in service
are not covered in this publication.
z/OS V2R2b
z/VM V7R1
z/VSE V6.2
z/TPF V1R1
The use of specific features depends on the operating system. In all cases, program
temporary fixes (PTFs) might be required with the operating system level that is indicated.
Check the z/OS fix categories, or the subsets of the 3931DEVICE (IBM z16 A01) PSP
buckets for z/VM and z/VSE.
The fix categories and the PSP buckets are continuously updated. They contain the latest
information about maintenance.
Hardware and software buckets contain installation information, hardware and software
service levels, service guidelines, and cross-product dependencies.
For more information about Linux on IBM Z distributions and KVM hypervisor, see the
distributor’s support information.
Features and functions that are available on previous servers, but no longer supported by IBM
z16 servers, are not documented here.
For more information about supported functions that are based on operating systems, see
7.3, “IBM z16 features and function support overview” on page 252. Tables are built by
function and feature classification to help you determine, by a quick scan, what is supported
and the minimum operating system level that is required.
7.2.1 z/OS
z/OS Version 2 Release 3 is the earliest in-service release that supports IBM z16 servers.
Consider the following points:
Service support for z/OS Version 2 Release 2 ended in September of 2020; however, a
fee-based extension for defect support (for up to three years) can be obtained by ordering
IBM Software Support Services - Service Extension for z/OS V2.R2.
IBM z16 capabilities differ depending on the z/OS release. Toleration support is provided
on z/OS V2R2. Exploitation support is provided on z/OS V2R3 and later only1.
Fixes are grouped into the following categories (for more information about IBM Fix
Categories, see this IBM Support web page):
Base support is provided by PTFs identified by:
IBM.Device.Server.-3931.RequiredService
Fixes that are required to run z/OS on the IBM z16 servers and must be installed before
migration.
Use of many functions is provided by PTFs identified by:
IBM.Device.Server.-3931.Exploitation
Fixes that are required to use the capabilities of the IBM z16. They must be installed only
if you use the function.
Recommended service is identified by:
IBM.Device.Server.-3931.RecommendedService
Fixes that are recommended to run z/OS on the IBM z16. These fixes also are listed in the
Recommended Service section of the hardware PSP bucket. They represent fixes that
were recommended by IBM Service. It is recommended that you review and install these
PTFs.
For more information about supported functions and their minimum required support levels,
see 7.3, “IBM z16 features and function support overview” on page 252.
1 Use support for select features by way of PTFs. Toleration support for new hardware might also require PTFs.
Note: At the time of this writing, z/VM V7.3 is planned for third quarter 2022 availability.
z/VM Compatibility Support enables guest use for several new facilities:
Embedded Artificial Intelligence Acceleration
– Designed to reduce the overall time required to execute CPU operations for neural
networking processing functions, and help support real-time applications, such as
fraud detection.
Compliance-ready CPACF Counters support
– Provides a means for guests to track crypto compliance and instruction usage.
Breaking Event Address Register (BEAR) Enhancement Facility;
– Facilitates debugging wild branches.
Vector Packed Decimal Enhancements 2
– New instructions intended to provide performance improvements.
Reset DAT protection Facility
– Provides a more efficient way to disable DAT protection, such as during copy-on-write
or page change tracking operations.
RoCE Express3 adapter
– Allows guests to use Routable RoCE, Zero Touch RoCE, and SMC-R V2 support.
Guest Enablement for the CEX8S crypto adapter and assorted crypto enhancements
– Including Quantum Safe API Guest Exploitation Support that is available to dedicated
guests.
CPU/Core topology location information within z/VM monitor data
– Provides a better picture of the system for diagnostic and tuning purposes.
Consolidated Boot Loader for guest IPL from SCSI
For more information about supported functions and their minimum required support levels,
see 7.3, “IBM z16 features and function support overview” on page 252.
2 z/VM 7.3 preview announcement. See the IBM z/VM continuous delivery page.
For more information about supported functions and their minimum required support levels,
see 7.3, “IBM z16 features and function support overview” on page 252.
7.2.5 z/TPF
IBM z16 support is provided by z/TPF V1R1 with PTFs. For more information about
supported functions and their minimum required support levels, see 7.3, “IBM z16 features
and function support overview” on page 252.
The service levels of SUSE, Red Hat, and Ubuntu releases that are supported at the time of
this writing are listed in Table 7-2.
Ubuntu 21.10c
For more information about supported Linux distributions on IBM Z servers, see the Tested
platforms for Linux page of the IBM IT infrastructure website.
IBM is working with Linux distribution Business Partners to provide further use of selected
IBM z16 functions in future Linux on IBM Z distribution releases.
For more information about KVM support, see the IBM Z website.
Information about Linux on IBM Z refers exclusively to the suitable distributions of SUSE,
Red Hat, and Ubuntu.
The tables in this section list but do not specifically mark all the features that require fixes that
are required by the corresponding operating system for toleration or use.
Maximum processor unit (PUs) per system image 200 200 200 80c 80c 80c
Dynamic PU add Y Y Y Y Y Y
Program-directed re-IPL - - - Y Y Y
Transactional Executione Y Y Y Yf Yf Yf
Flexible Capacity Y Y Y Y Y Y
CF level 25 Enhancements Yk Yk Yk - - -
HiperDispatch Optimization Y Y Y Yh Yh Yh
System Recovery Boost Enhancements (IBM z16) Yk Yk N/A N/A N/A N/A
ICSF Enhancements Y Y Y - - -
The supported base CPC functions for z/VSE, z/TPF, and Linux on IBM Z are listed in
Table 7-4.
Table 7-4 Supported base CPC functions for z/VSE, z/TPF, and Linux on IBM Z
Functiona z/VSE z/TPF Linux on
V6R2 V1R1 IBM Zb
Dynamic PU add Y N Y
Program-directed re-IPL Y N Y
HiperDispatch N N Y
AI accelerator exploitation N N Yh
Note: z/OS V2R2 support ended as of September 2020. No new function is provided for
the use of the new hardware features (toleration support only). Although extended
(fee-based) support for z/OS V2.R2 can be obtained, support for z/OS V2.R2 is not
covered extensively in this document.
Table 7-5 Supported coupling and clustering functions for z/OS and z/VM
Functiona z/OS z/OS z/OS z/VM z/VM z/VM
V2R5 V2R4 V2R3 V7R3 V7R2 V7R1
In addition to operating system support that is listed in Table 7-5, Server Time Protocol is
supported on z/TPF V1R1 and Linux on IBM Z. Also, CFCC Level 23, Level 24, and Level 25
are supported for z/TPF V1R1.
Table 7-6 Supported storage connectivity functions for z/OS and z/VM
Functiona z/OS z/OS z/OS z/VM z/VM z/VM
V2R5 V2R4 V2R3 V7R3 V7R2 V7R1
Table 7-7 Supported storage connectivity functions for z/VSE, z/TPF, and Linux on IBM Z
Functiona z/VSE z/TPF Linux on
V6R2 V1R1 IBM Zb
Table 7-8 Supported network connectivity functions for z/OS and z/VM
Functiona z/OS z/OS z/OS z/VM z/VM z/VM
V2R4 V2R3 V2R2 V7R2 V7R1 V6R4
HiperSockets
HiperSocketsc Y Y Y Y Y Y
The supported network connectivity functions for z/VSE, z/TPF, and Linux on IBM Z are listed
in Table 7-9.
Table 7-9 Supported network connectivity functions for z/VSE, z/TPF, and Linux on IBM Z
Functiona z/VSE z/TPF Linux on
V6R2 V1R1 IBM Zb
HiperSockets
HiperSocketsd Y N Y
Note: z/OS V2R2 support ended as of September 2020. No new function is provided for
leveraging the new HW features (toleration support only). Although extended (fee-based)
support for z/OS V2.R2 can be obtained, support for z/OS V2.R2 is not covered
extensively in this document.
Crypto Express8S Y Y Y Yb Yb Yb
Crypto Express7S Y Y Y Yb Yb Yb
Crypto Express6S Y Y Y Yb Yb Yb
the IBM z16 supported cryptography functions for z/VSE, z/TPF, and Linux on IBM Z are
listed in Table 7-11.
Table 7-11 Supported cryptography functions for z/VSE, z/TPF, and Linux on IBM Z
Functiona z/VSE z/TPF Linux on
V6R2 V1R1 IBM Zb
Crypto Express8S Y Y Y
Crypto Express7S Y Y Y
Crypto Express6S Y Y Y
Note: z/OS V2R2 support ended as of September 2020. No new function is provided for
the use of the new hardware features (toleration support only). Although extended
(fee-based) support for z/OS V2.R2 can be obtained, support for z/OS V2.R2 is not
covered extensively in this document.
z/VSE V6.2 and later z/VSE Turbo Dispatcher can use up to 4 CPs, and tolerates up to
10-way LPARs
Linux on IBM Z SUSE Linux Enterprise Server 12 and later: 256 CPs or IFLs.
Red Hat RHEL 7 and later: 256 CPs or IFLs.
Ubuntu 20.04.1 LTS and later: 256 CPs or IFLs.
KVM Hypervisor The KVM hypervisor is offered with the following Linux distributions
-- 256CPs or IFLs--:
SLES 12 SP5 and later
RHEL 7.9 and later
Ubuntu 20.04.1 LTS and later
z/VM z/VM V7R1 supports 2 TB while z/VM 7.2b and z/VM 7.3 support 4 TB
Dynamic PU add
Planning an LPAR configuration includes defining reserved PUs that can be brought online
when extra capacity is needed. Operating system support is required to use this capability
without an IPL; that is, nondisruptively. This support is available in z/OS for some time.
The dynamic PU add function enhances this support by allowing you to dynamically define
and change the number and type of reserved PUs in an LPAR profile, which removes any
planning requirements. The new resources are immediately made available to the operating
system and in the case of z/VM, to its guests.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
z/VM virtualizes this support to its guests, which now also can increase their memory
nondisruptively if supported by the guest operating system. Currently, releasing memory from
z/VM is supported on z/VM V7.2 with PTFs3. Releasing memory from the z/VM guest
depends on the guest’s operating system support.
Linux on IBM Z also supports acquiring and releasing memory nondisruptively. This feature is
enabled for SUSE Linux Enterprise Server 12 and RHEL 7.9and later releases.
The Capacity Provisioning Manager, which is a feature that was first available with z/OS
V1R9, interfaces with z/OS Workload Manager (WLM) and implements capacity provisioning
policies. Several implementation options are available, from an analysis mode that issues
only guidelines, to an autonomic mode that provides fully automated operations.
Program-directed re-IPL
Program directed re-IPL allows an operating system to re-IPL without operator intervention.
This function is supported for SCSI and IBM extended count key data (IBM ECKD) devices.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
IOCP
All IBM Z servers require a description of their I/O configuration. This description is stored in
I/O configuration data set (IOCDS) files. The I/O configuration program (IOCP) allows for the
creation of the IOCDS file from a source file that is known as the I/O configuration source
(IOCS).
The IOCS file contains definitions of LPARs and channel subsystems. It also includes detailed
information for each channel and path assignment, control unit, and device in the
configuration.
3
z/VM Dynamic Memory Downgrade (releasing memory from z/VM LPAR) made available with PTFs for APAR
VM66271. For more information, see: http://www.vm.ibm.com/newfunction/#dmd
For more information, see IBM Dynamic Partition Manager (DPM) Guide, SB10-7182.
HiperDispatch
The HIPERDISPATCH=YES/NO parameter in the IEAOPTxx member of SYS1.PARMLIB and on
the SET OPT=xx command controls whether HiperDispatch is enabled or disabled for a z/OS
image. It can be changed dynamically, without an IPL or any outage.
In z/OS, the IEAOPTxx keyword HIPERDISPATCH defaults to YES when it is running on an IBM
z16, IBM z15, or IBM IBM z14system.
The use of SMT on IBM z16 systems requires that HiperDispatch is enabled on the operating
system. For more information, see “Simultaneous multithreading” on page 273.
Also, any LPAR that is running with more than 64 logical processors is required to operate in
HiperDispatch Management Mode.
HiperDispatch on IBM z16 systems uses chip and CPC drawer configuration to improve the
access cache performance. It optimizes the system PU allocation with Chip/cluster/drawer
cache structure on IBM Z servers.
The base support for IBM z16 is provided by PTFs that are identified by:
IBM.device.server.IBM z16-3931.requiredservice
PR/SM on IBM z16 servers preferentially assigns memory for a system in one CPC drawer
that is striped across the clusters of that drawer to take advantage of the lower latency
memory access in a drawer. Also, PR/SM tries to consolidate storage onto drawers with the
most processor entitlement.
With HiperDispatch enabled, PR/SM seeks to assign logical processors of a partition to the
smallest number of PU chips within a drawer in cooperation with operating system to optimize
shared cache usage.
PR/SM automatically keeps a partition’s memory and logical processors on the same CPC
drawer where possible. This arrangement looks simple for a partition, but it is a complex
optimization for multiple logical partitions because some must be split among processors
drawers.
All IBM z16 processor types can be dynamically reassigned except IFPs.
To use HiperDispatch effectively, WLM goal adjustment might be required. Review the WLM
policies and goals and update them as necessary.
WLM policies can be changed without turning off HiperDispatch. A health check is provided to
verify whether HiperDispatch is enabled on a system image.
z/TPF
z/TPF on IBM z16 can use more processors immediately without reactivating the LPAR or
IPLing the z/TPF system.
When z/TPF is running in a shared processor configuration, the achieved MIPS is higher
when z/TPF uses a minimum set of processors.
In low-use periods, z/TPF minimizes the processor footprint by compressing TPF workload
onto a minimal set of I-streams (engines), which reduces the effect on other LPARs and
allows the entire CPC to operate more efficiently.
As a consequence, z/OS and z/VM experience less contention from the z/TPF system when
the z/TPF system is operating at periods of low demand.
zIIP support
zIIPs do not change the model capacity identifier of IBM z16 servers. IBM software product
license charges that are based on the model capacity identifier are not affected by the
addition of zIIPs.
No changes to applications are required to use zIIPs. They can be used by the following
applications:
Db2 V8 and later for z/OS data serving for applications that use data Distributed Relational
Database Architecture (DRDA) over TCP/IP, such as data serving, data warehousing, and
selected utilities.
z/OS XML services.
z/OS CIM Server.
z/OS Communications Server for network encryption (Internet Protocol Security [IPsec])
and for large messages that are sent by HiperSockets.
IBM GBS Scalable Architecture for Financial Reporting.
IBM z/OS Global Mirror (formerly XRC) and System Data Mover.
IBM z/OS Container Extensions.
IBM OMEGAMON® XE on z/OS, OMEGAMON XE on Db2 Performance Expert, and Db2
Performance Monitor.
Any Java application that uses the current IBM SDK.
Java IBM Semeru Runtime offloading enablement for DLC models that use Integrated
Accelerator for AI.
WebSphere Application Server V5R1 and later, and products that are based on it, such as
WebSphere Portal, WebSphere Enterprise Service Bus (WebSphere ESB), and
WebSphere Business Integration (WBI) for z/OS.
CICS/TS V2R3 and later.
Db2 UDB for z/OS Version 8 and later.
IMS Version 8 and later.
zIIP Assisted HiperSockets for large messages.
z/OSMF (z/OS Management Facility).
IBM z/OS Platform for Apache Spark.
IBM Watson® Machine Learning for z/OS.
z/OS System Recovery Boost.
Approved third-party vendor products.
The use of a zIIP is transparent to application programs. The supported operating systems
are listed in Table 7-3 on page 253.
On IBM z16 servers, the zIIP processor is designed to run in SMT mode, with up to two
threads per processor. This function is designed to help improve throughput for zIIP
workloads and provide appropriate performance measurement, capacity planning, and SMF
accounting data. zIIP support is available on all currently supported z/OS versions.
The field APPL% IIPCP of the Workload Activity Report listing by WLM service class
indicates the percentage of a processor that is zIIP eligible. Because of the zIIP’s lower price
as compared to a CP, even a utilization as low as 10% can provide cost benefits.
Transactional Execution4
Transactional Execution (TX) is known in academia and industry as hardware transactional
memory. Transactional execution is implemented on IBM Z servers.
This feature enables software to indicate to the hardware the beginning and end of a group of
instructions that must be treated in an atomic way. All of their results occur or none occur, in
true transactional style. The execution is optimistic.
The hardware provides a memory area to record the original contents of affected registers
and memory as instruction execution occurs. If the transactional execution group is canceled
or must be rolled back, the hardware transactional memory is used to reset the values.
Software can implement a fallback capability.
This capability increases the software’s efficiency by providing a way to avoid locks (lock
elision). This advantage is of special importance for speculative code generation and highly
parallelized applications.
TX is used by IBM Java virtual machine (JVM) and might be used by other software. The
supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on page 254.
4
Statement of Direction: In a future IBM Z hardware system family, the transactional execution and constrained
transactional execution facility will no longer be supported. Users of the facility on current servers should always
check the facility indications before use.
Automation
The client’s automation product can be used to automate and control the following System
Recovery Boost activities:
To activate and deactivate the eBod temporary capacity record to provide more physical
zIIPs for an IPL or Shutdown Boost.
To dynamically modify LPAR weights, as might be needed to modify sharing physical zIIP
capacity during a Boost period.
To drive the invocation of the PROC that indicates the beginning of a shutdown process
(and the start of the shut-down Boost).
To take advantage of new composite hardware API reconfiguration actions.
To control the level of parallelism that is present in the workload at startup (for example,
starting middleware regions) and shutdown (for example, performing an orderly shutdown
of middleware).
Simultaneous multithreading
SMT is the hardware capability to process up to two simultaneous threads in a single core,
which shares the resources of the core, such as cache, translation lookaside buffer (TLB),
and execution resources. This sharing improves system capacity and efficiency by reducing
processor delays, which increases the overall throughput of the system.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
An operating system that uses SMT controls each core and is responsible for maximizing its
throughput and meeting workload goals with the smallest number of cores. In z/OS, consider
HiperDispatch cache optimization when you must choose the two threads to be dispatched in
the same processor.
HiperDispatch attempts to dispatch guest virtual CPUs on the same logical processor on
which they ran. PR/SM attempts to dispatch a vertical low logical processor in the same
physical processor. If that process is not possible, it attempts to dispatch it in the same node,
or then the same CPC drawer where it was dispatched before to maximize cache reuse.
From the perspective of an application, SMT is transparent and no changes are required in
the application for it to run in an SMT environment, as shown in Figure 7-1.
MT Ignorant
z/OS z/VM
PR/SM Hypervisor MT Aware
z/OS
The use of SMT on z/OS requires enabling HiperDispatch, and defining the processor view
(PROCVIEW) control statement in the LOADxx parmlib member and the MT_ZIIP_MODE
parameter in the IEAOPTxx parmlib member.
The PROCVIEW statement is defined for the life of IPL, and can have the following values:
CORE: This value specifies that z/OS should configure a processor view of core, in which a
core can include one or more threads. The number of threads is limited by IBM z16 to two
threads. If the underlying hardware does not support SMT, a core is limited to one thread.
CPU: This value is the default. It specifies that z/OS should configure a traditional processor
view of CPU and not use SMT.
CORE,CPU_OK: This value specifies that z/OS should configure a processor view of core (as
with the CORE value) but the CPU parameter is accepted as an alias for applicable
commands.
The MT_ZIIP_MODE parameter in the IEAOPTxx controls zIIP SMT mode. It can be 1 (the
default), where only one thread can be running in a core, or 2, where up two threads can be
running in a core. If PROCVIEW CPU is specified, the MT_ZIIP_MODE is always 1. Otherwise, the
use of SMT to dispatch two threads in a single zIIP logical processor (MT_ZIIP_MODE=2) can be
changed dynamically by using the SET OPT=xx setting in the IEAOPTxx parmlib. Changing the
MT mode for all cores can take some time to complete.
PROCVIEW CORE requires DISPLAY M=CORE and CONFIG CORE to display the core states and
configure an entire core.
With the introduction of Multi-Threading support for SAPs, a maximum of 88 logical SAPs can
be used. RMF is updated to support this change by implementing page break support in the
I/O Queuing Activity report that is generated by the RMF Post processor.
The default in z/VM is multithreading disabled. Dynamic SMT enables dynamically varying the
active threads per core. The number of active threads per core can be changed dynamically
without a system outage and potential capacity gains going from SMT-1 to SMT-2 (one to two
threads per core) can be achieved dynamically.
z/VM V7R3, V7R2, and V7R1 support up to 40 multithreaded cores (80 threads) for IFLs, and
each thread is treated as an independent processor. z/VM dispatches virtual IFLs on the IFL
logical processor so that the same or different guests can share a core. Each core has a
single dispatch vector, and z/VM attempts to place virtual sibling IFLs on the same dispatch
vector to maximize cache reuses.
z/VM guests have no awareness of SMT, and cannot use it directly. z/VM SMT use does not
include guest support for multithreading. The value of this support for guests is that the
first-level z/VM host of the guests can achieve higher throughput from the multi-threaded IFL
cores.
The following minimum releases of Linux on IBM Z distributions are supported on IBM z16
(native SMT support):
SUSE:
– SLES 16
– SLES 15 SP3 with service
– SUSE SLES 12 SP5 with service
The KVM hypervisor is supported on the same Linux on IBM Z distributions in this list.
For more information about the most current support, see the Linux on IBM Z Tested
platforms website.
Single-instruction multiple-data
The SIMD feature introduces a new set of instructions to enable parallel computing that can
accelerate code with string, character, integer, and floating point data types. The SIMD
instructions allow many operands to be processed with a single complex instruction.
IBM z16 is equipped with a set of instructions to improve the performance of complex
mathematical models and analytic workloads through vector processing and complex
instructions, which can process numerous data with a single instruction. This set of
instructions, which is known as SIMD, enables more consolidation of analytic workloads and
business transactions on IBM Z servers.
SIMD on IBM z16 has support for enhanced math libraries that provide performance
improvements for analytical workloads by processing more information with a single CPU
instruction.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254. Operating System support includes the following features6:
Enablement of vector registers.
A math library with an optimized and tuned math function (Mathematical Acceleration
Subsystem [MASS]) that can be used in place of some of the C standard math functions. It
includes a SIMD vectorized and nonvectorized version.
A specialized math library, which is known as Automatically Tuned Linear Algebra
Software (ATLAS), that is optimized for the hardware.
IBM Language Environment® for C runtime function enablement for ATLAS.
DBX to support the disassembly of the new vector instructions, and to display and set
vector registers.
XML SS exploitation to use new vector processing instructions to improve performance.
MASS and ATLAS can reduce the time and effort for middleware and application developers.
IBM provides compiler built-in functions for SIMD that software applications can use as
needed, such as for using string instructions.
6 The features that are listed here might not be available on all operating systems that are listed in the tables.
Vector programming support is extended for IBM z16 to provide access to the new
instructions that were introduced by the VEF 27 specification.
Older levels of z/OS XL C/C++ compilers do not provide IBM z16 exploitation; however, the
z/OS V2R5 XL C/C++ compiler can be used to generate code for the older levels of z/OS
running on IBM z16.
Code must be developed to use the SIMD functions. Applications with SIMD instructions
abend if they run on a lower hardware level system that do not support SIMD. Some
mathematical function replacement can be done without code changes by including the scalar
MASS library before the standard math library.
Because the MASS and standard math library include different accuracies, assess the
accuracy of the functions in the context of the user application before deciding whether to use
the MASS and ATLAS libraries.
The SIMD functions can be disabled in z/OS partitions at IPL time by using the MACHMIG
parameter in the LOADxx member. To disable SIMD code, use the MACHMIG VEF
hardware-based vector facility. If you do not specify a MACHMIG statement, which is the default,
the system not limited in its use of the Vector Facility for z/Architecture (SIMD).
IBM z16 introduces COBOL optimization for Hexadecimal Floating Point (HFP) <--> Binary
Coded Decimal (BCD) conversion, and Numeric Editing, and Zoned Decimal operations.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254. For more information, see 7.5.8, “z/OS XL C/C++ considerations” on page 321.
Out-of-order execution
Out-of-order (OOO) execution yields significant performance benefits for compute-intensive
applications by reordering instruction execution, which allows later (newer) instructions to be
run ahead of a stalled instruction, and reordering storage accesses and parallel storage
accesses. OOO maintains good performance growth for traditional applications.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254. For more information, see “3.4.3, “Out-of-Order execution” on page 81.
z/OS DFSORT takes advantage of Z Sort, and provides users with significant performance
boosts for their sort workloads. With z/OS DFSORT's Z Sort algorithm, clients can see batch
sort job elapsed time improvements of up to 20 - 30% (depending on record size) and CPU
time improvements of up to 40% compared to IBM z14.
The function is used on z/OS V2.R5 and enabled on z/OS V2.R4 and V2.R3 with PTFs for
APAR PH03207.
The sort jobs must meet certain eligibility criteria. For the full list and other considerations,
see the DFSORT User Guide for PH03207.
The supported operating systems are listed in Table 7-3 on page 253.
For more information about this function, see this IBM Support web page.
For more information about the CPU Measurement Facility, see this IBM Support web page.
For more information, see “12.2, “IBM z16 Large System Performance Reference ratio” on
page 476.
SMF 30 records were enhanced to include new crypto counter sections that contain counters
for CPACF cryptographic instructions that are used by a job in a specific period. These
sections are produced only for those instructions that are used.
These counters are correlated with z/OS jobs and users for the determination of the
algorithms, bit lengths, and key security that is used by a specific workload. This data can aid
in compliance, performance, and configuration.
The SMF 30 self-defining section indicates the length and number of crypto counter sections.
The SMF 30 product and subsystem section indicates whether the crypto counters are active.
This feature is supported on z/OS 2.4 and later. It requires APAR OA61511.
z/OS V2.R5 allows 2 GB LFAREA to exceed the 4 TB limit, up to 16 TB. All online real storage
over 4 TB is part of the 2 GB LFAREA, in addition to what was specified in LFAREA.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
AI accelerator exploitation
With the IBM z16 Integrated Accelerator for AI, customers can benefit from the acceleration of
AI operations, such as fraud detection, customer behavior predictions, and streamlining of
supply chain operations; all in real time. AI accelerator is designed to deliver AI inference in
real time, at large scale and rate, with no transaction left behind so that customers can
instantly derive the valuable insights from their data.
The AI capability is applied directly into the running transaction, shifting the traditional
paradigm of applying AI to the transactions that were completed. This innovative technology
can be used for intelligent IT workloads placement algorithms and contribute to better overall
system performance. The co-processor is driven by the new Neural Networks Processing
Assist (NNPA) instruction.
IBM Virtual Flash Memory is designed to help improve availability and handling of paging
workload spikes when running z/OS. With this support, z/OS is designed to help improve
system availability and responsiveness by using VFM across transitional workload events,
such as market openings, and diagnostic data collection. z/OS also helps improve processor
performance by supporting middleware use of pageable large (1 MB) pages.
VFM can help organizations meet their most demanding service level agreements and
compete more effectively. VFM is easily configurable, and provides rapid time to value.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
z/OS
GSF support allows an area of storage to be identified such that an Exit routine assumes
control if a reference is made to that storage. GSF is managed by new instructions that define
Guarded Storage Controls and system code to maintain that control information across
undispatch and redispatch.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
Through enhanced hardware features (based on DAT table entry bit) and specific software
requests to obtain memory areas as nonexecutable, areas of memory can be protected from
unauthorized execution. A Protection Exception occurs if an attempt is made to fetch an
instruction from an address in such an element or if an address in such an element is the
target of an execute-type instruction.
z/OS
To use IEP, Real Storage Manager (RSM) is enhanced to request nonexecutable memory
allocation. Use new keyword EXECUTABLE=YES|NO on STORAGE OBTAIN or IARV64 to indicate
whether memory that is to be used contains executable code. Recovery Termination Manager
(RTM) writes LOGREC record of any program-check that results from IEP.
IEP support is for z/OS V2.R2 and later running on IBM z16 with APARs OA51030 and
OA51643 installed.
z/VM
Guest exploitation support for the Instruction Execution Protection Facility is provided with
APAR VM65986.
The supported operating systems are listed in Table 7-3 on page 253 and Table 7-4 on
page 254.
Secure Boot
Secure Boot verification ensures that the Linux distribution kernel comes from an official
provider and was not compromised. If the signature of the distribution cannot be verified, the
process of booting the operating system is stopped.
For more information, see Appendix C, “IBM Integrated Accelerator for zEnterprise Data
Compression” on page 507.
Each PU chip includes one on-chip compression unit, which is designed to replace the
zEnterprise Data Compression (zEDC) Express PCIe feature that is available on IBM z14 and
earlier.
The IBM Integrated Accelerator for zEDC maintains software compatibility with zEDC
Express use cases. For more information, see Integrated Acceleratorfor zEnterprise Data
Compression.
All data interchange with zEDC compressed data remains compatible as IBM z16 and zEDC
capable machines coexist (accessing same data). Data that is compressed and written with
zEDC can be read and decompressed by IBM z16 well into the future.
Function support for the IBM Integrated Accelerator for zEDC is listed in Table 7-3 on
page 253 and Table 7-4 on page 254.
For more information, see Appendix C, “IBM Integrated Accelerator for zEnterprise Data
Compression” on page 507.
CFCC Level 25
CFCC Level 25 is delivered on IBM z16 servers with driver level 51 and introduces the
following enhancements:
Scalability Improvements
Processing and dispatching enhancements that result in meaningful scaling of effective
throughput up to the limit of 16 ICF processors.
Request latency/performance improvements
CFCC and coupling link firmware and hardware improvements to reduce link latency.
Elimination of VSAM RLS orphaned cast-out lock problems and improved VSAM RLS
Structure Full recovery processing.
Addresses reoccurring problems that are encountered by installations running VSAM RLS
data sharing.
Retry Buffer support that is used on list and lock commands is extended to nonidempotent
cache commands and optimized lock commands.
The new support also allows connectors to lock structures to specify a percentage of
record data entries to be reserved. These reserved entries are off limits to normal
requests to the coupling facility and can be used only if a new keyword is used on lock
requests that generate record data entries.
Cache residency time metrics
The CF calculates in microseconds by way of a moving weighted average the elapsed
time a data area or directory entry resides in a storage class before it is reclaimed. XES
returns this information on an IXLCACHE REQUEST=READSTGSTATS and IXLMG
STRNAME=strname,STGCLASS=stgclass request.
DYNDISP=ON|OFF is deprecated
For CFCC Level 25, DYNDISP=THIN is the only available behavior for shared-engine CF
dispatching.
Specifying OFF or ON in CF commands and the CF configuration file is preserved for
compatibility, but a warning message is issued to indicate that these options are no longer
supported, and that DYNDISP=THIN behavior is to be used.
Before you begin the migration process, install the compatibility and coexistence PTFs. A
planned outage is required when you upgrade the CF or CF LPAR to CFCC Level 25.
CFCC Level 24
CFCC Level 24 is delivered on IBM z16 servers with driver level 41. CFCC Level 24
introduced the following enhancements:
CFCC Fair Latch Manager
This enhancement to the internals of the Coupling Facility (CFCC) dispatcher provides CF
work management efficiency and processor scalability improvements. It also improves the
“fairness” of arbitration for internal CF resource latches across tasks
For more information about CFCC code levels, see the Parallel Sysplex page of the IBM IT
infrastructure website.
For more information about the latest CFCC code levels, see the current exception letter that
is published on Resource Link website (login is required).
CF structure sizing changes are expected when upgrading from a previous CFCC Level to
CFCC Level 25. In fact, CFLEVEL 25 can have more noticeable CF structure size increases
associated with it, especially for smaller structures, because of task-related memory
increases that are associated with the increased number of CF tasks in CFLEVEL 25
Alternatively, the batch SIZER utility also can be used for re-sizing your CF structures as
needed. Make sure to update CFRM Policy INITISIZE or SIZE values as needed.
For more information, see 4.6.4, “Parallel Sysplex connectivity” on page 187.
For more information, see “Coupling Thin Interrupts” on page 102. The supported operating
systems are listed in Table 7-5 on page 256.
Asynchronous CF Duplexing for lock structures requires the following software support:
z/OS V2R5, V2R4
z/OS V2R3 with PTFs for APAR OA47796 and OA49148
z/VM V7R3, V7R2 or V7R1
Db2 V12 with PTFs for APAR PI66689
IRLM V2.R3 with PTFs for APAR PI68378
The supported operating systems are listed in Table 7-5 on page 256.
Instead of performing XI signals synchronously on every cache update request that causes
them, data managers can “opt in” for the CF to perform these XIs asynchronously (and then
synchronize them with the CF at or before the transaction is completed). Data integrity is
maintained if all XI signals complete by the time transaction locks are released.
The feature enables faster completion of cache update CF requests, especially with the
cross-site distance that is involved. It also provides improved cache structure service times
and coupling efficiency. It requires specific data manager use or participation, which is not
transparent to the data manager. No SMF data changes were made for CF monitoring and
reporting.
8 Coupling Express LR features are not carried forward to IBM z16. Coupling Express2 LR is available.
This function refers exclusively to the z/VM dynamic I/O support of InfiniBand9 and ICA
coupling links. Support is available for the CIB and CS5 CHPID type in the z/VM dynamic
commands, including the change channel path dynamic I/O command.
Specifying and changing the system name when entering and leaving configuration mode are
also supported. z/VM does not use InfiniBand or ICA, and does not support the use of
InfiniBand or ICA coupling links by guests. The supported operating systems are listed in
Table 7-5 on page 256.
zHyperlink Express
IBM IBM z14 introduced IBM zHyperLink Express as a brand new IBM Z input/output (I/O)
channel link technology since FICON. The zHyperLink Express 1.1 feature is available with
new IBM z16 systems and is designed to help bring data close to processing power, increase
the scalability of transaction processing, and lower I/O latency.
zHyperLink Express is designed for up to 5x lower latency than High-Performance FICON for
IBM Z (zHPF) by directly connecting the IBM Z central processor complex (CPC) to the I/O
Bay of the DS8000 (DS8880 or later). This short distance (up to 150 m [492.1 feet]), direct
connection is intended to speed Db2 for z/OS transaction processing and improve active log
throughput.
The improved performance of zHyperLink Express allows the Processing Unit (PU) to make a
synchronous request for the data that is in the DS8000 cache. This process eliminates the
undispatch of the running request, the queuing delays to resume the request, and the PU
cache disruption.
Support for zHyperLink Writes can accelerate Db2 log writes to help deliver superior service
levels by processing high-volume Db2 transactions at speed. IBM zHyperLink Express
requires compatible levels of DS8000/F hardware, firmware R8.5.1 or later, and Db2 12 with
PTFs.
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258.
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258.
FICON Express16SA
FICON Express16SA (carry forward to IBM z16 server) supports a link data rate of
16 gigabits per second (Gbps) and auto negotiation to 8 Gbps for synergy with switches,
directors, and storage devices. With support for native FICON, High-Performance FICON for
IBM Z (zHPF), and Fibre Channel Protocol (FCP), the IBM z16 server enables you to position
your SAN for even higher performance, which helps you to prepare for an end-to-end 16 Gbps
infrastructure to meet the lower latency and increased bandwidth demands of your
applications.
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258.
FICON Express16S+
FICON Express16S+ (carry forward to IBM z16) supports a link data rate of 16 Gbps and auto
negotiation to 4 or 8 Gbps for synergy with switches, directors, and storage devices. With
support for native FICON, High-Performance FICON for Z (zHPF), and Fibre Channel
Protocol (FCP), the IBM Z enable you to position your SAN for even higher performance,
which helps you to prepare for an end-to-end 16 Gbps infrastructure to meet the lower latency
and increased bandwidth demands of your applications.
The new FICON Express16S+ channel works with your fiber optic cabling environment
(single mode and multimode optical cables). The FICON Express16S+ feature running at
end-to-end 16 Gbps link speeds provides reduced latency for large read/write operations and
increased bandwidth compared to the FICON Express8S feature.
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258.
To use this enhancement, the control unit must support the new IU pacing protocol. IBM
System Storage DS8000 series supports extended distance FICON for IBM Z environments.
The channel defaults to current pacing values when it operates with control units that cannot
use extended distance FICON.
High-performance FICON
High-performance FICON (zHPF) was first provided on IBM System z10®, and is a FICON
architecture for protocol simplification and efficiency. It reduces the number of information
units (IUs) that are processed. Enhancements were made to the z/Architecture and the
FICON interface architecture to provide optimizations for online transaction processing
(OLTP) workloads.
When used by the FICON channel, the z/OS operating system, and the DS8000 control unit
or other subsystems, the FICON channel processor usage can be reduced and performance
improved. Suitable levels of Licensed Internal Code (LIC) are required.
Also, the changes to the architectures provide end-to-end system enhancements to improve
reliability, availability, and serviceability (RAS).
For example, the zHPF channel programs can be used by the z/OS OLTP I/O workloads,
Db2, VSAM, the partitioned data set extended (PDSE), and the z/OS file system (zFS).
At the zHPF announcement, zHPF supported the transfer of small blocks of fixed size data
(4 K) from a single track. This capability was extended, first to 64 KB, and then to multi-track
operations. The 64 KB data transfer limit on multi-track operations was removed by z196. This
improvement allows the channel to fully use the bandwidth of FICON channels, which results
in higher throughputs and lower response times.
zHPF is enhanced to allow all large write operations (greater than 64 KB) at distances up to
100 km (62.13 miles) to be run in a single round trip to the control unit. This process does not
elongate the I/O service time for these write operations at extended distances. This
enhancement to zHPF removes a key inhibitor for customers that are adopting zHPF over
extended distances, especially when the IBM HyperSwap capability of z/OS is used.
From the z/OS perspective, the FICON architecture is called command mode and the zHPF
architecture is called transport mode. During link initialization, the channel node and the
control unit node indicate whether they support zHPF.
Requirement: All FICON channel path identifiers (CHPIDs) that are defined to the same
LCU must support zHPF. The inclusion of any non-compliant zHPF features in the path
group causes the entire path group to support command mode only.
The mode that is used for an I/O operation depends on the control unit that supports zHPF
and its settings in the z/OS operating system. For z/OS use, a parameter is available in the
IECIOSxx member of SYS1.PARMLIB (ZHPF=YES or NO) and in the SETIOS system command
to control whether zHPF is enabled or disabled. The default is ZHPF=NO.
Support also is added for the D IOS,ZHPF system command to indicate whether zHPF is
enabled, disabled, or not supported on the server.
Similar to the existing FICON channel architecture, the application or access method provides
the channel program (CCWs). How zHPF (transport mode) manages channel program
operations is different from the CCW operation for the existing FICON architecture (command
mode).
While in command mode, each CCW is sent to the control unit for execution. In transport
mode, multiple channel commands are packaged together and sent over the link to the
control unit in a single control block. Fewer processors are used compared to the existing
FICON architecture. Specific complex CCW chains are not supported by zHPF.
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258.
For more information about FICON channel performance, see the performance technical
papers that are available at the IBM Z I/O connectivity page of the IBM IT infrastructure
website.
The MIDAW facility is a system architecture and software feature that is designed to improve
FICON performance. This facility was first made available on IBM System z9® servers, and is
used by the Media Manager in z/OS.
The MIDAW facility provides a more efficient CCW/IDAW structure for certain categories of
data-chaining I/O operations.
MIDAW can improve channel use and I/O response time. It also reduces FICON channel
connect time, director ports, and control unit processor usage.
IBM laboratory tests indicate that applications that use EF data sets, such as Db2, or long
chains of small blocks can gain significant performance benefits by using the MIDAW facility.
MIDAW is supported on FICON channels that are configured as CHPID type FC. The
supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on page 258.
Figure 7-2 shows a single CCW that controls the transfer of data that spans noncontiguous
4 K frames in main storage. When the IDAW flag is set, the data address in the CCW points to
a list of words (IDAWs). Each IDAW contains an address that designates a data area within
real storage.
The number of required IDAWs for a CCW is determined by the following factors:
IDAW format as specified in the operation request block (ORB)
Count field of the CCW
Data address in the initial IDAW
10
Exceptions are made to this statement, and many details are omitted in this description. In this section, we
assume that you can merge this brief description with an understanding of I/O operations in a virtual memory
environment.
CCWs with data chaining can be used to process I/O data blocks that have a more complex
internal structure, in which portions of the data block are directed into separate buffer areas.
This process is sometimes known as scatter-read or scatter-write. However, as technology
evolves and link speed increases, data chaining techniques become less efficient because of
switch fabrics, control unit processing and exchanges, and other issues.
The MIDAW facility is a method of gathering and scattering data from and into discontinuous
storage locations during an I/O operation. The MIDAW format is shown in Figure 7-3. It is
16 bytes long and aligned on a quadword.
The combination of the address and count in a MIDAW cannot cross a page boundary.
Therefore, the largest possible count is 4 K. The maximum data count of all the MIDAWs in a
list cannot exceed 64 K, which is the maximum count of the associated CCW.
The scatter-read or scatter-write effect of the MIDAWs makes it possible to efficiently send
small control blocks that are embedded in a disk record to separate buffers from those that
are used for larger data areas within the record. MIDAW operations are on a single I/O block,
in the manner of data chaining. Do not confuse this operation with CCW command chaining.
VSAM and non-VSAM (DSORG=PS) sets can be defined as EF data sets. For non-VSAM
data sets, a 32-byte suffix is appended to the end of every physical record (that is, block) on
disk. VSAM appends the suffix to the end of every control interval (CI), which normally
corresponds to a physical record.
A 32 K CI is split into two records to span tracks. This suffix is used to improve data reliability,
and facilitates other functions that are described next. For example, if the DCB BLKSIZE or
VSAM CI size is equal to 8192, the block on storage consists of 8224 bytes. The control unit
does not distinguish between suffixes and user data. The suffix is transparent to the access
method and database.
EA is useful for creating large Db2 partitions (larger than 4 GB). Striping can be used to
increase sequential throughput, or to spread random I/Os across multiple logical volumes.
DFSMS striping is useful for the use of multiple channels in parallel for one data set. The Db2
logs are often striped to optimize the performance of Db2 sequential inserts.
Processing an I/O operation to an EF data set normally requires at least two CCWs with data
chaining. One CCW is used for the 32-byte suffix of the EF data set. With MIDAW, the extra
CCW for the EF data set suffix is eliminated.
MIDAWs benefit EF and non-EF data sets. For example, to read 12 4 K records from a
non-EF data set on a 3390 track, Media Manager chains together 12 CCWs by using data
chaining. To read 12 4 K records from an EF data set, 24 CCWs are chained (two CCWs per
4 K record). By using Media Manager track-level command operations and MIDAWs, an
entire track can be transferred by using a single CCW.
Performance benefits
z/OS Media Manager includes I/O channel program support for implementing EF data sets,
and automatically uses MIDAWs when appropriate. Most disk I/Os in the system are
generated by using Media Manager.
The MIDAW facility removes the 4 K boundary restrictions of IDAWs and for EF data sets,
which reduces the number of CCWs. Decreasing the number of CCWs helps to reduce the
FICON channel processor use. Media Manager and MIDAWs do not cause the bits to move
any faster across the FICON link. However, they reduce the number of frames and sequences
that flow across the link and use the channel resources more efficiently.
The performance of a specific workload can vary based on the conditions and hardware
configuration of the environment. IBM laboratory tests found that Db2 gains significant
performance benefits by using the MIDAW facility in the following areas:
Table scans
Logging
Utilities
Use of DFSMS striping for Db2 data sets
Media Manager with the MIDAW facility can provide significant performance benefits when
used in combination applications that use EF data sets (such as Db2) or long chains of small
blocks.
For more information about FICON and MIDAW, see the following resources:
The I/O Connectivity page of the IBM IT infrastructure website includes information about
FICON channel performance
DS8000 Performance Monitoring and Tuning, SG24-8013
ICKDSF
Device Support Facilities, ICKDSF, Release 17 is required on all systems that share disk
subsystems with an IBM z16 processor.
ICKDSF supports a modified format of the CPU information field that contains a two-digit
LPAR identifier. ICKDSF uses the CPU information field instead of CCW reserve/release for
concurrent media maintenance. It prevents multiple systems from running ICKDSF on the
same volume, and at the same time allows user applications to run while ICKDSF is
processing. To prevent data corruption, ICKDSF must determine all sharing systems that
might run ICKDSF. Therefore, this support is required for IBM z16.
Remember: The need for ICKDSF Release 17 also applies to systems that are not part of
the same sysplex, or are running an operating system other than z/OS, such as z/VM.
The zDAC function is integrated into the hardware configuration definition (HCD). Clients can
define a policy that can include preferences for availability and bandwidth that include parallel
access volume (PAV) definitions, control unit numbers, and device number ranges. When new
controllers are added to an I/O configuration or changes are made to existing controllers, the
system discovers them and proposes configuration changes that are based on that policy.
All added or changed logical control units and devices are added into the proposed
configuration. They are assigned proposed control unit and device numbers, and channel
paths that are based on the defined policy. zDAC uses channel path chosen algorithms to
minimize single points of failure. The zDAC proposed configurations are created as work I/O
definition files (IODFs) that can be converted to production IODFs and activated.
zDAC is designed to run discovery for all systems in a sysplex that support the function.
Therefore, zDAC helps to simplify I/O configuration on IBM z16 systems that run z/OS, and
reduces complexity and setup time.
zDAC applies to all FICON features that are supported on IBM z16 when configured as
CHPID type FC. The supported operating systems are listed in Table 7-6 on page 257 and
Table 7-7 on page 258.
Information about the channels that are connected to a fabric (if registered) allows other
nodes or storage area network (SAN) managers to query the name server to determine what
is connected to the fabric.
The following attributes are registered for the IBM z16 systems:
Platform information
Channel information
Worldwide port name (WWPN)
Port type (N_Port_ID)
FC-4 types that are supported
Classes of service that are supported by the channel
The platform and name server registration service are defined in the Fibre Channel Generic
Services 4 (FC-GS-4) standard.
The informal name, 63.75-K subchannels, represents 65280 subchannels, as shown in the
following equation:
63 x 1024 + 0.75 x 1024 = 65280
This equation is applicable for subchannel set 0. For subchannel sets 1, 2 and 3, the available
subchannels are derived by using the following equation:
(64 X 1024) -1=65535
Current z/VM versions MSS support for mirrored direct access storage device (DASD)
provides a subset of host support for the MSS facility to allow the use of an alternative
subchannel set for Peer-to-Peer Remote Copy (PPRC) secondary volumes.
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258. For more information about channel subsystem, see Chapter 5, “Central processor
complex channel subsystem” on page 197.
Subchannel sets
IBM z16 A01 supports four subchannel sets (SS0, SS1, SS2, and SS3).
Subchannel sets SS1, SS2, and SS3 can be used for disk alias devices of primary and
secondary devices, and as Metro Mirror secondary devices. This set helps facilitate storage
growth and complements other functions, such as extended address volume (EAV) and
Hyper Parallel Access Volumes (HyperPAV).
See Table 7-6 on page 257 and Table 7-7 on page 258 for list of supported operating
systems.
See Table 7-6 on page 257 and Table 7-7 on page 258 for list of supported operating
systems. For more information, see “IPL from an alternative subchannel set” on page 295.
32 K subchannels
To help facilitate growth and continue to enable server consolidation, the IBM z16 supports up
to 32 K subchannels per FICON Express32S, FICON Express16SA, and FICON
Express16S+ channels (CHPID). More devices can be defined per FICON channel, which
includes primary, secondary, and alias devices. The maximum number of subchannels across
all device types that are addressable within an LPAR remains at 63.75 K for subchannel set 0
and 64 K (64 X 1024)-1 for subchannel sets 1, 2, and 3.
This support is available to the IBM z16, IBM z15, z14, IBM z13, and IBM z13s servers and
applies to the FICON Express32S, FICON Express16SA, and FICON Express16S+ features
(defined as CHPID type FC).
The supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on
page 258.
No action is required on z/OS to enable the health check; it is automatically enabled at IPL
and reacts to changes that might cause problems. The health check can be disabled by using
the PARMLIB or SDSF modify commands.
The supported operating systems are listed in Table 7-6 on page 257. For more information
about FICON Dynamic Routing (FIDR), see Chapter 4, “Central processor complex I/O
structure” on page 145.
The supported operating systems are listed in Table 7-6 on page 257.
For more information about FCP channel performance, see the performance technical papers
that are available at the IBM Z I/O connectivity page of the IBM IT infrastructure website.
The FCP protocol is supported by z/VM, z/VSE, and Linux on IBM Z. The supported
operating systems are listed in Table 7-6 on page 257 and Table 7-7 on page 258.
T10-DIF support
American National Standards Institute (ANSI) T10 Data Integrity Field (DIF) standard is
supported on IBM Z for SCSI end-to-end data protection on fixed block (FB) LUN volumes.
IBM Z provides added end-to-end data protection between the operating system and the
DS8870 unit. This support adds protection information that consists of Cyclic Redundancy
Checking (CRC), Logical Block Address (LBA), and host application tags to each sector of FB
data on a logical volume.
IBM Z support applies to FCP channels only. The supported operating systems are listed in
Table 7-6 on page 257 and Table 7-7 on page 258.
N_Port ID Virtualization
N_Port ID Virtualization (NPIV) allows multiple system images (in LPARs or z/VM guests) to
use a single FCP channel as though each were the sole user of the channel. First introduced
with z9 EC, this feature can be used with supported FICON features on IBM z16 servers. The
supported operating systems are listed in Table 7-6 on page 257 and Table 7-7 on page 258.
The capabilities of the WWPN are extended to calculate and show WWPNs for virtual and
physical ports ahead of system installation.
The tool assigns WWPNs to each virtual FCP channel or port by using the same WWPN
assignment algorithms that a system uses when assigning WWPNs for channels that use
NPIV. Therefore, the SAN can be set up in advance, which allows operations to proceed
much faster after the server is installed. In addition, the SAN configuration can be retained
instead of altered by assigning the WWPN to physical FCP ports when a FICON feature is
replaced.
The WWPN tool takes a .csv file that contains the FCP-specific I/O device definitions and
creates the WWPN assignments that are required to set up the SAN. A binary configuration
file that can be imported later by the system is also created. The .csv file can be created
manually or exported from the HCD/HCM. The supported operating systems are listed in
Table 7-6 on page 257 and Table 7-7 on page 258.
The WWPN tool is applicable to all FICON channels that are defined as CHPID type FCP (for
communication with SCSI devices) on IBM z16. It is available for download at the Resource
Link at the following website (log in is required).
Note: An optional feature can be ordered for WWPN persistency before shipment to keep
the same I/O serial number on the new CPC. Current information must be provided during
the ordering process.
The 25 GbE RoCE Express3 are native PCIe features that not use a CHPID and are defined
by using the IOCP FUNCTION statement or in the hardware configuration definition (HCD).
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID).
The 25 GbE RoCE Express3 SR (FC 0452) feature connects by using SR optics and
multimode fiber that is terminated with LC connector. The 25 GbE RoCE Express3 LR (FC
0453) connects by using LR optics and single mode fiber that is terminated with LC
connector.
Support for select Linux on IBM Z distributions is provided for Shared Memory
Communications over Remote Direct Memory Access (SMC-R) by using RoCE Express
features. For more information, see this Linux on IBM Z Blogspot web page.
The 10 GbE RoCE Express3 are native PCIe features that do not use a CHPID and are
defined by using the IOCP FUNCTION statement or in the HCD.
The virtualization capabilities for IBM z16 are 63 Virtual Functions per port (126 VFs per
feature/PCHID). One RoCE Express feature can be shared by up to 126 partitions (LPARs)
(one adapter is one PCHID).
The 25 GbE RoCE Express3 SR (FC 0452) feature connects by using SR optics and
multimode fiber that is terminated with LC connector. The 25 GbE RoCE Express3 LR (FC
0453) connects by using LR optics and single mode fiber that is terminated with LC
connector.
The 25 GbE RoCE Express2 has one PCHID and the same virtualization characteristics and
the 10 GbE RoCE Express2 (FC 0412 and FC 0432); that is, 126 Virtual Functions per
PCHID.
z/OS requires fixes for APAR OA55686. RMF 2.2 and later also is enhanced to recognize the
CX4 card type and properly display CX4 cards in the PCIe Activity reports.
25 GbE RoCE Express2 feature also are used by Linux on IBM Z for applications that are
coded to the native RoCE verb interface or use Ethernet (such as TCP/IP). This native
exploitation does not require a peer OSA.
Support for select Linux on IBM Z distributions is now provided for Shared Memory
Communications over Remote Direct Memory Access (SMC-R) by using RoCE Express
features. For more information, see this Linux on IBM Z Blogspot web page.
The RoCE Express3 features also can provide LAN connectivity for Linux on IBM Z, and
comply with IEEE standards. In addition, RoCE Express features assume several functions of
the TCP/IP stack that normally are performed by the PU, which allows significant
performance benefits by offloading processing from the operating system.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
z/OS Communications Server (CS) provides a new software device driver ConnectX4 (CX4)
for RoCE Express2. The device driver is not apparent to the upper layers of the CS (the
SMC-R and TCP/IP stack) and the application software (that uses TCP sockets). RoCE
Express2 introduces a minor change in how the physical port is configured.
RMF 2.2 and later also is enhanced to recognize the new CX4 card type and correctly display
CX4 cards in the PCIe Activity reports.
Support in select Linux on IBM Z distributions is now provided for Shared Memory
Communications over Remote Direct Memory Access (SMC-R) by using the supported RoCE
Express features. For more information, see this Linux on IBM Z Blogspot web page.
The RoCE Express3 features also can provide LAN connectivity for Linux on IBM Z, and
comply with IEEE standards. In addition, RoCE Express features assume several functions of
the TCP/IP stack that normally are performed by the PU, which allows significant
performance benefits by offloading processing from the operating system.
The 10 GbE RoCE Express2 feature also is used by Linux on IBM Z for applications that are
coded to the native RoCE verb interface or use Ethernet (such as TCP/IP). This native use
does not require a peer OSA.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Similar to SMC-R, this protocol uses shared memory architectural concepts that eliminate
TCP/IP processing in the data path, yet preserve TCP/IP Qualities of Service for connection
management purposes.
Support in select Linux on IBM Z distributions is now provided for Shared Memory
Communications over Direct Memory Access (SMC-D). For more information, see this Linux
on IBM Z Blogspot web page.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Because the initial version of SMC was limited to TCP/IP connections over the same layer 2
network, it was not routable across multiple IP subnets. The associated TCP/IP connection
was limited to hosts within a single IP subnet that required the hosts to have direct access to
the same physical layer 2 network (that is, the same Ethernet LAN over a single VLAN ID).
The scope of eligible TCP/IP connections for SMC was limited to and defined by the single IP
subnet.
SMC Version 2 (SMCv2) provides support for SMC over multiple IP subnets for both SMC-D
and SMC-R and is referred to as SMC-Dv2 and SMC-Rv2. SMCv2 requires updates to the
underlying network technology. SMC-Dv2 requires ISMv2 and SMC-Rv2 requires RoCEv2.
The SMCv2 protocol is downward compatible, which allows SMCv2 hosts to continue to
communicate with SMCv1 previous hosts.
Although SMCv2 changes the SMC connection protocol to enable multiple IP subnet support,
SMCv2 does not change how user TCP socket data is transferred, which preserves the
benefits of SMC to TCP workloads.
TCP/IP connections that require IPsec are not eligible for any form of SMC.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Layer 2 support can help facilitate server consolidation. Complexity can be reduced, network
configuration is simplified and intuitive, and LAN administrators can configure and maintain
the mainframe environment the same way as they do a nonmainframe environment.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Linux on IBM Z tools can be used to format, edit, and process the trace records for analysis
by system programmers and network administrators.
OSA-Express7S 1.2 25 GbE SR (FC 0459) and OSA-Express7S 1.2 25 GbE LR (FC 0460)
are installed in the PCIe+ I/O Drawer and have 25 GbE physical port. New with the generation
is the Long Reach version, which uses single mode fiber and can be point to point connected
to a distance of up to 10 km (6.2 miles). The features connect to a 25 GbE switch and do not
support auto-negotiation to a different speed.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
OSA-Express7S 1.2 10 GbE SR (FC 0457) and OSA-Express7S 1.2 10 GbE LR (FC 0456)
are installed in the PCIe+ I/O Drawer and have 10 GbE physical port. The features connect to
a 10 GbE switch and do not support auto-negotiation to a different speed.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
OSA-Express7S 1.2 GbE SX (FC 0455) and OSA-Express7S 1.2 10 GbE LX (FC 0454) are
installed in the PCIe+ I/O Drawer and have two GbE physical ports. The features connect to a
GbE switch and do not support auto-negotiation to a different speed.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express6S 1000BASE-T Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express6S 1000BASE-T Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Note: Operating system support is required to recognize and use the second port on the
OSA-Express6S 1000BASE-T Ethernet feature.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Note: OSA-ICC supports up to 48 secure sessions per CHPID (the overall maximum of
120 connections is unchanged).
OSA-ICC Enhancements
With HMC 2.14.1 and newer the following enhancements are available:
The IPv6 communications protocol is supported by OSA-ICC 3270 so that clients can
comply with regulations that require all computer purchases to support IPv6.
TLS negotiation levels (the supported TLS protocol levels) for the OSA-ICC 3270 client
connection can now be specified:
– TLS 1.0 OSA-ICC 3270 server permits TLS 1.0, TLS 1.1, and TLS 1.2 client
connections.
– TLS 1.1 OSA-ICC 3270 server permits TLS 1.1 and TLS 1.2 client connections.
– TLS 1.2 OSA-ICC 3270 server permits only TLS 1.2 client connections.
Separate and unique OSA-ICC 3270 certificates are supported (for each PCHID) for the
benefit of customers who host workloads across multiple business units or data centers
where cross-site coordination is required. Customers can avoid interruption of all the TLS
connections at the same time when having to renew expired certificates. OSA-ICC also
continues to support a single certificate for all OSA-ICC PCHIDs in the system.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Checksum offload provides checksum offload for several types of traffic and is supported by
the following features when configured as CHPID type OSD (QDIO mode only):
OSA-Express7S 1.2, OSA_Express7S 1.1, and OSA-Express7S 25 GbE
OSA-Express7S and OSA-Express7S 1.2 10 GbE
OSA-Express7S and OSA-Express7S 1.2 GbE
OSA-Express7S and OSA-Express7S 1.2 1000BASE-T Ethernet
OSA-Express6S 10 GbE
OSA-Express6S GbE
OSA-Express6S 1000BASE-T Ethernet
When multiple IP stacks share an OSA-Express, and an IP stack sends a packet to a next
hop address that is owned by another IP stack that is sharing the OSA-Express,
OSA-Express sends the IP packet directly to the other IP stack. The packet does not have to
be placed out on the LAN, which is termed LPAR-to-LPAR traffic. Checksum offload is
enhanced to support the LPAR-to-LPAR traffic, which was not originally available.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The use of display OSAINFO (z/OS) or NETSTAT OSAINFO (z/VM) allows the operator to monitor
and verify the current OSA configuration and helps improve the overall management,
serviceability, and usability of OSA-Express cards.
These commands apply to CHPID type OSD. The supported operating systems are listed in
Table 7-8 on page 259.
QDIO data connection isolation allows disabling internal routing for each QDIO connected. It
also provides a means for creating security zones and preventing network traffic between the
zones.
QDIO data connection isolation is supported by all OSA-Express features on IBM z16. The
supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on page 261.
QDIO interface isolation is supported on all OSA-Express features on IBM z16. The
supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on page 261.
The supported operating systems are listed in Table 7-8 on page 259.
Each input queue is a unique type of workload, and has unique service and processing
requirements. The IWQ function allows z/OS to preassign the appropriate processing
resources for each input queue. This approach allows multiple concurrent z/OS processing
threads to process each unique input queue (workload), which avoids traditional resource
contention.
IWQ reduces the conventional z/OS processing that is required to identify and separate
unique workloads. This advantage results in improved overall system performance and
scalability.
11 Only OSA-Express6S and OSA-Express7S cards are supported on IBM z16 as carry forward.
The following types of z/OS workloads are identified and assigned to unique input queues:
z/OS Sysplex Distributor traffic
Network traffic that is associated with a distributed virtual Internet Protocol address (VIPA)
is assigned to a unique input queue. This configuration allows the Sysplex Distributor
traffic to be immediately distributed to the target host.
z/OS bulk data traffic
Network traffic that is dynamically associated with a streaming (bulk data) TCP connection
is assigned to a unique input queue. This configuration allows the bulk data processing to
be assigned the suitable resources and isolated from critical interactive workloads.
EE (Enterprise Extender / SNA traffic)
IWQ for the OSA-Express features is enhanced to differentiate and separate inbound
Enterprise Extender traffic to a dedicated input queue.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
Link aggregation (trunking) combines multiple physical OSA-Express ports into a single
logical link. This configuration increases throughput, and provides nondisruptive failover if a
port becomes unavailable. The target links for aggregation must be of the same type.
Link aggregation is applicable to CHPID type OSD (QDIO). The supported operating systems
are listed in Table 7-8 on page 259 and Table 7-9 on page 261.
Large send support for IPv6 packets applies to the OSA-Express7S 1.2, OSA-Express7S,
and OSA-Express6S11 features (CHPID type OSD) on IBM z16, IBM z15, and IBM z14.
OSA-Express6S added TCP checksum on large send, which reduces the cost (CPU time) of
error detection for large send.
The supported operating systems are listed in Table 7-8 on page 259 and Table 7-9 on
page 261.
In all cases, the TCP/IP stack determines the best setting that is based on the current system
and environmental conditions, such as inbound workload volume, processor use, and traffic
patterns. It can then dynamically update the settings.
Supported OSA-Express features adapt to the changes, which avoids thrashing and frequent
updates to the OAT. Based on the TCP/IP settings, OSA holds the packets before presenting
them to the host. A dynamic setting is designed to avoid or minimize host interrupts.
OSA Layer 3 VMAC is supported by all OSA-Express features on IBM z16 when in QDIO
mode (CHPID type OSD). The supported operating systems are listed in Table 7-8 on
page 259.
The Network Traffic Analyzer is supported by all OSA-Express features on IBM z16 when in
QDIO mode (CHPID type OSD). The supported operating systems are listed in Table 7-8 on
page 259.
The minimum software support levels are described in the following sections. Review the
current PSP buckets to ensure that the latest support levels are known and included as part
of the implementation plan.
Database administrators can use z/OS Dataset Encryption, z/OS Coupling Facility
Encryption, z/VM encrypted hypervisor paging, and z/TPF transparent database encryption,
which use the performance enhancements in the hardware.
In addition, the IBM z16 and IBM z15 cores implement a Modulo Arithmetic unit in support of
Elliptic Curve Cryptography.
CPACF is used by several IBM software product offerings for z/OS, such as IBM WebSphere
Application Server for z/OS. For more information, see 6.4, “CP Assist for Cryptographic
Functions” on page 219.
The supported operating systems are listed in Table 7-10 on page 263 and Table 7-11 on
page 264.
Support of Crypto Express8S functions varies by operating system and release and by the
way that the card is configured as a coprocessor or an accelerator. The supported operating
systems are listed in Table 7-10 on page 263 and Table 7-11 on page 264.
The supported operating systems are listed in Table 7-10 on page 263 and Table 7-11 on
page 264.
Web deliverables
For more information about web-deliverable code on z/OS, see the z/OS downloads website.
For Linux on IBM Z, support is delivered through IBM and the distribution partners. For more
information, see Linux on IBM Z on the IBM Developer website.
IBM categorized the following security functions according to International Organization for
Standardization (ISO) standard 7498-2:
Identification and authentication: Includes the ability to identify users to the system and
provide proof that they are who they claim to be.
Access control: Determines which users can access which resources.
Data confidentiality: Protects an organization’s sensitive data from being disclosed to
unauthorized persons.
Data integrity: Ensures that data is in its original form and that nothing altered it.
Security management: Administers, controls, and reviews a business security policy.
Nonrepudiation: Assures that the suitable individual sent the message.
Only cryptographic services can provide the data confidentiality and the identity
authentication that is required to protect business commerce on the internet13.
ICSF support for IBM z16 is provided with PTFs, not as previously was the case, through Web
deliverables.
Supported levels of ICSF automatically detect what hardware cryptographic capabilities are
available where it is running. Then, it enables functions accordingly. No toleration of new
hardware is necessary because it is “just there”. ICSF maintenance is necessary if you want
to use new capabilities.
Use of new functions is supplied in ICSF PTFs on z/OS V2.R2-V2.R4 (Web deliverable
HCR77D1) or V2.R5 (base, which is HCR77D2).
For more information about ICSF versions and FMID cross-references, see this IBM Support
web page.
Reporting can be done at an LPAR/domain level to provide more granular reports for capacity
planning and diagnosing problems. This feature requires fix for APAR OA54952.
The supported operating systems are listed in Table 7-10 on page 263.
Policy-driven z/OS Data Set Encryption enables users to perform the following tasks:
De-couple encryption from data classification; encrypt data automatically independent of
labor-intensive data classification work.
Encrypt data immediately and efficiently at the time that it is written.
Reduce risks that are associated with mis-classified or undiscovered sensitive data.
Help protect digital assets automatically.
Achieve application transparent encryption.
IBM Db2 for z/OS and IBM Information Management System (IMS) intend to use z/OS Data
Set Encryption.
With z/OS, Data Set Encryption DFSMS enhances data security with support for data set
level encryption by using DFSMS access methods. This function is designed to give users the
ability to encrypt their data sets without changing their application programs.
DFSMS users can identify which data sets require encryption by using JCL, Data Class, or
the RACF data set profile. Data set level encryption can allow the data to remain encrypted
during functions, such as backup and restore, migration and recall, and replication.
z/OS Data Set Encryption requires CP Assist for Cryptographic Functions (CPACF).
Considering the significant enhancements that were introduced with z14, the encryption
mode of XTS is used by access method encryption to obtain the best performance possible. It
is not recommended to enable z/OS data set encryption until all sharing systems, fallback,
backup, and DR systems support encryption.
In addition to applying PTFs enabling the support, ICSF configuration is required. The
supported operating systems are listed in Table 7-10 on page 263.
The following Quantum-safe enhancements were introduced with IBM z16 to accomplish this
encryption:
Key generation
Hybrid key exchange schemes
Dual digital signature schemes
Included in this support is the ability to dynamically control whether a running z/VM system is
encrypting this data. This support protects guest paging data from administrators or users
with access to volumes. Enabled with AES encryption, z/VM Encrypted Paging includes low
overhead by using CPACF.
The supported operating systems are listed in Table 7-10 on page 263.
Because of the potential costs and overhead, most of the organizations avoid the use of
host-based network encryption today. By using enhanced CPACF and SIMD on IBM z16, TLS
and IPsec can use hardware performance gains while benefiting from transparent
enablement. Reduced cost of encryption enables broad use of network encryption.
A new z/OSMF Compliance fact collection REST API sends an ENF86 signal to all systems.
Participating products and components collect and write compliance data to new SMF 1154
records that are associated with its unique subtype. These new SMF 1154 records can be
integrated into solutions, such as the IBM z16 IBM Z Security and Compliance Center.
This support requires PTFs for z/OS 2.4 and z/OS 2.5. The PTFs are identified by a fix
category that is designated specifically for Compliance data collection support named
IBM.Function.Compliance.DataCollection. For more information about how to use this fix
category to identify and install the specific PTFs that enable compliance data collection, see
“IBM Fix Category Values and Descriptions”.
The following prerequisite operating system versions are supported for the collection of
compliance data:
Red Hat Enterprise Linux 8.0 (RHEL) on IBM Z, or later
SUSE Linux Enterprise Server (SLES) on IBM Z 15
Ubuntu Server LTS for IBM Z 22.04
Although IBM z16 servers do not require any “functional” software, it is recommended to
install all IBM z16 service before upgrading to the new server. The support matrix for z/OS
releases and the IBM Z servers that support them are listed in Table 7-16, where “X” indicates
that the hardware model is supported.
New: ICSF support for IBM z16 is provided with PTFs, not Web deliverables
14
For example, the use of Crypto Express7S requires the Cryptographic Support for z/OS V2R2 - z/OS V2R3 web
deliverable.
The use of many functions covers fixes that are required to use the capabilities of the IBM z16
servers. They are identified by:
IBM.Device.Server.IBM z16-3931.Exploitation
General z/OS support documentation provided in the PSP Bucket. For IBM z16 A01: PSP
Bucket Upgrade = 3931DEVICE, Subset = 3931/ZOS.
Use the SMP/E REPORT MISSINGFIX command to determine whether any FIXCAT APARs exist
that are applicable and are not yet installed, and whether any SYSMODs are available to
satisfy the missing FIXCAT APARs.
For more information about IBM Fix Category Values and Descriptions, see this IBM Support
web page.
15
For more information, see the Tool to Compare IBM z16 Instruction Mnemonics with Macro Libraries IBM
Technote.
For more information about this release, see this announcement letter.
Consider the following points before migrating z/OS V2.R3 to IBM z16:
IBM z/OS V2.R3 with IBM z16 requires a minimum of 8 GB of memory. When running as a
z/VM guest or on an IBM System z® Personal Development Tool, a minimum of 2 GB is
required for z/OS V2.R3. If the minimum is not met, a warning WTOR is issued at IPL.
Continuing with less than the minimum memory might affect availability. A migration health
check warns if the system is configured with less than 8 GB.
Dynamic splitting and merging of Coordinated Timing Network (CTN) is available with IBM
z16.
RMF support is provided to collect SMC-D related performance measurements in SMF 73
Channel Path Activity and SMF 74 subtype 9 PCIe Activity records. It also provides these
measurements in the RMF Postprocessor and Monitor III PCIe and Channel Activity
reports.
Configurations with a Coupling Facility on one of these servers can add an IBM z16 Server to
their Sysplex for a z/OS or a Coupling Facility image. IBM z16 does not support participating
in a Parallel Sysplex with System IBM z13/IBM z13s and earlier systems.
Coupling connectivity is available only when other systems also support the same type of
coupling. For more information about supported coupling link technologies on IBM z16, see
4.6.4, “Parallel Sysplex connectivity” on page 187, and the Coupling Facility Configuration
Options white paper.
Note: IBM z16 does not support InfiniBand coupling links. More planning might be required
to integrate the IBM z16 in a Parallel Sysplex with IBM IBM z14servers.
The following new functions provide performance improvements for applications by using new
IBM z16 instructions:
High-performance math libraries
MASS
Replace Atlas with OpenBLAS
Metal C for modernizing HLASM applications and systems programming
To enable the use of new functions, specify ARCH(14) and VECTOR for compilation. The binaries
that are produced by the compiler on IBM z16 can be run only on IBM z16 and above
because it uses the vector facility on IBM z16 for new functions. The use of older versions of
the compiler on IBM z16 does not enable new functions.
z/OS V2R4 can use the latest level (14) of the following C/C++ compiler options:
ARCHITECTURE: This option selects the minimum level of system architecture on which the
program can run. Specific features that are provided by the compiler require a minimum
architecture level. ARCH(14) uses instructions that are available on the IBM z16.
TUNE: This option allows optimization of the application for a specific system architecture
within the constraints that are imposed by the ARCHITECTURE option. The TUNE level must
not be lower than the setting in the ARCHITECTURE option.
Important: Use the previous IBM Z ARCHITECTURE or TUNE options for C/C++ programs if
the same applications run on the previous IBM Z servers. However, if C/C++ applications
run on IBM z16 servers only, use the latest ARCHITECTURE and TUNE options to ensure that
the best performance possible is delivered through the latest instruction set additions.
For more information about the ARCHITECTURE, TUNE, and VECTOR compiler options, see z/OS
XL C/C++ User’s Guide, SC14-7307-40.
z/OS XL C/C++ Web deliverables are available at no charge to z/OS XL C/C++ customers:
Based on open-source LLVM infrastructure; supports up to date C++ language standards
64-bit, UNIX System Services only
Statement of Direction: IBM will continue to adopt the LLVM and Clang compiler
infrastructure in future C/C++ offerings on IBM Za.
a. Any statements regarding IBM's future direction, intent, or product plans are subject to change
or withdrawal without notice.
Consider the following general guidelines when you are migrating z/VSE environment to IBM
z16 servers:
Collect reference information before migration
This information includes baseline data that reflects the status of, for example,
performance data, CPU use of reference workload, I/O activity, and elapsed times.
This information is required to size IBM z16 and is the only way to compare workload
characteristics after migration.
For more information, see z/VSE Release and Hardware Upgrade.
Apply required maintenance for IBM z16
Review the Preventive Service Planning (PSP) bucket 8561DEVICE for IBM z16 and apply
the required PTFs for IBM and independent software vendor (ISV) products.
This section provides an overview of the following IBM Z software licensing options that are
available for IBM z16 software, including MLC, zIPLA, subcapacity, sysplex, and Taylor Fit
Pricing:
Monthly license charge (MLC)
MLC is a recurring charge that is applied monthly. It includes the right to use the product
and provides access to IBM product support during the support period. Select an MLC
pricing metric that is based on your goals and environment.
The selected metric is used to price MLC products, such as z/OS, z/TPF, z/VSE,
middleware, compilers, and selected systems management tools and utilities:
– Key MLC metrics and offerings
MLC metrics include various offerings. The metrics and pricing schemes that are
available on IBM z14, IBM z15, and IBM z16 are listed in Table 7-17 on page 324.
zIPLA licensing
International Program License Agreement (IPLA) programs include a one-time charge
(OTC) and an optional annual maintenance charge, called Subscription and Support. This
annual charge includes access to IBM technical support and enables you to obtain version
upgrades at no charge for products that generally fall under the zIPLA such as application
development tools, CICS tools, data management tools, WebSphere for IBM Z products,
Linux on IBM Z middleware and z/VM.
The following pricing metrics apply to IBM Z IPLA products:
– Value Unit pricing applies to most IPLA products that run on z/OS. Value Unit pricing is
typically based on a number of MSUs and allows for lower cost of incremental growth.
– z/VM V5, V6, V7, and specific z/VM middleware include pricing that is based on the
number of engines. Engine-based Value Unit pricing allows for a lower cost of
incremental growth with more engine-based licenses that are purchased.
– Most Linux middleware also is priced based on the number of engines. The number of
engines is converted into Processor Value Units under the IBM Passport Advantage®
terms and conditions.
For more information, see this web page.
Subcapacity licensing
Subcapacity licensing includes software charges for specific IBM products that are based
on the use capacity of the logical partitions (LPARs) on which the product runs.
Subcapacity licensing removes the dependency between the software charges and CPC
(hardware) installed capacity.
The subcapacity licensed products are charged monthly based on the highest observed
4-hour rolling average use of the logical partitions in which the product runs.
Technology Update Pricing for the IBM z16 extends the software price and performance that
is provided by AWLC for IBM z16 servers. The new and revised Transition Charges for
Sysplexes offerings provide a transition to Technology Update Pricing for the IBM z16 for
customers who have not fully migrated to IBM z16 servers. This transition ensures that
aggregation benefits are maintained and phases in the benefits of Technology Update Pricing
for the IBM z16 pricing as customers migrate.
When an IBM z16 server is in an actively coupled Parallel Sysplex or a Loosely Coupled
Complex, you might choose aggregated Advanced Workload License Charges (AWLC)
pricing or aggregated Parallel Sysplex License Charges (PSLC) pricing (subject to all
applicable terms and conditions).
When an IBM z16 server is part of a Multiplex under Country Multiplex Pricing (CMP) terms,
Country Multiplex License Charges (CMLC), Multiplex zNALC (MzNALC), and Flat Workload
License Charges (FWLC) are the only pricing metrics that are available (subject to all
applicable terms and conditions).
When an IBM z16 server is running z/VSE, you can choose Mid-Range Workload License
Charges (MWLC), which are subject to all applicable terms and conditions.
7.9 References
For more information about planning, see the home pages for the following operating
systems:
z/OS
z/VM
z/VSE
z/TPF
Linux on IBM Z
KVM for IBM Z
IBM z16 servers support dynamic provisioning features to give clients exceptional flexibility
and control over system capacity and costs.
8.2.1 Overview
Upgrades can be categorized as described in this section.
Tip: An MES provides system upgrades that can result in more enabled processors, a
different central processor (CP) capacity level, more processor drawers, memory,
PCIe+ I/O drawers, and I/O features (physical upgrade). Extra planning tasks are
required for nondisruptive logical upgrades. An MES is ordered through your IBM
representative and installed by IBM service support representatives (IBM SSRs).
System Recovery Boost zIIP capacity is a pre-paid offering that is available on IBM z15
T01 and IBM z16 A01. It is intended to provide temporary zIIP capacity to be used to boost
CPU performance for boost events. For more information, see Introducing IBM X System
Recovery Boost, REDP-5563.
Flexible Capacity for Cyber Resiliency is new type of temporary record that is introduced
with IBM z16. This record holds the Flexible Capacity Entitlements for IBM z16 machines
across two or more sites.
Activated capacity Capacity that is purchased and activated. Purchased capacity can be greater than the activated
capacity.
Billable capacity Capacity that helps handle workload peaks (expected or unexpected). The one billable offering
that is available is On/Off CoD.
Capacity Hardware resources (processor and memory) that can process the workload can be added to
the system through various capacity offerings.
Capacity Backup This capacity allows you to place model capacity or specialty engines in a backup system. CBU
(CBU) is used in an unforeseen loss of system capacity because of an emergency or for Disaster
Recovery testing.
Capacity for Planned Used when temporary replacement capacity is needed for a short-term event. CPE activates
Event (CPE) processor capacity temporarily to facilitate moving systems between data centers, upgrades,
and other routine management tasks. CPE is an offering of CoD.
Capacity levels Can be full capacity or sub-capacity. For an IBM z16 A01 system, capacity levels for the CP
engine are 7, 6, 5, and 4.
Capacity setting Derived from the capacity level and the number of processors.
For the IBM z16 A01 system, the capacity levels are 7nn, 6yy, 5yy, and 4xx, where xx, yy, or
nn indicates the number of active CPs.
Customer Initiated A web-based facility where you can request processor and memory upgrades by using the IBM
Upgrade (CIU) Resource Link and the system’s Remote Support Facility (RSF) connection.
Capacity on Demand The ability of a system to increase or decrease its performance capacity as needed to meet
(CoD) fluctuations in demand.
Capacity As a component of z/OS Capacity Provisioning, CPM monitors business-critical workloads that
Provisioning are running z/OS on IBM z16.
Manager (CPM)
Customer profile This information is on Resource Link and contains client and system information. A customer
profile can contain information about systems that are related to their IBM customer numbers.
Flexible Capacity for Available on IBM z16 A01 servers, the optional Flexible Capacity Record is an orderable
Cyber Resiliency feature that entitles a customer to active MIPS flexibility for all engine types between IBM z16
servers across two or more sites. It allows capacity swaps for an extended term.
Full capacity CP For IBM z16 servers, feature (CP7) provides full capacity. Capacity settings 7nn are full
feature capacity settings with the ranges of 1 - 99 in decimal and A0 - K0, where A0 represents 100 and
K0 represents 200, for capacity level 7nn.
Installed record The LICCC record is downloaded, staged to the Support Element (SE), and is installed on the
central processor complex (CPC). A maximum of eight different records can be concurrently
installed.
Model capacity Shows the current active capacity on the system, including all replacement and billable capacity.
identifier (MCI) For IBM z16 A01 servers, the model capacity identifier is in the form of 4xx, 5yy, 6yy, or 7nn,
where xx, yy, or nn indicates the number of active CPs:
xx can have a range of 00 - 39. An all IFL or an all ICF system has a capacity level of 400.
yy can have a range of 01 - 39.
1 - 99 in decimal and A0 - K0, where A0 represents 100 and K0 represents 200, for
capacity level 7nn.
Model Permanent Keeps information about the capacity settings that are active before any temporary capacity is
Capacity Identifier activated.
(MPCI)
Model Temporary Reflects the permanent capacity with billable capacity only, without replacement capacity. If no
Capacity Identifier billable temporary capacity is active, MTCI equals the MPCI.
(MTCI)
On/Off Capacity on Represents a function that allows spare capacity in a CPC to be made available to increase the
Demand total capacity of a CPC. For example, On/Off CoD can be used to acquire capacity for handling
a workload peak.
Permanent capacity The capacity that a client purchases and activates. This amount might be less capacity than the
total capacity purchased.
Permanent upgrade LICC that is licensed by IBM to enable the activation of applicable computing resources, such
as processors or memory, for a specific CIU-eligible system on a permanent basis.
Purchased capacity Capacity that is delivered to and owned by the client. It can be higher than the permanent
capacity.
Permanent/ The internal representation of a temporary (TER) or permanent (PER) capacity upgrade that is
Temporary processed by the CIU facility. An entitlement record contains the encrypted representation of
entitlement record the upgrade configuration with the associated time limit conditions.
Replacement A temporary capacity that is used for situations in which processing capacity in other parts of
capacity the enterprise is lost. This loss can be a planned event or an unexpected disaster. The two
replacement offerings available are Capacity for Planned Events and Capacity Backup.
Resource Link The IBM Resource Link website a technical support website that provides a comprehensive set
of tools and resources (log in required).
Secondary approval An option that is selected by the client that requires second approver control for each CoD order.
When a secondary approval is required, the request is sent for approval or cancellation to the
Resource Link secondary user ID.
Staged record The point when a record that represents a temporary or permanent capacity upgrade is
retrieved and loaded on the SE disk.
Subcapacity For IBM z16 A01 servers, CP features (CP4, CP5, and CP6) provide reduced capacity relative
to the full capacity CP feature (CP7).
System Recovery Available on IBM z16 A01 servers, the optional System Recovery Boost Upgrade is an
Boost Upgrade orderable feature that provides more capacity for a limited time to enable speeding up
record shutdown, restart, and catchup processing for a limited event duration.
Temporary capacity An optional capacity that is added to the current system capacity for a limited amount of time. It
can be capacity that is owned or not owned by the client.
Vital product data Information that uniquely defines system, hardware, software, and microcode elements of a
(VPD) processing system.
Tip: The use of the CIU facility for a system requires that the online CoD buying feature
(FC 9900) is installed on the system. The CIU facility is enabled through the permanent
upgrade authorization feature code (FC 9898).
Considerations: Most of the MESs can be concurrently applied without disrupting the
workload. For more information, see 8.3, “Concurrent upgrades” on page 334. However,
specific MES changes might be disruptive, such as adding PCIe IO drawers.
memory
nondisruptively if multiple CPC drawers are available and the
option is used.
flexible
Memory upgrades that require dual inline memory module (DIMM) changes can be made
Prepaid On/Off CoD tokens: Beginning with IBM z16, new prepaid On/Off CoD tokens
that are purchased do not carry forward to future systems.
CBU
This offering allows you to replace model capacity or specialty engines in a backup system
that is used in an unforeseen loss of system capacity because of a disaster.
CPE1
This offering allows you to replace model capacity or specialty engines because of a
relocation of workload during system migrations or a data center move.
System Recovery Boost Upgrade record
This offering allows you to add up to 20 zIIPs for use with the System Recovery Boost
facility. System Recovery Boost provides temporary extra capacity for CP workloads to
allow rapid shutdown, restart, and recovery of eligible systems. System Recovery Boost
records are prepaid, licensed for 1 - 5 years, and can be renewed at any time.
Flexible Capacity Record
This offering allows you to move CPU capacity between machines across two or more
sites. Capacity can be moved between sites a maximum of 12 times per year for a
maximum of 12 months per move.
Billable capacity
To handle a peak workload, you can activate up to double the purchased capacity of any
processor unit (PU) type temporarily. You are charged daily.
This capability is based on the flexibility of the design and structure, which allows concurrent
hardware installation and Licensed Internal Control Code (LICC) configuration changes.
The sub-capacity models allow more configuration granularity within the family. The added
granularity is available for models that are configured with up to 39 CPs, and provides 117
extra capacity settings. Sub-capacity models provide for CP capacity increase in two
dimensions that can be used together to deliver configuration granularity. The first dimension
is adding CPs to the configuration. The second is changing the capacity setting of the CPs
currently installed to a higher model capacity identifier.
IBM z16 allows the concurrent and nondisruptive addition of processors to a running logical
partition (LPAR). As a result, you can have a flexible infrastructure to which you can add
capacity. This function is supported by z/OS, z/VM, and z/VSE. This addition is made by using
one of the following methods:
With planning ahead for the future need of extra processors. Reserved processors can be
specified in the LPAR’s profile. When the extra processors are installed, the number of
active processors for that LPAR can be increased without the need for a partition
reactivation and initial program load (IPL).
Another (easier) way is to enable the dynamic addition of processors through the z/OS
LOADxx member. Set the DYNCPADD parameter in member LOADxx to ENABLE.
Another function concerns the system assist processor (SAP). When more SAPs are
concurrently added to the configuration, the SAP-to-channel affinity is dynamically remapped
on all SAPs on the system to rebalance the I/O configuration.
The model capacity identifier can be concurrently changed. Concurrent upgrades can be
performed for permanent and temporary upgrades.
Tip: A CPC drawer feature upgrade can be performed concurrently only for a Max39 or a
Max82 machine if feature codes 2981 or 2982 were ordered with the base machine.
2 The IBM z16 zero CP MCI is 400. This setting applies to an all-IFL or all-ICF system.
The concurrent I/O upgrade capability can be better used if a future target configuration is
considered during the initial configuration.
Important: The LICCC-based PU conversions require that at least one PU (CP, ICF, or
IFL), remains unchanged. Otherwise, the conversion is disruptive. The PU conversion
generates a LICCC that can be installed concurrently in two steps:
1. Remove the assigned PU from the configuration.
2. Activate the newly available PU as the new PU type.
LPARs also might have to free the PUs to be converted. The operating systems must include
support to configure processors offline or online so that the PU conversion can be done
nondisruptively.
Considerations: Client planning and operator action are required to use concurrent PU
conversion. Consider the following points about PU conversion:
It is disruptive if all current PUs are converted to different types.
It might require individual LPAR outages if dedicated PUs are converted.
The CIU facility is controlled through the permanent upgrade authorization FC 9898. A
prerequisite to FC 9898 is the online CoD buying feature code (FC 9900). Although FC 9898
can be installed on your IBM z16 servers at any time, often it is added when ordering an
IBM z16.
After you place an order through the CIU facility, you receive a notice that the order is ready
for download. You can then download and apply the upgrade by using functions that are
available through the Hardware Management Console (HMC), along with the RSF. After all of
the prerequisites are met, the entire process (from ordering to activation of the upgrade) is
performed by the customer, and does not require any onsite presence of IBM SSRs.
As part of the setup, provide one resource link ID for configuring and placing CIU orders and,
if required, a second ID as an approver. The IDs are then set up for access to the CIU
support. The CIU facility allows upgrades to be ordered and delivered much faster than
through the regular MES process.
To order and activate the upgrade, log on to the IBM Resource Link website, and start the CIU
application to upgrade a system for processors or memory. You can request a client order
approval to conform to your operational policies. You also can allow the definition of more IDs
to be authorized to access the CIU. More IDs can be authorized to enter or approve CIU
orders, or only view orders.
Permanent upgrades
Permanent upgrades can be ordered by using the CIU facility. Through the CIU facility, you
can generate online permanent upgrade orders to concurrently add processors (CPs, ICFs,
zIIPs, IFLs, and SAPs) and memory, or change the model capacity identifier. You can do so
up to the limits of the installed processor drawers on a system.
Temporary upgrades
The IBM z16 A01 base model describes permanent and dormant capacity by using the
capacity marker and the number of PU features that are installed on the system. Up to eight
temporary offerings can be present. Each offering includes its own policies and controls, and
each can be activated or deactivated independently in any sequence and combination.
Although multiple offerings can be active at any time, only one On/Off CoD offering can be
active at any time if enough resources are available to fulfill the offering specifications.
Temporary upgrades are represented in the system by a record. All temporary upgrade
records are on the SE hard disk drive (HDD). The records can be downloaded from the RSF
or installed from portable media. At the time of activation, you can control everything locally.
API
HMC Application
Query Activation
R3 Up to 8 records installed
R1 R2 R4 R5 R6 R7 R8 and active
Dormant
capacity
Base Change permanent capacity
model through CIU or MES order
Purchased
capacity
The authorization layer enables administrative control over the temporary offerings. The
activation and deactivation can be driven manually or under the control of an application
through a documented application programming interface (API)3.
By using the API approach, you can customize at activation time the resources that are
necessary to respond to the current situation up to the maximum that is specified in the order
record. If the situation changes, you can add or remove resources without having to return to
the base configuration. This process eliminates the need for temporary upgrade
specifications for all possible scenarios.
3 API details can be found in z/OS MVS Programming: Callable Services for High-Level Languages, SA23-1377
R1 R2 R3 R4
R3 R1
CBU CPE
R4 R4 R3 R3 R1
CBU CBU CBU CBU CPE
R2 R2 R2 R2 R2 R3
OOCoD OOCoD OOCoD OOCoD OOCoD CBU
As shown in Figure 8-2, if R2, R3, and R1 are active at the same time, only parts of R1 can be
activated because not enough resources are available to fulfill all of R1. When R2 is
deactivated, the remaining parts of R1 can be activated as shown.
Temporary capacity can be billable as On/Off CoD, or replacement capacity as CBU, CPE, or
System Recovery Boost. Consider the following points:
On/Off CoD is a function that enables concurrent and temporary capacity growth of the
system.
On/Off CoD can be used for client peak workload requirements, for any length of time, and
includes a daily hardware and maintenance charge. The software charges can vary
according to the license agreement for the individual products. For more information,
contact your IBM Software Group representative.
On/Off CoD can concurrently add processors (CPs, ICFs, zIIPs, IFLs, and SAPs),
increase the model capacity identifier, or both. It can do so up to the limit of the installed
processor drawers of a system. It is restricted to twice the installed capacity. On/Off CoD
requires a contractual agreement between you and IBM.
You decide whether to pre-pay or post-pay On/Off CoD. Capacity tokens that are inside the
records are used to control activation time and resources.
CBU is a concurrent and temporary activation of more CPs, ICFs, zIIPs, IFLs, and SAPs;
or an increase of the model capacity identifier; or both.
Note: CBU cannot be used for peak workload management in any form.
The CPE feature requires unused capacity to be available on installed processor drawers
of the backup system. The capacity must be available as unused PUs, as a possibility to
increase the model capacity identifier on a sub-capacity system, or as both.
A CPE contract must be in place before the special code that enables this capability can
be loaded on the system. The standard CPE contract provides for one 3-day planned
activation at a specific date. For more information, contact your IBM representative.
The System Recovery Boost Upgrade allows a concurrent activation of extra zIIPs.
The System Recovery Boost Upgrade record offering can be used to provide extra zIIP
capacity that can be used by the System Recovery Boost facility. You might want to
consider the use of this offering if your server is a full capacity model (7nn) and can benefit
from more CP capacity (running on zIIPs) for system shutdown and restart. The capacity
is delivered as zIIPs that can perform CP work during the boost periods for an LPAR.
A System Recovery Boost Record contract must be in place before the special code that
enables this capability can be loaded on the system. The standard contract provides for
one 6-hour activation for the specific purpose of System Recovery Boost only. For more
information, contact your IBM representative.
Activation of System Recovery Boost Record does not change the MCI of your system.
Permanent MES CPs, ICFs, zIIPs, IFLs, SAPs, processor Installed by IBM SSRs
drawer, memory, and I/Os
Online permanent CPs, ICFs, zIIPs, IFLs, SAPs, and Performed through the CIU facility
upgrade memory
Temporary On/Off CoD CPs, ICFs, zIIPs, IFLs, and SAPs Performed through the On/Off CoD
facility
CBU CPs, ICFs, zIIPs, IFLs, and SAPs Activated through model conversion
CPEa CPs, ICFs, zIIPs, IFLs, and SAPs Activated through model conversion
Flexible Capacity CPs, ICFs, zIIPs, IFLs, and SAPs Activated through model conversion
Record
a. On IBM z16, only ICFs are supported.
The MES upgrade can be performed by using LICCC only, installing more processor drawers,
adding PCIe+ I/O drawers, adding I/O4 features, or using the following combinations:
MES upgrades for processors are done by any of the following methods:
– LICCC assigning and activating unassigned PUs up to the limit of the installed
processor drawers.
– LICCC to adjust the number and types of PUs to change the capacity setting, or both.
– Installing more processor drawers and LICCC assigning and activating unassigned
PUs on the installed processor drawers.
MES upgrades for memory are done by one of the following methods:
– By using LICCC to activate more memory capacity up to the limit of the memory cards
on the currently installed processor drawers. Flexible memory features enable you to
implement better control over future memory upgrades. For more information about the
memory features, see 2.5.7, “Flexible Memory Option” on page 46.
– Installing more processor drawers and the use of LICCC to activate more memory
capacity on installed processor drawers.
4
Other adapter types, such as zHyperlink, Coupling Express LR, and Remote Direct Memory Access (RDMA) over
Converged Ethernet (RoCE), also can be added to the PCIe+ I/O drawers through an MES.
An MES upgrade requires IBM SSRs for the installation. In most cases, the time that is
required for installing the LICCC and completing the upgrade is short, depending on how up
to date the machine microcode levels are.
To better use the MES upgrade function, carefully plan the initial configuration to allow a
concurrent upgrade to a target configuration. The availability of PCIe+ I/O drawers improves
the flexibility to perform unplanned I/O configuration changes concurrently.
The Store System Information (STSI) instruction gives more useful and detailed information
about the base configuration and temporary upgrades.
The model and model capacity identifiers that are returned by the STSI instruction are
updated to coincide with the upgrade. For more information, see “Store System Information
instruction” on page 375.
Upgrades: An MES provides the physical upgrade, which results in more enabled
processors, different capacity settings for the CPs, and more memory, I/O ports, I/O
adapters, and I/O drawers. Extra planning tasks are required for nondisruptive logical
upgrades. For more information, see “Guidelines to avoid disruptive upgrades” on
page 377.
Limits: The sum of CPs, inactive CPs, ICFs, unassigned ICFs, zIIPs, unassigned zIIPs,
IFLs, unassigned IFLs, and SAPs cannot exceed the maximum limit of PUs available for
client use. The number of zIIPs cannot exceed twice the number of purchased CPs.
An IBM z16 model A01 Max39 (one processor drawer), model capacity identifier 708 (eight
CPs), is concurrently upgraded to a model A01 Max82 (two processor drawers), with MCI 742
(42 CPs). The model upgrade requires adding a processor drawer and assigning and
activating 42 PUs as CPs. Then, model Max82, MCI 742, is concurrently upgraded to a
capacity identifier 743 (43 CPs) with two IFLs. This process is done by assigning and
activating three more unassigned PUs (one as CP and two as IFLs). If needed, LPARs can be
created concurrently to use the newly added processors.
The example that is shown in Figure 8-3 shows how the addition of PUs as CPs and IFLs and
the addition of a processor drawer works. The addition of a processor drawer to an IBM z16
Max 39 upgrades the machine to Max82.
After the second CPC drawer addition, CPC drawer 0 has 39 configurable PUs and CPC
drawer 1 has 43 configurable PUs, which allows 82 PUs to be characterized on the new
Max82 configuration.
z/OS V2R3 and later 200 PUs per z/OS LPAR in non-SMT mode and 128 PUs per
z/OS LPAR in SMT mode. For both, the PU total is the sum of CPs
and zIIPs.
z/TPF 86 CPs.
Linux on IBM z16 The IBM z16 limit is 200 CPs although Linuxa supports 256 cores
without SMT and 128 cores with SMT (256 threads).
a. Supported Linux on IBM Z distributions (for more information, see Chapter 7, “Operating
system support” on page 247).
Software charges, which are based on the total capacity of the system on which the software
is installed, are adjusted to the new capacity after the MES upgrade.
Software products that use Workload License Charges (WLC) or Taylor Fit Pricing (TFP)
might not be affected by the system upgrade. Their charges are based on partition usage, not
on the system total capacity. For more information about WLC, see 7.8, “Software licensing”
on page 323.
The Flexible Memory Feature is available to allow better control over future memory
upgrades. For more information about flexible memory features, see 2.5.7, “Flexible Memory
Option” on page 46.
If the IBM z16 is a multiple processor drawer configuration, you can use the EDA feature to
remove a processor drawer and add DIMM memory cards. It also can be used to upgrade the
installed memory cards to a larger capacity size. You can then use LICCC to enable the extra
memory.
With suitable planning, memory can be added nondisruptively to z/OS partitions and z/VM
partitions. If necessary, new LPARs can be created nondisruptively to use the newly added
memory.
Concurrency: Upgrades that require DIMM changes can be concurrent by using the EDA
feature. Planning is required to see whether this option is a viable for your configuration.
The use of the flexible memory option ensures that EDA can work with the least disruption.
You also can add memory by concurrently adding a second processor drawer with sufficient
memory into the configuration and then, using LICCC to enable that memory. Changing
DIMMs in a single CPC drawer system is disruptive.
An LPAR can dynamically take advantage of a memory upgrade if reserved storage is defined
to that LPAR. The reserved storage is defined to the LPAR as part of the image profile.
Reserved memory can be configured online to the LPAR by using the LPAR dynamic storage
reconfiguration (DSR) function. DSR allows a z/OS operating system image and z/VM
partitions to add reserved storage to their configuration if any unused storage exists.
The nondisruptive addition of storage to a z/OS and z/VM partition requires the correct
operating system parameters to be set. If reserved storage is not defined to the LPAR, the
LPAR must be deactivated, the image profile changed, and the LPAR reactivated. This
process allows the extra storage resources to be available to the operating system image.
For more information about PCIe+ I/O drawers, see 4.2, “I/O system overview” on page 148.
The number of PCIe+ I/O drawers that can be present in an IBM z16 depends on how many
CPC drawers are present and whether the configuration includes the Bulk Power Assembly
(BPA) offering. It also depends on whether the CPC drawer reserve features are present.
The number of drawers for a PDU-based IBM z16 configuration options is listed in Table 8-4.
Options for a BPA-based IBM z16 are Table 8-5 on page 346. Both tables are based on no
CPC drawer reserve options being configured.
Note: The maximum number of I/O drawers in the table is reduced by one for each CPC
drawer reserve feature that is present.
Table 8-4 PCIe+ I/O drawer limit for PDU-based IBM z16 systems
Number of frames Max39 Max82 Max125 Max168/Max200
1 3 2 1 -
2 6 7 6 4
3 - 12 11 9
4 - - - 12
1 1 0 0 -
2 5 4 4 1
3 6 7 7 5
4 - 10 10 10
Depending on the number of I/O features, the configurator determines the number of PCIe+
I/O drawers that is required.
To better use the MES for I/O capability, carefully plan the initial configuration to allow
concurrent upgrades up to the target configuration.
If a PCIe+ I/O drawer is added to an IBM z16 and original features must be physically moved
to another PCIe+ I/O drawer, original card moves are disruptive.
z/VSE, z/TPF, and Linux on Z do not provide dynamic I/O configuration support. Although
installing the new hardware is done concurrently, defining the new hardware to these
operating systems requires an IPL.
Tip: IBM z16 A01 features a hardware system area (HSA) of 256 GB (same as IBM z15
T01). IBM z14 M0x servers have a 192 GB HSA. HSA is not part of the client-purchased
memory.
A FoD record can be installed only once. If it is removed, a new FoD record is needed to
reinstall. A remove action cannot be undone.
Tip: Accurate planning and the definition of the target configuration allows you to maximize
the value of these plan-ahead features.
Adding permanent upgrades to a system through the CIU facility requires that the permanent
upgrade enablement feature (FC 9898) is installed on the system. A permanent upgrade
might change the system model capacity identifier (4xx, 5yy, 6yy, or 7nn) if more CPs are
requested, or if the capacity identifier is changed as part of the permanent upgrade. If
necessary, more LPARs can be created concurrently to use the newly added processors.
Software charges that are based on the total capacity of the system on which the software is
installed are adjusted to the new capacity after the permanent upgrade is installed. Software
products that use WLC or customers with TFP might not be affected by the system upgrade
because their charges are based on LPAR usage rather than system total capacity.
For more information about WLC, see 7.8, “Software licensing” on page 323.
Customer ibm.com/servers/resourcelink
Internet
Online
permanent
Optional customer order
secondary order
approval
Remote Support
Facility
Figure 8-5 Permanent upgrade order example
The following sample sequence shows how to start an order on the IBM Resource Link:
1. Sign on to Resource Link.
2. Select Customer Initiated Upgrade from the main Resource Link page. Client and
system information that is associated with the user ID are displayed.
3. Select the system to receive the upgrade. The current configuration (PU allocation and
memory) is shown for the selected system.
4. Select Order Permanent Upgrade. The Resource Link limits the options to those options
that are valid or possible for the selected configuration (system).
5. After the target configuration is verified by the system, accept or cancel the order. An order
is created and verified against the pre-established agreement.
6. Accept or reject the price that is quoted. A secondary order approval is optional. Upon
confirmation, the order is processed. The LICCC for the upgrade is available within hours.
IBM z16
8.5.1 Ordering
IBM Resource Link provides the interface that enables you to order a concurrent upgrade for
a system. You can create, cancel, or view the order, and view the history of orders that were
placed through this interface.
Configuration rules enforce that only valid configurations are generated within the limits of the
individual system. Warning messages are issued if you select invalid upgrade options. The
process allows only one permanent CIU-eligible order for each system to be placed at a time.
The initial view of the Machine profile on Resource Link is shown in Figure 8-7.
The number of CPs, ICFs, zIIPs, IFLs, SAPs, memory size, and unassigned IFLs on the
current configuration are displayed on the left side of the page.
Resource Link retrieves and stores relevant data that is associated with the processor
configuration, such as the number of CPs and installed memory cards. It allows you to select
only those upgrade options that are deemed valid by the order process. It also allows
upgrades only within the bounds of the currently installed hardware.
When the order is available for download, you receive an email that contains an activation
number. You can then retrieve the order by using the Perform Model Conversion task from the
SE, or through the Single Object Operation to the SE from an HMC.
The window provides several possible options. If you select the Retrieve and apply data
option, you are prompted to enter the order activation number to start the permanent
upgrade, as shown in Figure 8-9.
8.6.1 Overview
The capacity for CPs is expressed in millions of service units (MSUs). Capacity for specialty
engines is expressed in number of specialty engines. Capacity tokens are used to limit the
resource consumption for all types of processor capacity.
Capacity tokens were introduced to provide better control over resource consumption when
On/Off CoD offerings are activated. Tokens represent the following resource consumptions:
For CP capacity, each token represents the amount of CP capacity that results in one
MSU of software cost for one day (an MSU-day token).
For specialty engines, each token is equivalent to one specialty engine capacity for one
day (an engine-day token).
Each specialty engine type features its own tokens, and each On/Off CoD record includes
separate token pools for each capacity type. During the ordering sessions on IBM Resource
Link, select how many tokens of each type to create for an offering record. Each engine type
must include tokens for that engine type to be activated. Capacity that has no tokens cannot
be activated.
When resources from an On/Off CoD offering record that contains capacity tokens are
activated, a billing window is started. A billing window is always 24 hours. Billing occurs at the
end of each billing window.
The resources that are billed are the highest resource use inside each billing window for each
capacity type. An activation period is one or more complete billing windows. The activation
period is the time from the first activation of resources in a record until the end of the billing
window in which the last resource in a record is deactivated.
At the end of each billing window, the tokens are decremented by the highest usage of each
resource during the billing window. If any resource in a record does not have enough tokens
to cover usage for the next billing window, the entire record is deactivated.
Note: On/Off CoD requires that the Online CoD Buying feature (FC 9900) is installed on
the system that you want to upgrade.
The On/Off CoD to Permanent Upgrade Option gives customers a window of opportunity to
assess capacity additions to your permanent configurations by using On/Off CoD. If a
purchase is made, the hardware On/Off CoD charges during this window (three days or less)
are waived. If no purchase is made, you are charged for the temporary use.
Unassigned PUs that are on the installed processor drawers can be temporarily and
concurrently activated as CPs, ICFs, zIIPs, IFLs, and SAPs through LICCC. You can assign
PUs up to twice the currently installed CP capacity, and up to twice the number of ICFs, zIIPs,
or IFLs.
An On/Off CoD upgrade cannot change the system capacity feature. The addition of
processor drawers is not supported. However, the activation of an On/Off CoD upgrade can
increase the model capacity identifier (4xx, 5yy, 6yy, or 7nn).
The use of the capacity provisioning function requires the following prerequisites:
TCP/IP connectivity to observed systems
RMF Distributed Data Server is active
CIM server is active
Security and CIM are customized
Capacity Provisioning Manager is customized
The Capacity Provisioning Manager Console is provided as part of z/OSMF, which provides a
browser-based interface for managing z/OS systems.
8.6.3 Ordering
Concurrently installing temporary capacity by ordering On/Off CoD is possible in the following
manner:
CP features equal to the MSU capacity of installed CPs
IFL features up to the number of installed IFLs
ICF features up to the number of installed ICFs
zIIP features up to the number of installed zIIPs
SAPs: Up to eight
On/Off CoD can be ordered as prepaid or postpaid. A prepaid On/Off CoD offering record
contains resource descriptions, MSUs, specialty engines, and tokens that describe the total
capacity that can be used. For CP capacity, the token contains MSU-days. For specialty
engines, the token contains specialty engine-days.
When resources on a prepaid offering are activated, they must have enough capacity tokens
to allow the activation for an entire billing window, which is 24 hours. The resources remain
active until you deactivate them or until one resource uses all of its capacity tokens. Then, all
activated resources from the record are deactivated.
A postpaid On/Off CoD offering record contains resource descriptions, MSUs, specialty
engines, and can contain capacity tokens that denote MSU-days and specialty engine-days.
When resources in a postpaid offering record without capacity tokens are activated, those
resources remain active until they are deactivated, or until the offering record expires. The
record normally expires 180 days after its installation.
When resources in a postpaid offering record with capacity tokens are activated, those
resources must include enough capacity tokens to allow the activation for an entire billing
window (24 hours). The resources remain active until they are deactivated until all of the
resource tokens are used, or until the record expires. The record usually expires 180 days
after its installation. If one capacity token type is used, resources from the entire record are
deactivated.
For example, for an IBM z16 with capacity identifier 502 (two CPs), a capacity upgrade
through On/Off CoD can be delivered in the following ways:
Add CPs of the same capacity setting. With this option, the model capacity identifier can
be changed to a 503, which adds another CP to make it a three-way CP. It also can be
changed to a 504, which adds two CPs and makes it a four-way CP.
Change to a different capacity level of the current CPs and change the model capacity
identifier to a 602 or 702. The capacity level of the CPs is increased, but no other CPs are
added. The 502 also can be temporarily upgraded to a 603, which increases the capacity
level and adds a processor. The capacity setting 439 does not have an upgrade path
through On/Off CoD because you cannot reduce the number of CPs and a 539 is more
than twice the capacity.
The On/Off CoD hardware capacity is charged on a 24-hour basis. A grace period is granted
at the end of the On/Off CoD day. This grace period allows up to an hour after the 24-hour
billing period to change the On/Off CoD configuration for the next 24-hour billing period or
deactivate the current On/Off CoD configuration. The times when the capacity is activated
and deactivated are maintained in the IBM z16 and returned to the IBM support systems.
If On/Off capacity is active, On/Off capacity can be added without having to return the system
to its original capacity. If the capacity is increased multiple times within a 24-hour period, the
charges apply to the highest amount of capacity active in that period.
If capacity is added from an active record that contains capacity tokens, the system checks
whether the resource has enough capacity to be active for an entire billing window (24 hours).
If that criteria is not met, no extra resources are activated from the record.
If necessary, more LPARs can be activated concurrently to use the newly added processor
resources.
Consideration: On/Off CoD provides a concurrent hardware upgrade that results in more
capacity being made available to a system configuration. Extra planning tasks are required
for nondisruptive upgrades. For more information, see “Guidelines to avoid disruptive
upgrades” on page 377.
To participate in this offering, you must accept contractual terms for purchasing capacity
through the Resource Link, establish a profile, and install an On/Off CoD enablement feature
on the system. Later, you can concurrently install temporary capacity up to the limits in On/Off
CoD and use it for up to 180 days.
Monitoring occurs through the system call-home facility. An invoice is generated if the
capacity is enabled during the calendar month. You are billed for the use of temporary
capacity until the system is returned to the original configuration. Remove the enablement
code if the On/Off CoD support is no longer needed.
On/Off CoD orders can be pre-staged in Resource Link to allow multiple optional
configurations. The pricing of the orders is done at the time that you order them, and the
pricing can vary from quarter to quarter. Staged orders can have different pricing.
When the order is downloaded and activated, the daily costs are based on the pricing at the
time of the order. The staged orders do not have to be installed in the order sequence. If a
staged order is installed out of sequence and later a higher-priced order is staged, the daily
cost is based on the lower price.
Another possibility is to store multiple On/Off CoD LICCC records on the SE with the same or
different capacities, which gives you greater flexibility to enable quickly needed temporary
capacity. Each record is easily identified with descriptive names, and you can select from a
list of records that can be activated.
Resource Link provides the interface to order a dynamic upgrade for a specific system. You
can create, cancel, and view the order. Configuration rules are enforced, and only valid
configurations are generated based on the configuration of the individual system. After you
complete the prerequisites, orders for the On/Off CoD can be placed. The order process uses
the CIU facility on Resource Link.
An individual record can be activated only once. Subsequent sessions require a new order to
be generated, which produces a new LICCC record for that specific order.
Alternatively, you can use an auto-renewal feature to eliminate the need for a manual
replenishment of the On/Off CoD order. This feature is implemented in IBM Resource Link,
and you must also select this feature in the machine profile, as shown in Figure 8-10.
This test can have a maximum duration of 24 hours, which commences upon the activation of
any capacity resource that is contained in the On/Off CoD record. Activation levels of capacity
can change during the 24-hour test period. The On/Off CoD test automatically stops at the
end of the 24-hour period.
You also can perform administrative testing. Although capacity is not added to the system,
you can test all of the procedures and automation for the management of the On/Off CoD
facility.
The example order that is shown in Figure 8-11 is an On/Off CoD order for 0% more CP
capacity (system is at capacity level 7), and for two more ICFs and two more zIIPs. The
maximum number of CPs, ICFs, zIIPs, and IFLs is limited by the current number of available
unused PUs of the installed processor drawers. The maximum number of SAPs is determined
by the model number and the number of available PUs on the already installed processor
drawers.
To finalize the order, you must accept Terms and Conditions for the order, as shown in
Figure 8-12.
If the On/Off CoD offering record does not contain resource tokens, you must deactivate the
temporary capacity manually. Deactivation is done from the SE and is nondisruptive.
Depending on how the capacity was added to the LPARs, you might be required to perform
tasks at the LPAR level to remove it. For example, you might have to configure offline any CPs
that were added to the partition, deactivate LPARs that were created to use the temporary
capacity, or both.
On/Off CoD orders can be staged in Resource Link so that multiple orders are available. An
order can be downloaded and activated only once. If a different On/Off CoD order is required
or a permanent upgrade is needed, it can be downloaded and activated without having to
restore the system to its original purchased capacity.
In support of automation, an API is available that allows the activation of the On/Off CoD
records. The activation is performed from the HMC and requires specifying the order number.
With this API, automation code can be used to send an activation command along with the
order number to the HMC to enable the order.
8.6.6 Termination
A customer is contractually obligated to end the On/Off CoD right-to-use feature when a
transfer in asset ownership occurs. A customer also can choose to end the On/Off CoD
right-to-use feature without transferring ownership.
Removing FC 9898 ends the right to use the On/Off CoD. This feature cannot be ordered if a
temporary session is active. Similarly, the CIU enablement feature cannot be removed if a
temporary session is active. When the CIU enablement feature is removed, the On/Off CoD
right-to-use feature is simultaneously removed. Reactivating the right-to-use feature subjects
the client to the terms and fees that apply then.
Monitoring
When you activate an On/Off CoD upgrade, an indicator is set in vital product data. This
indicator is part of the call-home data transmission, which is sent on a scheduled basis. A
timestamp is placed into the call-home data when the facility is deactivated. At the end of
each calendar month, the data is used to generate an invoice for the On/Off CoD that was
used during that month.
Software
Software Parallel Sysplex license charge (PSLC) clients are billed at the MSU level that is
represented by the combined permanent and temporary capacity. All PSLC products are
billed at the peak MSUs that are enabled during the month, regardless of usage. Customers
with WLC licenses are billed by product at the highest four-hour rolling average for the month.
In this instance, temporary capacity does not increase the software bill until that capacity is
allocated to LPARs and used.
Results from the STSI instruction reflect the current permanent and temporary CPs. For more
information, see “Store System Information instruction” on page 375.
z/OS Capacity Provisioning is delivered as part of the z/OS MVS Base Control Program
(BCP).
The Provisioning Manager monitors the workload on a set of z/OS systems and organizes the
provisioning of extra capacity to these systems when required. You define the systems to be
observed in a domain configuration file.
The details of extra capacity and the rules for its provisioning are stored in a policy file. These
two files are created and maintained through the Capacity Provisioning Management Console
(CPMC).
The operational flow of Capacity Provisioning is shown in Figure 8-13 on page 360.
The z/OS WLM manages the workload by goals and business importance on each z/OS
system. WLM metrics are available through existing interfaces, and are reported through IBM
Resource Measurement Facility (RMF) Monitor III, with one RMF gatherer for each z/OS
system.
Sysplex-wide data aggregation and propagation occur in the RMF Distributed Data Server
(DDS). The RMF Common Information Model (CIM) providers and associated CIM models
publish the RMF Monitor III data.
CPM retrieves critical metrics from one or more z/OS systems’ CIM structures and protocols.
CPM communicates to local and remote SEs and HMCs by using the Simple Network
Management Protocol (SNMP).
CPM can see the resources in the individual offering records and the capacity tokens. When
CPM activates resources, a check is run to determine whether enough capacity tokens
remain for the specified resource to be activated for at least 24 hours. If insufficient tokens
remain, no resource from the On/Off CoD record is activated.
If a capacity token is used during an activation that is driven by the CPM, the corresponding
On/Off CoD record is deactivated prematurely by the system. This process occurs even if the
CPM activates this record, or parts of it. However, you receive warning messages five days
before a capacity token is fully used.
The five days are based on the assumption that the consumption is constant for the five days.
You must put operational procedures in place to handle these situations. You can deactivate
the record manually, allow it to occur automatically, or replenish the specified capacity token
by using the Resource Link application.
The CPD configuration defines the CPCs and z/OS systems that are controlled by an
instance of the CPM. One or more CPCs, sysplexes, and z/OS systems can be defined into a
domain. Although sysplexes and CPCs do not have to be contained in a domain, they must
not belong to more than one domain.
CPM operates in the following modes, which allows four different levels of automation:
Manual
Use this command-driven mode when no CPM policy is active.
Analysis
In analysis mode, CPM processes capacity-provisioning policies and informs the operator
when a provisioning or deprovisioning action is required according to policy criteria.
Also, the operator determines whether to ignore the information or to manually upgrade or
downgrade the system by using the HMC, SE, or available CPM commands.
Several reports are available in all modes that contain information about the workload,
provisioning status, and the rationale for provisioning guidelines. User interfaces are provided
through the z/OS console and the CPMC application.
The provisioning policy defines the circumstances under which more capacity can be
provisioned (when, which, and how). The criteria features the following elements:
A time condition is when provisioning is allowed:
– Start time indicates when provisioning can begin.
– Deadline indicates that provisioning of more capacity is no longer allowed.
– End time indicates that deactivation of capacity must begin.
A workload condition is which work qualifies for provisioning. It can have the following
parameters:
– The z/OS systems that can run eligible work.
– The importance filter indicates eligible service class periods, which are identified by
WLM importance.
– Performance Index (PI) criteria:
• Activation threshold: PI of service class periods must exceed the activation
threshold for a specified duration before the work is considered to be suffering.
• Deactivation threshold: PI of service class periods must fall below the deactivation
threshold for a specified duration before the work is considered to no longer be
suffering.
– Included service classes are eligible service class periods.
– Excluded service classes are service class periods that must not be considered.
Tip: If no workload condition is specified, the full capacity that is described in the policy
is activated and deactivated at the start and end times that are specified in the policy.
Provisioning scope is how much more capacity can be activated and is expressed in
MSUs.
The number of zIIPs must be one specification per CPC that is part of the CPD and are
specified in MSUs.
The maximum provisioning scope is the maximum extra capacity that can be activated for
all the rules in the CPD.
In the specified time interval, the provisioning rule is that up to the defined extra capacity can
be activated if the specified workload is behind its objective.
The rules and conditions are named and stored in the Capacity Provisioning Policy.
For more information about z/OS Capacity Provisioning functions, see z/OS MVS Capacity
Provisioning User’s Guide, SC34-2661.
The provisioning management routines can interrogate the installed offerings, their content,
and the status of the content of the offering. To avoid the decrease in capacity, create only
one On/Off CoD offering on the system by specifying the maximum allowable capacity. Then,
when an activation is needed, the CPM can activate a subset of the contents of the offering
sufficient to satisfy the demand. If more capacity is needed later, the Provisioning Manager
can activate more capacity up to the maximum allowed increase.
Multiple offering records can be pre-staged on the SE HDD. Changing the content of the
offerings (if necessary) also is possible.
Remember: CPM controls capacity tokens for the On/Off CoD records. In a situation
where a capacity token is used, the system deactivates the corresponding offering record.
Therefore, you must prepare routines for catching the warning messages about capacity
tokens being used, and have administrative procedures in place for such a situation.
The messages from the system begin five days before a capacity token is fully used. To
avoid capacity records being deactivated in this situation, replenish the necessary capacity
tokens before they are used.
The CPM operates based on Workload Manager (WLM) indications, and the construct that is
used is the Performance Index (PI) of a service class period. It is important to select service
class periods that are suitable for the business application that needs more capacity.
For example, the application in question might be running through several service class
periods, where the first period is the important one. The application might be defined as
importance level 2 or 3, but might depend on other work that is running with importance
level 1. Therefore, it is important to consider which workloads to control and which service
class periods to specify.
Important: The base System Recovery Boost capability is built into IBM z16 firmware and
does not require ordering more features. System Recovery Boost Upgrade (consisting of
FC 9930 and FC 6802) is an optional, orderable feature that provides more temporary zIIP
capacity for use during boost periods. Consider the following points:
FC 9930 is not required to use the base System Recovery Boost capability.
FC 9930 is only needed if more zIIP temporary capacity is required.
The System Recovery Boost Upgrade optional feature is offered with IBM z16 A01 to provide
more capacity for System Recovery Boost processing. For example, if a z/OS operating
system change requires a sequence of system shutdowns and restarts, and these
procedures can benefit from extra CPU capacity, the System Recovery Boost record can be
used to activate more zIIPs on the server at the commencement of the change window.
These zIIPs are used by the z/OS systems to run general CP work during the boost periods.
The System Recovery Boost Upgrade requires the following feature codes:
FC 6802: System Recovery Boost Record
This feature code provides extra zIIP capacity (20 zIIP records for 1 - 5 years, orderable by
FC 6799, and must be renewed if expired) for use in System Recovery Boost events
(shutdown, restart/IPL, and stand-alone memory dumps).
Note: At least one permanent zIIP record must be present for ordering System
Recovery Boost Upgrade (FC 6802).
Important: The System Recovery Boost Upgrade record is for System Recovery Boost
capacity only. It cannot be used for peak workload management. The customer must
deactivate the boost capacity at the end of the system boost procedure. The zIIP capacity
that is provided in the SRB Record is for LPAR shutdown and startup (IPL and Catchup)
and is not available for process boost events.
The zIIP processors that can be activated by System Recovery Boost record come from the
“dark core” capacity on the server. They can be added to an IBM z16 nondisruptively.
The base system configuration must have sufficient memory and channels to accommodate
the potential requirements of the larger capacity system.
Note: The System Recovery Boost configuration is activated temporarily and provides up
to a maximum of 20 extra zIIPs to the system’s original, permanent configuration and can
violate the 2:1 zIIP rule. The number of zIIPs that can be activated is limited by the unused
capacity that is available on the system.
When activating the System Recovery Boost record, the extra zIIPs are added to the zIIP pool
when they are activated. Review the LPAR zIIP assignments and weights in the image
profiles to ensure that the LPAR can use the extra capacity when it becomes available.
This feature can be ordered in the range of 1 - 5 years. Only one record can be installed on a
system and it must be replenished before it can be used again (if expired).
A System Recovery Boost Upgrade contract (through FC 9930) must be in place before the
special code that enables this capability can be installed on the system.
Flexible Capacity for Cyber Resiliency is a new CoD offering that is available on IBM z16
machines. This offering allows processing capacity flexibility between an organization’s
primary site and alternative data centers.
Flexible Capacity for Cyber Resiliency is designed to provide increased flexibility and control
to organizations that want to shift production capacity between participating IBM z16 servers
at different sites. The capacity of any engine type can be shifted up to 12 times annually and
stay at the target machine for up to 12 months after the flexible capacity record activation on
the target machine. Capacity shifts can be done under full customer control without IBM
intervention and can be fully automated by using IBM GDPS automation tools.
Flexible Capacity for Cyber Resiliency supports a broad set of scenarios and can be
combined with other IBM On-Demand offerings.
For more information, see “IBM Z Flexible Capacity for Cyber Resiliency” on page 520.
Important: CBU is for disaster and recovery purposes only. It cannot be used for peak
workload management or for a planned event.
5
All new CBU contract documents contain new CBU test terms to allow execution of production workload during
CBU test. CBU clients must sign the IBM client Agreement Amendment for IBM Z Capacity Backup Upgrade Tests
(US form #Z125-8145).
The CBU entitlement record (FC 6818) contains an expiration date that is established at the
time of the order. This date depends on the quantity of CBU years (FC 6817). You can extend
your CBU entitlements through the purchase of more CBU years.
The number of FC 6817 per instance of FC 6818 remains limited to five. Fractional years are
rounded up to the nearest whole integer when calculating this limit.
If two years and eight months exist before the expiration date at the time of the order, the
expiration date can be extended by no more than two years. One test activation is provided for
each CBU year that is added to the CBU entitlement record.
FC 6805 allows for ordering more tests in increments of one. The maximum number of tests
that is allowed is 15 for each FC 6818.
The processors that can be activated by CBU come from the available unassigned PUs on
any installed processor drawer. A maximum of 200 CBU features that can be ordered. The
number of features that can be activated is limited by the number of unused PUs on the
system; for example:
An IBM z16 Max 39 with Capacity Model Identifier 401 can activate up to 39 CBU features.
These CBU features can be used to change the capacity setting of the CPs, and to
activate unused PUs.
An IBM z16 Max 82 CMI 715 with 15 CPs, 4 IFLs, and 1 ICF has 62 unused PUs available.
It can activate up to 62 CBU features.
The ordering system allows for over-configuration in the order. You can order up to 200 CBU
features, regardless of the current configuration. However, at activation, only the capacity that
is installed can be activated. At activation, you can decide to activate only a subset of the
CBU features that are ordered for the system.
Sub-capacity makes a difference in the way that the CBU features are completed. On the
full-capacity models, the CBU features indicate the amount of extra capacity that is needed. If
the amount of necessary CBU capacity is equal to four CPs, the CBU configuration is four
CBU CPs.
For example, if the base configuration is a two-way 402, two CBU feature codes are required
to provide a CBU configuration of a four-way of the same capacity setting. If the required CBU
capacity changes the capacity setting of the CPs, going from model capacity identifier 402 to
a CBU configuration of a four-way 504 requires four CBU feature codes with a capacity setting
of 5yy.
If the capacity setting of the CPs is changed, more CBU features are required, not more
physical PUs. Therefore, your CBU contract requires more CBU features when the capacity
setting of the CPs is changed.
CBU can add CPs through LICCC only, and the IBM z16 A01 must have the correct number
of installed processor drawers to allow the required upgrade. CBU can change the model
capacity identifier to a higher value than the base setting (4xx, 5yy, or 6yy), but the CBU
feature cannot decrease the capacity setting.
A CBU contract must be in place before the special code that enables this capability can be
installed on the system. CBU features can be added to an IBM z16 nondisruptively. For each
system enabled for CBU, the authorization to use CBU is available for 1 - 5 years.
The alternative configuration is activated temporarily, and provides more capacity than the
system’s original, permanent configuration. At activation time, determine the capacity that you
require for that situation. You can decide to activate only a subset of the capacity that is
specified in the CBU contract.
The base system configuration must have sufficient memory and channels to accommodate
the potential requirements of the large CBU target system. Ensure that all required functions
and resources are available on the backup systems. These functions include CF LEVELs for
coupling facility partitions, memory, and cryptographic functions, and connectivity capabilities.
When the emergency is over (or the CBU test is complete), the system must be returned to its
original configuration. The CBU features can be deactivated at any time before the expiration
date. Failure to deactivate the CBU feature before the expiration date can cause the system to
downgrade resources gracefully to the original configuration. The system does not deactivate
dedicated engines, or the last of in-use shared engines.
Planning: CBU for processors provides a concurrent upgrade. This upgrade can result in
more enabled processors, changed capacity settings that are available to a system
configuration, or both. You can activate a subset of the CBU features that are ordered for
the system. Therefore, more planning and tasks are required for nondisruptive logical
upgrades. For more information, see “Guidelines to avoid disruptive upgrades” on
page 377.
For more information, see the Capacity on Demand User’s Guide, SC28-6846.
CBU activation
CBU is activated from:
The SE by using the HMC and SSO to the SE
By using the Perform Model Conversion task
Through automation by using the API on the SE or the HMC
During a real disaster, use the Activate CBU option to activate the 90-day period.
Image upgrades
After CBU activation, the IBM z16 can have more capacity, more active PUs, or both. The
extra resources go into the resource pools and are available to the LPARs. If the LPARs must
increase their share of the resources, the LPAR weight can be changed or the number of
logical processors can be concurrently increased by configuring reserved processors online.
The operating system must concurrently configure more processors online. If necessary,
more LPARs can be created to use the newly added capacity.
CBU deactivation
To deactivate the CBU, the extra resources must be released from the LPARs by the
operating systems. In some cases, this process involves varying the resources offline. In
other cases, it can mean shutting down operating systems or deactivating LPARs. After the
resources are released, the same facility on the HMC/SE is used to turn off CBU. To
deactivate CBU, select the Undo temporary upgrade option from the Perform Model
Conversion task on the SE.
CBU testing
Test CBUs are provided as part of the CBU contract. CBU is activated from the SE by using
the Perform Model Conversion task. Select the test option to start a 10-day test period. A
standard contract allows one test per CBU year. However, you can order more tests in
increments of one up to a maximum of 15 for each CBU order.
Tip: The CBU test activation is done the same way as the real activation; that is, by using
the same SE Perform a Model Conversion window and selecting the Temporary
upgrades option. The HMC windows were changed to avoid accidental real CBU
activations by setting the test activation as the default option.
The test CBU must be deactivated in the same way as the regular CBU. Failure to deactivate
the CBU feature before the expiration date can cause the system to degrade gracefully back
to its original configuration. The system does not deactivate dedicated engines or the last
in-use shared engine.
Alternatively, 4 CP CBU features can be used to change the MCI (in the example from a 504
to a 704) and then add the remaining 3 CP CBU features to upgrade to a 707 (the red path).
Note: System Recovery Boost record does not affect model capacity identifier.
The GDPS service is for z/OS only, or for z/OS in combination with Linux on Z.
For more information, see Appendix D, “Tailored Fit Pricing and IBM Z Flexible Capacity for
Cyber Resiliency” on page 513.
Flexible Capacity for Cyber Resiliency can be ordered by contacting your IBM hardware sales
representative. The offering requires that an order is placed against each serial number (SN)
that is involved in capacity transfer with one record per SN.
Installation and setup: The new Flexible Capacity Record is installed and set up on each
participating IBM z16 system. Consider the following points:
On the IBM z16 source system, the permanent capacity is unassigned to the base level.
The new Flexible Capacity Record is installed and activated on the IBM z16 source
system to restore capacity back to the purchased level.
On the IBM z16 target systems, the new Flexible Capacity Record enables clients to bring
the capacity up to the level of the production system when activated. The Flexible Capacity
Record remains inactive until capacity is shifted from the base system to the target
systems.
After deactivating the Flexible Capacity Record on the base system, the capacity active
through Flexible Capacity Transfer records on the target systems must not exceed the
capacity active on the base system before the swap.
If GDPS is used to automate the shift, it must be set up with the correct LIC records to add
capacity in the target systems and remove capacity in the base systems site.
A Flexible Capacity record is installed on the machines at each site. The machine at Site A
has the base capacity that is configured at the same level as the high water mark of the Site B
machine and has flex capacity that is activated to its HWM.
The flex capacity is then added to the machine at Site B, which brings it up to the Site A HWM
and workload is transferred.
After all workload is transferred (within 24 hours), the capacity of the Site A machine is
reduced to the base level. Workload can continue to run at Site B for up to a year.
Capacity transfer is shown in Figure 8-17. Flexible Capacity Record is ACTIVE in both sites
for up to 24 hours.
Figure 8-17 Flexible Capacity record at Site B is activated and workload transferred
After the site swap (capacity active in both sites for up to 24 hours), workload can stay in Site
B for up to one year (see Figure 8-18 on page 372).
IBM z16 allows concurrent upgrades, which means that dynamically adding capacity to the
system is possible. If the operating system images that run on the upgraded system do not
require disruptive tasks to use the new capacity, the upgrade is also nondisruptive. This
process avoids power-on resets (POR), LPAR deactivation, and IPLs.
If the concurrent upgrade is intended to satisfy an image upgrade to an LPAR, the operating
system that is running in this partition must concurrently configure more capacity online. z/OS
operating systems include this capability. z/VM can concurrently configure new processors
and I/O devices online, and memory can be dynamically added to z/VM partitions.
If the concurrent upgrade is intended to satisfy the need for more operating system images,
more LPARs can be created concurrently on the IBM z16 system. These LPARs include all
resources that are needed. These extra LPARs can be activated concurrently.
These enhanced configuration options are available through the HSA, which is an IBM
reserved area in system memory.
In general, Linux operating systems cannot add more resources concurrently. However, Linux
and other types of virtual machines that run under z/VM can benefit from the z/VM capability
to nondisruptively configure more resources online (processors and I/O).
With z/VM, Linux guests can manipulate their logical processors by using the Linux CPU
hotplug daemon. The daemon can start and stop logical processors that are based on the
Linux load average value. The daemon is available in Linux SLES 10 SP2 and later, and in
Red Hat Enterprise Linux (RHEL) V5R4 and up.
PUs
CPs, ICFs, zIIPs, IFLs, and SAPs can be added concurrently to an IBM z16 if unassigned
PUs are available on any installed processor drawer. The number of zIIPs cannot exceed
twice the number of CPs. The IBM z16 allows the concurrent addition of a second and third
processor drawer if the CPC reserve features are installed.
If necessary, more LPARs can be created concurrently to use the newly added processors.
The Coupling Facility Control Code (CFCC) also can configure more processors online to
coupling facility LPARs by using the CFCC image operations window.
Memory
Memory can be added concurrently up to the physical installed memory limit. More processor
drawers can be installed concurrently, which allows further memory upgrades by LICCC, and
enables memory capacity on the new processor drawers.
By using the previously defined reserved memory, z/OS operating system images, and z/VM
partitions, you can dynamically configure more memory online. This process allows
nondisruptive memory upgrades. Linux on Z supports Dynamic Storage Reconfiguration.
I/O
I/O features can be added concurrently if all the required infrastructure (I/O slots and PCIe
Fan-outs) is present in the configuration. PCIe+ I/O drawers can be added concurrently
without planning if free space is available in one of the frames and the configuration permits.
Dynamic I/O configuration changes are supported by specific operating systems (z/OS and
z/VM), which allows for nondisruptive I/O upgrades. Dynamic I/O reconfiguration on a
stand-alone coupling facility system also is possible by using the Dynamic I/O activation for
stand-alone CF CPCs features.
Cryptographic adapters
Crypto Express8S features can be added concurrently if all the required infrastructure is in
the configuration.
Enabling and using the extra processor capacity is not apparent to most applications.
However, specific programs depend on processor model-related information, such as ISV
products. Consider the effect on the software that is running on an IBM z16 when you perform
any of these configuration upgrades.
Processor identification
The following instructions are used to obtain processor information:
Store System Information (STSI) instruction
The STSI instruction can be used to obtain information about the current execution
environment and any processing level that is below the current environment. It can be
used to obtain processor model and model capacity identifier information from the basic
machine configuration form of the system information block (SYSIB). It supports
concurrent upgrades and is the recommended way to request processor information.
Store CPU ID (STIDP) instruction
STIDP returns information that identifies the execution environment, system serial
number, and machine type.
Note: To ensure unique identification of the configuration of the issuing CPU, use the STSI
instruction specifying basic machine configuration (SYSIB 1.1.1).
The model capacity identifier contains the base capacity, On/Off CoD, and CBU. The Model
Permanent Capacity Identifier and the Model Permanent Capacity Rating contain the base
capacity of the system. The Model Temporary Capacity Identifier and Model Temporary
Capacity Rating contain the base capacity and On/Off CoD.
For more information about the STSI instruction, see z/Architecture Principles of Operation,
SA22-7832.
For more information about the STIDP instruction, see z/Architecture Principles of Operation,
SA22-7832.
In a multi-site high-availability configuration, another option is the use of Flexible Capacity for
Cyber Resilience to move workload to another site while hardware maintenance is performed.
You can minimize the need for these outages by carefully planning and reviewing “Guidelines
to avoid disruptive upgrades” on page 377.
A fourth or fifth processor drawer can be installed at the IBM Manufacturing plant only.
One major client requirement was to eliminate the need for a client authorization connection
to the IBM Resource Link system when activating an offering. This requirement is met by the
IBM z14, IBM z15, and IBM z16 servers.
After the offerings are installed on the IBM z16 SE, they can be activated at any time at the
customer’s discretion. No intervention by IBM or IBM personnel is necessary. In addition, the
activation of CBU does not require a password.
The IBM z16 A01 can have up to eight offerings that are installed at the same time, with the
limitation that only one of them can be an On/Off CoD offering, and only one can be an SRB
Upgrade record. The others can be any combination.
The installed offerings can be activated fully or partially, and in any sequence and any
combination. The offerings can be controlled manually through command interfaces on the
HMC, or programmatically through a number of APIs. IBM applications, ISV programs, and
client-written applications can control the use of the offerings.
Resource usage (and therefore, financial exposure) can be controlled by using capacity
tokens in the On/Off CoD offering records.
The CPM is an example of an application that uses the CoD APIs to provision On/Off CoD
capacity that is based on the requirements of the workload. The CPM cannot control other
offerings.
For more information about any of the topics in this chapter, see Capacity on Demand User’s
Guide, SC28-6943.
Note: Throughout this chapter, IBM z16 refers to IBM z16 Model A01 (Machine Type
3931).
The key objectives, in order of priority, are to ensure data integrity, computational integrity,
reduce or eliminate unscheduled outages, reduce scheduled outages, reduce planned
outages, and reduce the number of Repair Actions.
RAS can be accomplished with improved concurrent replace, repair, and upgrade functions
for processors, memory, drawers, and I/O. RAS also extends to the nondisruptive capability
for installing Licensed Internal Code (LIC) updates. In most cases, a capacity upgrade can be
concurrent without a system outage. As an extension to the RAS capabilities, environmental
controls are implemented in the system to help reduce power consumption and meet cooling
requirements.
The following overriding RAS requirements are principles as shown in Figure 9-1:
Include existing (or equivalent) RAS characteristics from previous generations.
Learn from current field issues and addressing the deficiencies.
Understand the trend in technology reliability (hard and soft) and ensure that the RAS
design points are sufficiently robust.
Invest in RAS design enhancements (hardware and firmware) that provide IBM Z and
Customer valued differentiation.
9.2 Technology
This section introduces some of the RAS features that are incorporated in the IBM z16
design.
Two PU chips are packaged on a dual chip module (DCM), as shown in Figure 9-3. PU
chip to PU chip communication is performed through the M-Bus, which is a high-speed
bus that acts as ring-to-ring extension communication at 160 Gbps data rate, with the
following RAS characteristics:
– ECC on data path and snoop bus
– Parity on miscellaneous control lines
– One spare per ~50 data bit bus
Figure 9-3 IBM z16 Dual Chip Module (M-Bus connects the two chips)
IBM z16 processor memory and cache structure are shown in Figure 9-5. The physical L3
cache (on chip) and System Controller (SC) SCM (physical L4 cache for IBM z15), which was
implemented in EDRAM, were removed and replaced on IBM z16 virtual L3 and L4 cache
structure.
Note: All of these features are available without Flexible Memory, but all customer
purchased memory is not available for use in most cases. Some work might need to be
shut down or not restarted.
9.3 Structure
The IBM z16 was designed in a 19-inch frames format. The IBM z16 A01 can be delivered
only as an air-cooled system and fulfills the requirements for ASHRAE A3 class environment.
The IBM z16 A01 can have up to 10 PCIe+ I/O drawers when delivered with Bulk Power
Assembly (BPA) and 12 PCIe+ I/O drawers when delivered with Power Distribution Unit
(PDU). The structure of the IBM z16 is designed with the following goals:
Enhanced system modularity
Standardization to enable rapid integration
Platform simplification
Cables are keyed to ensure that correct lengths are plugged. Plug detection ensures correct
location, and custom latches ensure retention. Further improvements to the fabric bus include
symmetric multiprocessing (SMP) cables that connect the drawers.
The thermal RAS design also was improved for the field-replaceable water manifold for PU
cooling.
Independent channel recovery with replay buffers on all interfaces allows recovery of a single
DIMM channel, while other channels remain active. Further redundancies are incorporated in
I/O pins for clock lines to main memory, which eliminates the loss of memory clocks because
of connector (pin) failure. The following RAS enhancements reduce service complexity:
Continued use of RAIM ECC
RAIM logic was moved on DIMM
N+1 on-DIMM voltage regulation
Replay buffer for hardware retry on soft errors on the main memory interface
Redundant I/O pins for clock lines to main memory
Staggered refresh for performance enhancement
The new RAIM scheme achieves a higher ratio of data to ECC symbols, while also
providing another chip mark
Firmware was updated to improve filtering and resolution of errors that do not require action.
Enhanced integrated sparing in processor cores, cache relocates, N+1 SEEPROM and Point
of load (POL) N+2 redundancies, and DRAM marking also are incorporated to reduce
touches.
The new IBM Z Channel Subsystem Function performs periodic polling from the channel
to the end points for the logical paths that are established and reduces the number of
useless Repair Actions (RAs).
The RDP data history is used to validate Predictive Failure Algorithms and identify Fibre
Channel Links with degrading signal strength before errors start to occur. The new Fibre
Channel Extended Link Service (ELS) retrieves signal strength.
FICON Dynamic Routing
FICON Dynamic Routing (FIDR) enables the use of storage area network (SAN) dynamic
routing policies in the fabric. With the IBM z16, FICON channels are no longer restricted to
the use of static routing policies for inter-switch links (ISLs) for cascaded FICON directors.
FICON Dynamic Routing dynamically changes the routing between the channel and
control unit that is based on the Fibre Channel Exchange ID. Each I/O operation has a
unique exchange ID. FIDR supports static SAN routing policies and dynamic routing
policies.
FICON Dynamic Routing can help clients reduce costs by providing the following features:
– Share SANs between their FICON and FCP traffic.
– Improve performance because of SAN dynamic routing policies that better use all of
the available ISL bandwidth through higher use of the ISLs,
– Simplify management of their SAN fabrics by using static routing policies that assign
different ISL routes with each power-on-reset (POR), which makes the SAN fabric
performance difficult to predict.
Customers must ensure that all devices in their FICON SAN support FICON Dynamic
Routing before they implement this feature.
The difference between scheduled outages and planned outages might not be obvious. The
general consensus is that scheduled outages occur sometime soon. The timeframe is
approximately two weeks.
Planned outages are outages that are planned well in advance and go beyond this
approximate two-week timeframe. This chapter does not distinguish between scheduled and
planned outages.
Preventing unscheduled, scheduled, and planned outages was addressed by the IBM Z
system design for many years.
IBM z16 A01 has a fixed size HSA of 256 GB. This size helps eliminate planning
requirements for HSA and provides the flexibility to update dynamically the configuration. You
can perform the following tasks dynamically:1
Add:
– Logical partition (LPAR)
– Logical channel subsystem (LCSS)
– Subchannel set
– Logical PU to an LPAR
– Cryptographic coprocessor
– Memory
– Physical processor
Remove a cryptographic coprocessor
Enable I/O connections
Swap processor types
By addressing the elimination of planned outages, the following tasks also are possible:
Concurrent driver upgrades
Concurrent and flexible customer-initiated upgrades
For more information about the flexible upgrades that are started by customers, see 8.3.2,
“Customer Initiated Upgrade facility” on page 336.
1 Some planning considerations might exist. For more information, see Chapter 8, “System upgrades” on page 327.
Coupling Express2 LR does not support the “going away signal”; however, ECAR can be
used to assist with recovery in the following configurations:
Design of pervasive infrastructure controls in processor chips in memory ASICs.
Improved error checking in the processor recovery unit (RU) to better protect against word
line failures in the RU arrays.
The EDA procedure and careful planning help ensure that all the resources are still available
to run critical applications in an (n-1) drawer configuration. This process allows you to avoid
planned outages. Consider the flexible memory option to provide more memory resources
when you are replacing a drawer.
For more information about flexible memory, see 2.5.7, “Flexible Memory Option” on page 46.
To minimize the effect on current workloads, ensure that sufficient inactive physical resources
exist on the remaining drawers to complete a drawer removal. Also, consider deactivating
noncritical system images, such as test or development LPARs. After you stop these
noncritical LPARs and free their resources, you might find sufficient inactive resources to
contain critical workloads while completing a drawer replacement.
The following configurations especially enable the use of the EDA function. These IBM z16
features need enough spare capacity so that they can cover the resources of a fenced or
isolated drawer. This configuration imposes limits on the following number of the client-owned
PUs that can be activated when one drawer within a model is fenced:
A maximum of:
– 39 PUs are configured on the Max39
– 82 PUs are configured on the Max82
– 125 PUs are configured on the Max125
– 168 PUs are configured on the Max168
– 200 PUs are configured on the Max200
The system configuration must have sufficient dormant resources on the remaining drawers
in the system for the evacuation of the drawer that is to be replaced or upgraded. Dormant
resources include the following possibilities:
Unused PUs or memory that is not enabled by LICCC
Inactive resources that are enabled by LICCC (memory that is not being used by any
activated LPARs)
Memory that is purchased with the flexible memory option
Extra drawers
The I/O connectivity also must support drawer removal. Most of the paths to the I/O feature
redundant I/O interconnect support in the I/O infrastructure (drawers) that enable connections
through multiple fan-out cards.
If sufficient resources are not present on the remaining drawers, specific noncritical LPARs
might need to be deactivated. One or more PUs or storage might need to be configured
offline to reach the required level of available resources. Plan to address these possibilities to
help reduce operational errors.
Include the planning process as part of the initial installation and any follow-on upgrade that
modifies the operating environment. A customer can use the IBM Resource Link machine
information report to determine the number of drawers, active PUs, memory configuration,
and channel layout.
If the IBM z16 is installed, click Prepare for Enhanced Drawer Availability in the Perform
Model Conversion window of the EDA process on the Hardware Management Appliance
(HMA). This task helps you determine the resources that are required to support the removal
of a drawer with acceptable degradation to the operating system images.
The EDA process determines which resources (including memory, PUs, and I/O paths) are
free to allow for the removal of a drawer. You can run this preparation on each drawer to
determine which resource changes are necessary. Use the results as input in the planning
stage to help identify critical resources.
With this planning information, you can examine the LPAR configuration and workload
priorities to determine how resources might be reduced and still allow the drawer to be
concurrently removed.
When you perform the review, document the resources that can be made available if the EDA
is used. The resources on the drawers are allocated during a POR of the system and can
change after that process.
Perform a review when changes are made to your IBM z16, such as adding drawers, PUs,
memory, or channels. Also, perform a review when workloads are added or removed, or if the
HiperDispatch feature was enabled and disabled since the last time you performed a POR.
Processor availability
Processor resource availability for reallocation or deactivation is affected by the type and
quantity of the resources in use, such as the following examples:
Total number of PUs that are enabled through LICCC
PU definitions in the profiles that can be dedicated and dedicated reserved or shared
Active LPARs with dedicated resources at the time of the drawer repair or replacement
To maximize the PU availability option, ensure that sufficient inactive physical resources are
on the remaining drawers to complete a drawer removal.
Memory availability
Memory resource availability for reallocation or deactivation depends on the following factors:
Physically installed memory
Image profile memory allocations
Amount of memory that is enabled through LICCC
Flexible memory option
Virtual Flash Memory if enabled and configured
For more information, see 2.7.2, “Enhanced drawer availability” on page 54.
Preparation: The preparation step does not reallocate any resources. It is used only to
record customer choices and produce a configuration file on the SE that is used to run the
concurrent drawer replacement operation.
The preparation step can be done in advance. However, if any changes to the configuration
occur between the preparation and the physical removal of the drawer, you must rerun the
preparation phase.
The process can be run multiple times because it does not move any resources. To view the
results of the last preparation operation, click Display Previous Prepare Enhanced Drawer
Availability Results from the Perform Model Conversion window in the SE.
The preparation step can be run without performing a drawer replacement. You can use it to
dynamically adjust the operational configuration for drawer repair or replacement before IBM
SSR activity. The Perform Model Conversion window in you click Prepare for Enhanced
Drawer Availability is shown in Figure 9-7 on page 397.
After you click Prepare for Enhanced Drawer Availability, the Enhanced Drawer Availability
window opens. Select the drawer that is to be repaired or upgraded; then, select OK, as
shown in Figure 9-8. Only one target drawer can be selected at a time.
The system verifies the resources that are required for the removal, determines the required
actions, and presents the results for review. Depending on the configuration, the task can take
from a few seconds to several minutes.
The preparation step determines the readiness of the system for the removal of the targeted
drawer. The configured processors and the memory in the selected drawer are evaluated
against unused resources that are available across the remaining drawers. The system also
analyzes I/O connections that are associated with the removal of the targeted drawer for any
single path I/O connectivity.
If insufficient resources are available, the system identifies the conflicts so that you can free
other resources.
Preparation tabs
The results of the preparation are presented for review in a tabbed format. Each tab indicates
conditions that prevent the EDA option from being run. The following tab selections are
available:
Processors
Memory
Single I/O
Single Domain I/O
Single Alternative Path I/O
Only the tabs that feature conditions that prevent the drawer from being removed are
displayed. Each tab indicates the specific conditions and possible options to correct them.
For example, the preparation identifies single I/O paths that are associated with the removal
of the selected drawer. These paths must be varied offline to perform the drawer removal.
After you address the condition, rerun the preparation step to ensure that all the required
conditions are met.
After you review the reassignment results and make any necessary adjustments, click OK
(see Figure 9-10).
By understanding the system configuration and the LPAR allocation for memory, PUs, and
I/O, you can make the best decision about how to free the necessary resources to allow for
drawer removal.
Upon successful completion, the system is ready for the removal of the drawer.
Review the results. The result of the preparation task is a list of resources that must be made
available before the drawer replacement can occur.
Reserved storage: If you plan to use the EDA function with z/OS LPARs, set up
reserved storage and an RSU value. Use the RSU value to specify the number of
storage units that are to be kept free of long-term fixed storage allocations. This
configuration allows for storage elements to be varied offline.
When correctly configured, IBM z16 supports concurrently activating a selected new LIC
Driver level. Concurrent activation of the selected new LIC Driver level is supported only at
specific released sync points. Concurrently activating a selected new LIC Driver level
anywhere in the maintenance stream is not possible. Specific LIC updates do not allow a
concurrent update or upgrade.
The CDM function does not eliminate the need for planned outages for driver-level upgrades.
Upgrades might require a system level or a functional element scheduled outage to activate
the new LIC. The following circumstances require a scheduled outage:
Specific complex code changes might dictate a disruptive driver upgrade. You are alerted
in advance so that you can plan for the following changes:
– Design data or hardware initialization data fixes
– CFCC release level change
OSA CHPID code changes might require PCHID Vary OFF/ON to activate new code.
Crypto code changes might require PCHID Vary OFF/ON to activate new code.
Note: zUDX clients should contact their User Defined Extensions (UDX) provider
before installing Microcode Change Levels (MCLs). Any changes to Segments 2 and 3
from a previous MCL level might require a change to the customer’s UDX. Attempting to
install an incompatible UDX at this level results in a Crypto checkstop.
The native PCIe features (managed by Resource Group code) are listed Table 9-1.
Consider the following points for managing native PCIe adapters microcode levels:
Updates to the Resource Group require all native PCIe adapters that are installed in that
RG to be offline.
Updates to the native PCIe adapter require the adapter to be offline. If the adapter is not
defined, the MCL session automatically installs the maintenance that is related to the
adapter.
The PCIe native adapters are configured with Function IDs (FIDs) and might need to be
configured offline when changes to code are needed. To help alleviate the number of
adapters (and FIDs) that are affected by the Resource Group code update, IBM z16 have four
Resource Groups per system (CPC).
Note: Other adapter types, such as FICON Express, OSA Express, and Crypto Express
that are installed in the PCIe+ I/O drawer are not affected because they are not managed
by the Resource Groups.
2
If HMA feature is installed on the system, special upgrade procedure must be followed to ensure nondisruptive SE
upgrade.
This chapter describes the newest elements for the HMC and SE.
Tip: The Help function is a good starting point to get more information about all of the
functions that can be used by the HMC and SE. The Help feature is available by clicking
Help from the drop-down menu that appears when you click your user ID.
For more information, see the IBM Resource Link, select Library, the applicable server
and then, select Hardware Management Console Operations Guide or Support
Element Operations Guide.
The HMC runs a set of management applications. On IBM z16, two HMCs are delivered with
the Hardware Management Appliance (HMA) feature FC 0129. The HMC code runs on the
two integrated 1U rack-mounted servers on the top of the IBM z16 A frame. Stand-alone
outside the IBM z16 HMCs (Tower or Rack Mount) no longer can be ordered.
The following HMC feature codes can be carried forward from previous orders:
FC 0062
FC 0063
FC 0082
FC 0083
On the carry forward HMC features, Driver 51/Version 2.16.0 must be installed to support IBM
z16. Also, Driver 51/Version 2.16.0 can be installed on the SE/HMC hardware that is provided
with the HMA feature FC 0100 on IBM z15.
You can use the IBM z16 with carry forward HMCs and no initial order of FC 0129 (HMA for
IBM z16). Order of the HMA FC 0129 is possible later as MES.
With IBM z16, the HMA feature shares the integrated 1U server hardware (installed on top of
rack A with) the SE code. The SE code runs virtualized under the HMC on each of the two
integrated 1U rack-mounted servers. One SE is the Primary SE (active) and the other is the
Alternative SE (backup). As with the HMCs, the SEs are closed systems (appliances), and no
other applications can be installed on the same hardware.
The HMC is used to set up, manage, monitor, and operate one or more IBM Z CPCs. It
manages IBM Z hardware and its logical partitions (LPARs), and provides support
applications. At least one HMC is required to operate an IBM Z. An HMC can manage
multiple IBM Z CPCs.
When tasks are performed at the HMC, the commands are routed to the Primary SE of the
IBM z16. The SE then issues those commands to their CPC.
Note: The new Driver level for HMC and SE for IBM z16 is Driver 51. Driver 51 is
equivalent to Version 2.16.0.
With HMC and SE Driver 41/Version 2.15.0, several “traditional” SE-only functions were
moved to HMC tasks. On HMC and SE Driver 41/Version 2.15.0 and Driver 51/Version 2.16.0,
these functions appear as native HMC tasks, but run on the SE. These HMC functions run in
parallel with Single Object Operations (SOOs), which simplifies and streamlines system
management.
For more information about SOOs, see “Single Object Operations” on page 435.
For more information about the HMC, see the IBM Z Hardware Management Console Videos.
Note: HMC 2.16.0 supports managing IBM z14, IBM z15, and IBM z15 IBM Z server
generations (N-2),
For more information about HMC and SE functions, use the HMC and SE console help
system or see IBM Resource Link (login required), select Library, the applicable server and
then, select Hardware Management Console Operations Guide or Support Element
Operations Guide.
Figure 10-3 shows how you can change the permission from Edit to View only for a task in
User Management.
A weekly schedule for health and diagnostic data (which is also sent to IBM Resource
Link) is available in a new window (see Figure 10-7). The active health checking feeds
data daily to IBM for an automatic analysis of your IBM Z for potential problems.
We highly recommend enabling the System availability collection on your IBM Z to get the
full RAS support.
Figure 10-8 New sender option for HMC Event email notifications
The notification about whether you are on an HMC with the Replica role and open a
window where Data Replication is involved were improved. An example of an HMC with
the Replica role is shown in Figure 10-10.
Figure 10-10 Notification for a task with relation to Data Replication on an HMC with Replica role
Also, the wizard that is used to manage the settings for Data Replication were improved.
Remote Code Load
The Remote Code Load (RCL) for IBM Z allows IBM to upgrade a machine remotely by
working with the customer to schedule a date and time for the code load and monitor the
process to ensure that it completes successfully.
This feature allows you to schedule one or multiple Single Step Code Loads for an HMC or
SE.
On an IBM z16 system with the HMA feature, RCL can be scheduled to update both
HMCs. The firmware automatically manages Alternative SE switches to ensure that each
HMC can be updated without the Primary SE being present.
In general, you first must generate a token on the HMC task Manage Remote Firmware
Updates. Then, you can schedule an RCL on IBM Resource Link by selecting Fixes →
Licensed internal code → Remote Code Load request.
Note: For IBM Z, IBM Support personnel do not connect to your IBM Z from outside the
data center location to perform the RCL. Instead, the RCL is managed on your HMC.
The Figure 10-11 and Figure 10-12 on page 418 show examples of the automatic
firmware activation for Crypto and OSA adapters.
Figure 10-11 Crypto and OSA adapters selected for Update Firmware
The SE interface can be accessed from the HMC by using the Single Object Operation on the
HMC.
The HMA feature FC 0129 consists of the HMC code that is installed on the two 1U
rack-mounted servers on the top of the A frame, which are collocated with the SE code. The
servers are configured with processor, memory, storage, and networking resources to support
all processing and security requirements for running HMC and SE code.
The two HMCs (HMC1 and HMC2) from manufacturing (these names can be changed) are
configured as independent HMCs. They are not Primary or Alternative HMCs. HMC Data
Replication can be established, if wanted.
The SE code is running as a guest of the HMC. The two SE code instances are clustered for
high availability. One SE code runs the Primary SE the other Alternative SE. These SEs
perform data mirroring and their role can be switched for maintenance purposes.
Switching the Primary and Alternative SE roles is important because HMC microcode
maintenance can be performed only on the server that runs the Alternative SE as a guest.
If the HMC, which receives microcode updates, runs the Primary SE guest, SE switchover
must be performed. Figure 10-13 shows the HMA relation to the HMCs and SEs.
10.2.3 Rack Mount HMC and Tower HMC (carry forward only to IBM z16)
For IBM z16, stand-alone HMCs (Tower or Rack Mount) can no longer be ordered.
The following HMC feature codes can be carried forward from previous orders:
With these HMC feature codes, Driver 51/Version 2.16.0 can be installed to support IBM z16.
Also, Driver 51/Version 2.16.0 can be installed on the two HMCs that are provided with the
HMA feature FC 0100 on IBM z15.
You also can use the IBM z16 with previous ordered HMCs and no HMA. It is possible to
order the HMA FC 0129 later as an MES feature. It then provides new microcode, and no new
hardware.
The HMCs can be accessed by using a remote web browser and can manage N-2
generations systems (IBM z16 A01, IBM z15 T02, IBM z15 T01, IBM z14 ZR1, and IBM z14
M0x).
Important: With the IBM HMA, shutdown or restart of the HMC that includes the Primary
SE code as guest also restarts the Primary SE code. Only an application restart of the
HMC is not disruptive to the guest SE code.
For more information about the mini-KMM and how to attach it to the A frame, see 3931
Installation Manual, GC28-7017.
Microcode load
Microcode can be loaded by using the following options:
USB
If the HMC and SE code is included with a USB drive when a new system is ordered, the
load procedure is similar that which was used with a DVD.
Electronic
If USB load is not allowed, or if FC 0846 (No physical media options is ordered, an ISO
image is used for a firmware load over a local area network (LAN). The ISO image can be
downloaded through zRSF or an FTP (/FTPS/SFTP) server that is accessible by the LAN.
Important: The ISO image server must be in the same IP subnet with the target system to
load the HMC or SE ISO.
10.2.6 SE driver and version support with the HMC Driver 51/Version 2.16.0
The driver of the HMC and SE is equivalent to a specific HMC and SE version:
Driver 36 is equivalent to Version 2.14.1
Driver 41 is equivalent to Version 2.15.0
Driver 51 is equivalent to version 2.16.0
An HMC with Driver 51/Version 2.16.0 supports N-2 IBM Z server generations. Some
functions that are available on Driver 51/Version 2.16.0 and later are supported only when the
HMC is connected to an IBM Z with Driver 51/Version 2.16.0.
The SE drivers and versions that are supported by the HMC Driver 51/Version 2.16.0 are
listed in Table 10-1.
The following HMC feature codes can be carried forward from previous systems:
Tower FC 0082
Tower FC 0062
1U Rack Mount FC 0083
1U Rack Mount FC 0063
With these HMC feature codes, Driver 51/Version 2.16.0 can be installed to support IBM z16.
On the HMCs of the HMA for IBM z15, FC 0100, Driver 51/Version 2.16.0 also can be
installed.
For more information about the HMC settings that are related to access and security, see
IBM Resource Link. At the web page, select Library, the applicable server and then, select
Hardware Management Console Operations Guide or Support Element Operations
Guide.
With IBM z16 and HMA, the SE code runs virtualized on the two integrated HMCs on the two
integrated 1U rack-mounted servers on the top of the IBM z16 A frame. One SE is the
Primary SE (active) and the other is the Alternative SE (backup).
With these HMC feature codes, Driver 51/Version 2.16.0 can be installed to support IBM z16.
You can use the IBM z16 with previously ordered HMCs and no HMA. It is possible to order
the HMA FC 0129 later as MES.
The HMC communicates with the SE is through a customer-supplied Ethernet switch (two
switches are recommended for redundancy).
Each SE and each HMC has two Ethernet RJ45 ports. An example of these connections are
made without an HMA is shown in Figure 10-18.
Note: The HMC must be connected to the SEs by using a customer-provided switch. Direct
connection between the HMC and the SEs is not supported.
The connectivity for multiple CPC generations and mixing stand-alone HMCs and HMAs
environments (IBM z16 N-2 only) is shown in Figure 10-19.
Various methods are available for setting up the network. Designing and planning the HMC
and SE connectivity is the customers’ responsibility, based on the environment’s connectivity
and security requirements.
The following functions are examples that depend on the HMC connectivity:
Lightweight Directory Access Protocol (LDAP) support, which can be used for HMC user
authentication
Network Time Protocol (NTP)
RSF through broadband
HMC remote access and HMC Mobile
RSA SecurID support
MFA with Time-based One Time Password (TOTP)
FTPS is based on Secure Sockets Layer (SSL) cryptographic protocol and requires
certificates to authenticate the servers. SFTP is based on Secure Shell protocol (SSH) and
requires SSH keys to authenticate the servers. Certificates and key pairs are hosted on the
IBM z16 HMC.
The file transfer server choices for HMC are shown in Figure 10-20.
After the HMC console completes all FTP operations, the HMC console performs the FTP
operation on the SE’s behalf and returns the results. The IBM Z platform must be managed by
at least one HMC to allow FTP operations.
To allow greater flexibility in password selection, the password limit was increased to 64
characters and special characters are allowed for IBM z16.
For more information about HMC networks, see the following resources:
The HMC and SE Driver 51/Version 2.16.0 console help system
IBM Resource Link. At this web page, select Library, the applicable server and then,
select Hardware Management Console Operations Guide or Support Element
Operations Guide.
IBM Z 3931 Installation Manual for Physical Planning, GC28-7015.
Ethernet switches
Ethernet switches for HMC and SE connectivity are provided by the customer. Existing
supported switches can still be used.
The HMC uses IPv4 and IPv6 multi-casting1 to automatically discover the SEs. The HMC
Network Diagnostic Information task can be used to identify the IP addresses (IPv4 and
IPv6), which are used by the HMC to communicate to the SEs (of a CPC).
IPv6 addresses are easily identified. A fully qualified IPV6 address features 16 bytes. It is
written as eight 16-bit hexadecimal blocks that are separated by colons, as shown in the
following example:
2001:0db8:0000:0000:0202:b3ff:fe1e:8329
Because many IPv6 addresses are not fully qualified, shorthand notation can be used. In
shorthand notation, the leading zeros can be omitted, and a series of consecutive zeros can
be replaced with a double colon. The address in the previous example also can be written in
the following manner:
2001:db8::202:b3ff:fe1e:8329
If an IPv6 address is assigned to the HMC for remote operations that use a web browser,
browse to it by specifying that address. The address must be surrounded with square
brackets in the browser’s address field, as shown in the following example:
https://[fdab:1b89:fc07:1:201:6cff:fe72:ba7c]
1 For the customer-supplied switch, multicast must be enabled at the switch level.
The MFA first factor is the combination of login ID and password; the second factor is TOTP
(Time-based One-Time Password) that is sent to your smartphone, desktop, or application
(for example, Google Authenticator or IBM Verify). This TOTP is defined in RFC 6238
standard and uses a cryptographic hash function that combines a secret key with the current
time to generate a one-time password.
The secret key is generated by HMC/SE/TKE while the user is performing first factor log-on.
The secret key is known only to HMC/SE/TKE and to the user’s device. For that reason, it
must be protected as much as your first factor password.
The algorithm within the HMC that is responsible for MFA code generation changes the code
every 30 seconds. However, to ease the process, the HMC and SE console accepts current,
previous, and next MFA codes.
It is also important to have HMC, SE, and device clocks synchronized. If the clocks are not
synchronized, the MFA log-on attempt fails. Time zone differences are irrelevant because the
MFA code algorithm uses UTC.
On IBM z15, HMC Driver 41/Version 2.15.0 provides the integration of HMC authentication
and z/OS MFA support. Therefore, RSA SecurID authentication is achieved by way of
centralized support from IBM MFA for z/OS, with the MFA policy defined in RACF and the
HMC IDs assigned to RACF user IDs. The RSA SecurID passcode (from an RSA SecurID
Token) is verified by the RSA authentication server. This authentication is supported on HMC
only, not on the SE.
The following support was added with IBM z16 Driver 51/Version 2.16.0:
Enhanced MFA functions
MFA by way of Time-based One-time Password (TOTP) or IBM Z MFA (z/OS) and RSA
Secure ID is possible on the HMC.
New with Driver 51/Version 2.16.0 the following further MFA possibilities are supported:
– Certificates:
• Personal Identity Verification (PIV)
• Common Access Card (CAC)
• Certificates on USB keys
– Generic Remote Authentication Dia-In User Service (RADIUS) allows for support of all
various RADIUS factor types. Involves customer provided RADIUS server.
Also, Driver 51/Version 2.16.0 provides support of IBM Z MFA for Red Hat Enterprise Linux
Server or SUSE Linux Enterprise Server that is running on z/VM or native in an LPAR.
For more information about the benefits of Broadband RSF and the SSL/TLS-secured
protocol, and a sample configuration for the Broadband RSF connection, see Integrating the
HMC Broadband Remote Support Facility into Your Enterprise, SC28-7026.
10.4.2 RSF connections to IBM and Enhanced IBM Service Support System
To have the best availability and redundancy and to be prepared for the future, the HMC must
access IBM by using the internet through RSF in the following manner: Transmission to the
enhanced IBM Support System requires a domain name server (DNS). The DNS must be
configured on the HMC if a proxy for RSF is not used. If a proxy for RSF is used, the proxy
must provide the DNS.
The following hostnames and IP addresses are used and your network infrastructure must
allow the HMC to access RSF:
esupport.ibm.com on port 443
The use of IPv4 requires outbound connectivity to the following IP addresses:
– 129.42.54.189
– 129.42.56.189
– 129.42.60.189
The use of IPv6 requires outbound connectivity to the following IP addresses:
– 2620:0:6c0:200:129:42:54:189
– 2620:0:6c2:200:129:42:56:189
– 2620:0:6c4:200:129:42:60:189
For more information about these capabilities, see the HMC and SE Driver 51/Version 2.16.0
console help system or see the IBM Resource Link. At this web page, select Library, the
applicable server and then, select Hardware Management Console Operations Guide or
Support Element Operations Guide.
With the introduction of the DPM mode mainly for LinuxONE management, the user interface
and user interaction with the HMC changed significantly, although the capabilities are
remained the same. The figures and descriptions in this section cover only the traditional
Processor Resource/Systems Manager (PR/SM) mode.
The HMC is used to start the power-on reset (POR) of the system. During the POR,
processor units (PUs) are characterized and placed into their respective pools, memory is put
into a single storage pool, and the IOCDS is loaded into and started in the hardware system
area (HSA).
The hardware messages task displays hardware-related messages at the CPC, LPAR, or SE
level. It also displays hardware messages that relate to the HMC.
Because PR/SM must manage LPAR access to processors and the initial weights of each
partition, weights are used to prioritize partition access to processors.
You can use the Load task on the HMC to perform an IPL of an operating system. This task
causes a program to be read from a designated device, and starts that program. You can
perform the IPL of the operating system from storage, the USB flash memory drive (UFD), or
an FTP server.
When an LPAR is active and an operating system is running in it, you can use the HMC to
dynamically change certain LPAR parameters. The HMC provides an interface to change
partition weights, add logical processors to partitions, and add memory.
LPAR weights also can be changed through a scheduled operation. Use the Customize
Scheduled Operations task to define the weights that are set to LPARs at the scheduled time.
Channel paths can be dynamically configured on and off (as needed for each partition) from
an HMC.
The Change LPAR Controls task for IBM z16 can export the Change LPAR Controls table
data to a comma-separated value (.csv)-formatted file. This support is available to a user
when they are remotely connected to the HMC by a web browser.
Partition capping values can be scheduled and are specified on the Change LPAR Controls
scheduled operation support. Viewing more information about a Change LPAR Controls
scheduled operation is available on the SE.
One example of managing the LPAR settings is the absolute physical hardware LPAR
capacity setting. Driver 15 (zEC12/zBC12) introduced the capability to define (in the image
profile for shared processors) the absolute processor capacity that the image is allowed to
use (independent of the image weight or other cappings).
Following on to LPAR absolute capping, LPAR group absolute capping uses a similar method
to enforce the following components:
Customer licensing
Non-z/OS partitions where group soft capping is not an option
z/OS partitions where ISV does not support software capping
A group name, processor capping value, and partition membership are specified at the
hardware console, along with the following properties:
Set an absolute capacity cap by CPU type on a group of LPARs.
Allow each of the partitions to use capacity up to their individual limits if the group’s
aggregate consumption does not exceed the group absolute capacity limit.
Include updated SysEvent QVS support (used by vendors who implement software
pricing).
Only shared partitions are managed in these groups.
Specify caps for one or more processor types in the group.
Specify in absolute processor capacity (for example, 2.5 processors).
Use Change LPAR Group Controls (as with windows that are used for software
group-defined capacity), as shown in Figure 10-22.
The value is not tied to the Licensed Internal Code (LIC) configuration code (LICCC). Any
value 0.01 - 255.00 can be specified. This configuration makes the profiles more portable,
Although the absolute cap can be specified to hundredths of a processor, the exact amount
might not be that precise. The same factors that influence the “machine capacity” also
influence the precision with which the absolute capping works.
The following HMC feature codes can be carried forward from previous orders:
FC 0062
FC 0063
FC 0082
FC 0083
With these HMC feature codes, Driver 51/Version 2.16.0 can be installed to support IBM z16.
Also, Driver 51/Version 2.16.0 can be installed on the two HMCs that are provided with the
HMA FC 0100 on IBM z15. With these HMCs, you can still work local at an HMC.
Note: Remote web browser access is the default for the HMA HMCs.
Access to the USB Device on HMC and SE requires physical access to the HMC/SE.
Log-on security for a web browser is provided by the local HMC user log-on procedures.
Certificates for secure communications are provided and can be changed by the user.
Web browser access can be limited by specifying an IP address from the Customize Console
Services task. To enable or disable the Remote operation service, click Change... in the
Customize Console Services window, as shown in Figure 10-23.
Note: If the setting in Change Remote Access Setting → IP Access Control is set to
Allow specific IP addresses, but none or incorrect IP addresses are in the list, a remote
HMC connection is not available by using a web browser.
Microsoft Internet Explorer, Microsoft Edge, Mozilla Firefox, Safari, and Goggle Chrome were
tested as remote web browsers. For more information about web browser requirements, see
the HMC and SE console help system or see the IBM Resource Link. At this web page, select
Library, the applicable server and then, select Hardware Management Console
Operations Guide or Support Element Operations Guide.
Note: With HMC Driver 41/Version 2.15.0 and Driver 51/Version 2.16.0, specific tasks that
required in the past to access to the SE in SOO mode were implemented as HMC tasks.
With this enhancement, the HMC runs the tasks on the SE directly, without the need to lock
the SE in SOO mode.
You also can start, stop, or change the activation profile for a partition.
A full set of granular security controls is provided from the HMC, including MFA. This mobile
interface is optional and is disabled by default. More functions from the HMC also are planned
to be available on IBM HMC Mobile.
The HMC also provides integrated 3270 and ASCII consoles. These consoles allow an
operating system to be accessed without requiring other network or network devices, such as
TCP/IP or control units.
Use the Certificate Management feature if the certificates that are returned by the 3270
server are not signed by a well-known trusted certificate authority (CA) certificate, such as
VeriSign or Geotrust. An advanced action within the Certificate Management task, Manage
Trusted Signing Certificates, is used to add trusted signing certificates.
For example, if the certificate that is associated with the 3270 server on the IBM host is
signed and issued by a corporate certificate, it must be imported, as shown in Figure 10-24.
The import from the remote server option can be used if the connection between the console
and the IBM host can be trusted when the certificate is imported, as shown in Figure 10-25.
Otherwise, import the certificate by using removable media.
When a driver is upgraded, always check the Driver (51) Customer Exception Letter option in
the Fixes section at the IBM Resource Link.
Tip: The IBM Resource Link provides access to the system information for your IBM Z
system according to the system availability data that is sent on a scheduled basis. It
provides more information about the MCL status of your IBM Z. At the IBM Resource Link
web page, click Tools → Machine Information, select your IBM Z system and then, click
EC/MCL.
Microcode terms
The microcode features the following characteristics:
The driver contains engineering change (EC) streams.
Each EC stream covers the code for a specific component of IBM z16. It includes a
specific name and an ascending number.
The EC stream name and a specific number are one MCL.
MCLs from the same EC stream must be installed in sequence.
MCLs can include installation dependencies on other MCLs.
Combined MCLs from one or more EC streams are in one Bundle.
An MCL contains one or more Microcode Fixes (MCFs).
Driver / Version
bundle
Bundle
EC stream
MCL
MCF
MCF
MCF
MCF
Bundle
EC stream
MCL
MCF
MCF
MCF
MCF
EC stream
MCL
MCF
MCF
MCF
MCF
MCL EC stream
MCL
MCF
MCF
MCL MCF
MCF
MCF MCF
MCF
MCF MCF
MCF
MCF
MCF
EC stream
MCL
MCF
MCL MCF
MCF
MCF
MCF
MCF
MCF
MCF
bundle
Bundle bundle
Bundle
EC stream EC stream
MCL MCL
MCF MCF
MCF MCF
MCF MCF
MCF MCF
MCL Activation
By design and with planning, MCLs can be activated concurrently. Consider the following
points:
Most MCLs activate concurrently when applied.
Some MCLs need a disruptive configure off/on to activate the new loaded microcode.
Activate traditional I/O Feature Pended MCL – LIC on the hardware feature:
– Display Pending MCLs by using HMC function or Resource Link Machine Information
Reports
– Activate by using HMC function on a feature basis by PCHID individually – disruptive:
CONFIG the CHPID OFF to all sharing LPARs, activate, and then CONFIG ON to all
Activate Native PCIe Pended MCL – LIC on a hardware feature OR Resource Group (RG)
LIC:
– Display Pending MCLs by using HMC function or Resource Link Machine Information
Reports
Note: For hardware that does not need CHPID or a FUNCTION definition (for example,
Crypto Express), a different method that is specific to the feature is used.
Alternative: Apply and activate all Pended MCLs disruptively with a scheduled Power On
Reset (POR)
To discover this “Pended” situation, the following actions are completed whenever an MCL is
applied:
Log on to the HMC and select System Management → CPC
Change Management
System Information
Query Additional Actions
Or:
Log on to the SE and select System Management → CPC
Change Management
Query Channel/Crypto Configure Off/On Pending
The System Information window is enhanced to show a summary Bundle level for the
activated level, as shown in Figure 10-28.
Monitors Dashboard
The Monitors Dashboard in the Monitor group provides a tree-based view of resources.
Multiple graphical views are available for displaying data, including history charts. The
Monitors Dashboard monitors processor and channel usage. It also produces data that
includes power monitoring information, power consumption, and the air input temperature for
the system.
The data is presented in table format and graphical “histogram” format. The data also can be
exported to a .csv-formatted file so that the data can be imported into a spreadsheet. For this
task, you must use a web browser to connect to an HMC.
An example of the settings for Event Monitor Summary can be see on Figure 10-30.
The HMC for IBM z16 features the following CoD capabilities:
SNMP API support:
– API interfaces for granular activation and deactivation
– API interfaces for enhanced CoD query information
– API event notification for any CoD change activity on the system
– CoD API interfaces, such as On/Off CoD and Capacity Back Up (CBU)
SE window features (accessed through HMC Single Object Operations):
– Window controls for granular activation and deactivation
– History window for all CoD actions
– Description editing of CoD records
HMC/SE provides the following CoD information:
– Millions of service units (MSU) and processor tokens
– Last activation time
– Pending resources that are shown by processor type instead of only a total count
– Option to show more information about installed and staged permanent records
– More information for the Attention state by providing seven more flags
HMC and SE are a part of the z/OS Capacity Provisioning environment. The Capacity
Provisioning Manager (CPM) communicates with the HMC through IBM Z APIs, and enters
CoD requests. For this reason, SNMP must be configured and enabled by using the
Customize API Settings task on the HMC.
For more information about the use of and setting up CPM, see the following publications:
z/OS MVS Capacity Provisioning User’s Guide, SC34-2661
Capacity on-Demand User’s Guide, SC28-7025
Important: The Sysplex Time task on the SE was discontinued with IBM z15. Therefore,
an HMC with Driver 51/Version 2.16.0 is required to access the Manage System Time task
for IBM z16 A01, IBM z15 T01, and IBM z15 T02 CPCs.
IBM z16 implements the following major enhancements in support of the Server Time
Protocol function:
CPC direct Ethernet connectivity for the external time source (ETS)
In previous generation IBM Z, the ETS for the STP was provided by connecting the
Support element to the client network.
With IBM z16 the ETS, PTP (IEEE 1588) or NTP network connectivity is provided by using
the IBM z16 CPC oscillator (OSC) cards dedicated network ports (RJ45) to client LAN for
accessing the time synchronization information.
Pulse-per-second connectivity also is provided for the highest timing information accuracy.
Connection of the ETS direct to the IBM Z CPC provides less delay in accessing the time
source that connection through the Support Element.
n-mode Power STP Imminent Disruption Signal option
On IBM Z, losing a Preferred Time Server (PTS) has significant consequences to the
timing network and the overall workload execution environment of the IBM Z sysplex. The
IBM Z and the HMC feature longtime automated failover protection for various cases that
can arise.
New for IBM z16 because an integrated battery facility no long is available, support was
added by the HMC to allow the customer to configure an option to monitor for n-mode
power conditions (wall power or power cord loss). If detected, an automated failover
occurs to the Backup Time Server (BTS).
Note: Provide some backup power method to hold power for 60 seconds on the PTS to
allow failover to successfully complete.
Manage System Time user interface controls are available to manage failback to the PTS
when the full power state is restored.
ETS direct connectivity to the IBM z16 CPC is provided for both supported ETS protocols:
NTP and PTP.
An STP-only CTN can be managed by using different HMCs. However, the HMC must be at
the same driver level (or later) than any SE that is to be managed. Also, all SEs to be
managed must be known (defined) to that HMC.In a STP-only CTN, the HMC can be used to
perform the following tasks:
Start or modify the CTN ID.
Start the time (manually or by contacting an NTP server).
Start the time zone offset, Daylight Saving Time offset, and leap second offset.
Assign the roles of preferred, backup, and current time servers, and arbiter.
Adjust time by up to plus or minus 60 seconds.
Schedule changes to the offsets listed. STP can automatically schedule Daylight Saving
Time, based on the selected time zone.
Monitor the status of the CTN.
Monitor the status of the coupling links that are started for STP message exchanges.
For diagnostic purposes, the PPS port state on an IBM z16 can be displayed and fenced
ports can be reset individually.
Important: The Sysplex Time task on the SE was discontinued for IBM z15. Therefore, an
HMC at Version 2.15.0 is required to manage system time for IBM z15 CPCs.
Detailed instructions and guidelines are provided within task workflow. IBM z15 HMC
provides a visual representation of the CTN topology. A preview of any configuration action is
also shown in topological display. An example of the topology view is shown in Figure 10-31.
For STP, if CPC1 (Primary Time Server and Current Time Server) senses a shift to an
on-battery power condition, a signal indicating potential imminent server disruption of CPC1
is sent to CPC2 (the Backup Time Server).
If within 30 seconds the BTS (CPC2) does not receive a signal that power is back to fully
redundant on CPC1, CPC2 takes over as CTS. After normal power is restored to CPC1,
CPC1 can automatically return to the CTS role.
With IBM z16, the IBF was discontinued. As such, Coupling Facility Nonvolatility requires
Room-level, Row-level, or System-Level UPS for protecting against power failures.
New STP PTS/BTS methods enable better coverage across both BPA and iPDU
configurations for a much broader scope of clients that can take advantage of STP
high-reliability functions. The new strategy allows clients with UPS to more effectively monitor
and manage their power efficiency across their data center.
STP resiliency for PTS/CTS transition IBM Z from IBF Sensing to N-Mode Power Sensing.
To enable the N-Mode Power Sensing, a one-time setup step is required. In the HMC Manage
System Time task, a customer enables the automatic switch over function from CPC1 to
CPC2 for STP. After the function is enabled, the power subsystem (BPA and iPDU) detects
any power source loss (at the power cord or power side level).
If within 30 seconds CPC2 does not receive a signal that power is back to fully redundant on
CPC1, CPC2 takes over as CTS. After normal power is restored to CPC1, CPC1 can
automatically return to the CTS role.
ECAR support is faster than the original CAR support because the console path changes
from a 2-way path to a 1-way path. Also, almost no lag time is incurred between the system
checkstop and the start of CAR processing. Because the request is generated from the PTS
before system logging, it avoids the potential of recovery being held up.
Attention: IBM z16, IBM z15, and IBM z14 ZR1 do not support InfiniBand connectivity.
Although IBM z14 M0x supports InfiniBand, InfiniBand coupling and timing links cannot be
configured in a Syspex/CTN with IBM z16.
For more information about planning and setup, see IBM Z Server Time Protocol Guide,
SG24-8480.
For more information, see IBM Z Server Time Protocol Guide, SG24-8480.
CTN split
The HMC menus for Server Time Protocol (STP) provide support when one or more systems
must be split in to a separate CTN without interruption in the clock source.
The task is available under the Advanced Actions menu in the Manage System Time task.
Several checks are performed to avoid potential disruptive actions. If targeted CTN includes
only members with the roles, the task start fails with an error message.
If targeted CTN includes at least one system without any roles, the task starts. An
informational warning is presented to the user to acknowledge that sysplex workloads are
divided suitably.
During the transition state, most of the STP actions for the two affected CTNs are disabled.
After the merge is completed, STP actions are enabled again.
For more information about planning and understanding STP server roles, see the following
publications:
IBM Z Server Time Protocol Guide, SG24-8480
Server Time Protocol Planning Guide, SG24-7280
Server Time Protocol Implementation Guide, SG24-7281
Server Time Protocol Recovery Guide, SG24-7380
The NTP server becomes the single time source (the ETS) for STP and other servers that are
not IBM Z (such as AIX® and Microsoft Windows) that include NTP clients.
The HMC can act as an NTP server. With this support, the IBM z15 can receive the time from
the HMC without accessing a LAN other than the HMC and SE network. When the HMC is
used as an NTP server, it can be configured to receive the NTP source from the internet. For
this type of configuration, a LAN that is separate from the HMC/SE LAN can be used.
Note: For IBM z16, STP External Time Source (Ethernet) is connected directly to the CPC
drawer by using a dedicated LAN port. When HMC is configured as an NTP server, it must
be connected to the same client network with the dedicated CPC LAN port.
The HMC offers the following symmetric key and autokey authentication and NTP commands:
Symmetric key (NTP V3-V4) authentication
Symmetric key authentication is described in RFC 1305, which was made available in
NTP Version 3. Symmetric key encryption uses the same key for encryption and
decryption. Users that are exchanging data keep this key to themselves.
Messages that are encrypted with a secret key can be decrypted only with the same
secret key. Symmetric key authentication supports network address translation (NAT).
Symmetric key autokey (NTP V4) authentication
This autokey uses public key cryptography, as described in RFC 5906, which was made
available in NTP Version 4. You can generate keys for the HMC NTP by clicking Generate
Local Host Key in the Autokey Configuration window. This option issues the ntp-keygen
command to generate the specific key and certificate for this system. Autokey
authentication is not available with the NAT firewall.
Issue NTP commands
NTP command support is added to display the status of remote NTP servers and the
current NTP server (HMC).
For more information about planning and setup for STP and NTP, see the following
publications:
IBM Z Server Time Protocol Guide, SG24-8480
Server Time Protocol Planning Guide, SG24-7280
Server Time Protocol Implementation Guide, SG24-7281
Server Time Protocol Recovery Guide, SG24-7380
For HMC and SE Driver 51/Version 2.16.0, HDD encryption uses Trusted Platform Module
(TPM) and Linux Unified Key Setup (LUKS) technology.
With IBM z16, you can offload the following HMC and SE log files for customer audit:
Console event log
Console service history
Tasks performed log
Security logs
System log
Full log offload and delta log offload (since the last offload request) are provided. Offloading
to removable media and to remote locations by FTP is available. The offloading can be
manually started by the new Audit and Log Management task or scheduled by the Customize
Scheduled Operations task. The data can be offloaded in the HTML and XML formats.
Each HMC user ID template defines the specific authorization levels for the tasks and objects
for the user who is mapped to that template. The HMC user is mapped to a specific user ID
template by user ID pattern matching. The system then obtains the name of the user ID
template from content in the LDAP server schema data.
Any default user IDs that are part of a previous HMC level can be carried forward to new HMC
levels as part of a MES Upgrade or by way of selecting User Profile Data for the Save/Restore
Customizable Console Data or Configure Data Replication tasks.
The Secure FTP infrastructure allows HMC and SE applications to query whether a public key
is associated with a host address and to use the Secure FTP interface with the suitable public
key for a host. Tasks that use FTP now provide a selection for the secure host connection.
When selected, the task verifies that a public key is associated with the specified hostname. If
a public key is not provided, a message window opens that points to the Manage SSH Keys
task to enter a public key. The following tasks provide this support:
Import/Export IOCDS
Advanced Facilities FTP IBM Content Collector Load
Audit and Log Management (Scheduled Operations only)
FCP Configuration Import/Export
The information that is needed to manage a system’s I/O configuration must be obtained from
many separate sources. The System Input/Output Configuration Analyzer task enables the
system hardware administrator to access, from one location, the information from those
sources. Managing I/O configurations then becomes easier, particularly across multiple
systems.
The System Input/Output Configuration Analyzer is a view-only tool. It does not offer any
options other than viewing. The tool provides various sort options, and data can be exported
to a UFD for later viewing.
By using the tool, data is formatted and displayed in the following views:
PCHID Control Unit View
Shows PCHIDs, channel subsystems (CSS), CHPIDs, and their control units.
PCHID Partition View
Shows PCHIDS, CSS, CHPIDs, and the partitions in which they exist.
Control Unit View
Shows the control units, their PCHIDs, and their link addresses in each CSS.
Link Load View
Shows the Link address and the PCHIDs that use it.
Node ID View
Shows the Node ID data under the PCHIDs.
On IBM z16, the HMC APIs provide monitoring and control functions through SNMP. The API
can get and set a managed object’s attributes, issue commands, receive asynchronous
notifications, and generate SNMP traps.
The older system, such as IBM z13’s HMC, supports the CIM as an extra systems
management API. Starting with IBM z14, the CIM support is removed.
Cryptographic hardware
IBM z16 systems include standard cryptographic hardware and optional cryptographic
features for flexibility and growth capability.
The Crypto Express8S, which is a new Peripheral Component Interconnect Express (PCIe)
cryptographic coprocessor, is an optional IBM z16 exclusive feature. Crypto Express8S
provides a secure programming and hardware environment on which crypto processes are
run.
Each Crypto Express8S adapter can be configured by the installation as a Secure IBM CCA
coprocessor, a Secure IBM Enterprise Public Key Cryptography Standards (PKCS) #11
(EP11) coprocessor, or an accelerator.
When EP11 mode is selected, a unique Enterprise PKCS #11 firmware is loaded into the
cryptographic coprocessor. It is separate from the Common Cryptographic Architecture
(CCA) firmware that is loaded when a CCA coprocessor is selected. CCA firmware and
PKCS #11 firmware cannot coexist in a card.
The Trusted Key Entry (TKE) Workstation with smart card reader feature is required to
support the administration of the Crypto Express8S when configured as an Enterprise
PKCS #11 coprocessor.
The TKE10.0 is needed to support the new Crypto Express8S card. An example of the
Cryptographic Configuration window is shown in Figure 10-32.
The Usage Domain Zeroize feature is provided to clear the suitable partition crypto keys for a
usage domain when you remove a crypto card from a partition. Crypto Express8/7/6S in
EP11 mode is configured to the standby state after the Zeroize process.
For more information, see IBM z16 (3931) Configuration Setup, SG24-8960.
This operation ensures that any changes that are made to the data are detected during the
upgrade process by verifying the digital signature. It helps ensure that no malware can be
installed on IBM Z products during firmware updates. It also enables the IBM z16 Central
Processor Assist for Cryptographic Function (CPACF) functions to comply with Federal
Information Processing Standard (FIPS) 140-3 Level 1 planned for Cryptographic LIC
changes. The enhancement follows the IBM Z focus of security for the HMC and the SE.
The Crypto Express8S (CEX8S) is compliant with CCA PCI HSM. TKE workstation is optional
when used to manage a Crypto Express8S feature that is defined as a CCA coprocessor in
normal mode.
A system can be configured in DPM mode or in PR/SM mode (POR is required to switch
modes). In general, DPM supports the following functions:
Create, provision, and manage partitions (processor, memory, and adapters) and storage
Monitor and troubleshoot the environment
If DPM is enabled, the IBM z16 system cannot run z/OS, IBM z/VSE, and z/TPF LPARs.
The IBM z16 can be initialized in PR/SM mode or in DPM mode, but not both.
DPM provides a GUI for PR/SM (to manage resources). Tools, such as HCD are in DPM
mode and are not necessary.
This IBM Redbooks publication does not cover scenarios that use DPM.
For more information about the use of DPM, see IBM Dynamic Partition Manager (DPM)
Guide, SB10-7182.
Naming: Throughout this chapter, IBM z16 refers to IBM z16 Model A01 (Machine Type
3931), unless otherwise specified.
Removal of support for Bulk Power Assembly (BPA)a: IBM z16 is planned to be the last
generation of IBM Z server to support BPA.
a. Statements by IBM regarding its plans, directions, and intent are subject to change or
withdrawal without notice at the sole discretion of IBM. Information regarding potential future
products is intended to outline general product direction and should not be relied on in making
a purchasing decision.
For more information about physical planning, see 3931 Installation Manual for Physical
Planning, GC28-7015.
IBM z16 servers support installation on a raised floor or nonraised floor and are available in
the following power and cooling options:
Intelligent Power Distribution Unit-based power (iPDU) or PDU
Bulk Power Assembly-based power (BPA)
All IBM z16 models include radiator-based cooling (air cooled system).
A PDU-based system can have 2 - 8 power cords, depending on the configuration. The use of
iPDU on IBM z16 might enable fewer frames, which allows for more I/O slots to be available
and improves power efficiency to lower overall energy costs. It also offers some
standardization and ease of data center installation planning, which allows the IBM z16 to
easily coexist with other platforms within the data center.
The second, third, and fourth PDU pairs are installed dependent on other CPC or PCIe+ I/O
drawers installed. The locations of the PDU pairs and frames are listed in Table 11-1.
Power cords for the PDUs are attached to the options that are listed in Table 11-2.
A rear view of a maximum configured PDU-powered system with four CPC drawers and 12
PCIe+ I/O drawers is shown in Figure 11-1 on page 456.
The number of PDUs and power cords that are required based on the number of CPC
drawers and PCIe+ I/O drawers are in Table 11-3.
1 2 2 2 2 6 6 6 6
2 4 4 4 4 4 4 4 4 6 6 6 6 6
3 4 4 4 4 4 4 4 6 6 6 6 6 N/A
4 6 6 6 6 6 6 6 6 6 6 8 8 8
Power consumption
The utility power consumption for the IBM z16 for PDU option is listed in Table 11-4.
max82 8.8 9.8 10.8 11.7 12.6 13.5 14.5 15.4 16.4 17.3 18.2 19.1 20.0
FC0668 kw kw kw kw kw kw kw kw kw kw kw kw kw
max125 12.7 13.7 14.6 15.6 16.5 17.4 18.3 19.3 20.2 21.1 22.0 22.9
N/A
FC0669 kw kw kw kw kw kw kw kw kw kw kw kw
max168 13.5 14.5 15.6 16.6 17.7 18.7 19.8 20.8 21.9 22.9 24.0 25.0 25.8
FC0670 kw kw kw kw kw kw kw kw kw kw kw kw kw
max200 17.1 18.1 19.0 20.0 20.9 21.8 22.7 23.6 24.5 25.4 26.3 27.2 28.1
FC0671 kw kw kw kw kw kw kw kw kw kw kw kw kw
Power estimation for any configuration, power source, and room condition can be obtained
by using the power estimation tool that is available at the IBM Resource Link website (login
required).
On the Resource Link page, click Tools → Power and weight estimation.
Removal of IBF support: IBM z16 cannot be configured with an internal battery feature
(IBM). Extra support for data center power planning can be requested through your IBM
Sales contact.
The BPA option is required for customers who order an Internal Battery Feature (IBF), Water
Cooling Unit (WCU), or Balanced Power. The BPA requires two or four power cords. All BPAs
are concurrently repairable in the field.
If power is interrupted to one of the power supplies, the other power supply assumes the
entire load and the system continues to operate without interruption. Therefore, the power
cords for each power supply must be wired to support the entire power load of the system.
For the most reliable availability, the power cords in the rear of the frame should be powered
from different PDUs. All power cords exit through the rear of the frame. The utility current
distribution across the phase conductors (phase current balance) depends on the system
configuration. Each front-end power supply is provided with phase switching redundancy.
The loss of an input phase is detected and the total input current is switched to the remaining
phase pair without any power interruption. Depending on the configuration input power draw,
the system can run from several minutes to indefinitely in this condition. Because most single
phase losses are transients that recover in seconds, this redundancy provides protection
against virtually all single phase outages.
Power cords for the BPAs are attached to the options that are listed in Table 11-5.
A rear view of a maximum configured BPA powered system with two BPA pairs, IBF, four CPC
drawers, and 10 PCIe+ I/O drawers is shown in Figure 11-2.
1 1 1 2 2 2 2 2 3 3 3 4
2 1 2 2 2 2 2 3 3 3 4 4
3 1 2 2 2 2 2 3 3 3 4 4
4 2 2 3 3 3 3 3 4 4 4 4
The number of power cords and number of Bulk Power Regulators that are required are
based on the number of CPC processor drawers and number of PCIe+ I/O drawers in the
configuration is shown in Table 11-8.
0 1 2 3 4 5 6
1 CPC 1 1 2 2 2 2 2
2 CPC 2 2 2 2 2 3 3
3 CPC 2 2 3 3 3 3 -
Note: Balanced power feature includes all BPRs that are plugged in all BPAs in the system.
0 1 2 3 4 5 6 7 8 9 10
max39 1 7.0 8.0 9.0 9.9 10.8 11.7 14.0 N/A N/A N/A N/A
FC0667 kw kw kw kw kw kw kw
max82 2 10.7 11.6 12.5 13.4 14.3 16.6 17.9 19.0 20.1 21.1 22.1
FC0668 kw kw kw kw kw kw kw kw kw kw kw
max125 3 14.1 15.1 16.1 17.1 18.0 20.4 21.7 18.7 23.9 24.9 25.9
FC0669 kw kw kw kw kw kw kw kw kw kw kw
max168 4 20.6 21.6 22.6 23.6 24.6 25.5 26.5 27.4 28.3 29.2 30.0
FC0670 kw kw kw kw kw kw kw kw kw kw kw
max200 4 20.6 21.6 22.6 23.6 24.6 25.5 26.5 27.4 28.3 29.2 30.0
FC0671 kw kw kw kw kw kw kw kw kw kw kw
Power estimation for any configuration, power source, and room condition can be obtained
by using the power estimation tool at IBM Resource Link website (authentication required).
On the Resource Link page, click Tools → Power and weight estimation.
For more information about cooling requirements, see 3931 Installation Manual for Physical
Planning, GC28-7015.
The radiator cooling system requires chilled air to fulfill the air-cooling requirements. IBM z16
system airflow is from the front (intake, chilled air) to the rear (exhausts, warm air) of the
frames. The chilled air is provided through perforated floor panels in front of the system
The hot and cold airflow and the arrangement of server aisles are shown in Figure 11-3.
As shown in Figure 11-3, rows of servers must be placed front-to-front. Chilled air is provided
through perforated floor panels that are placed in rows between the fronts of servers (the cold
aisles). Perforated tiles generally are not placed in the hot aisles.
If your computer room causes the temperature in the hot aisles to exceed a comfortable
temperature, add as many perforated tiles as necessary to create a satisfactory comfort level.
Heated exhaust air exits the computer room above the computing equipment.
For more information about the requirements for air-cooling options, see 3931 Installation
Manual for Physical Planning, GC28-7015.
The IBM z16 can be installed on a raised or nonraised floor. (For more information about
weight distribution and floor loading tables, see IBM 3931 Installation Manual for Physical
Planning, GC28-7015). This data is used with the maximum frame weight, frame width, and
frame depth to calculate the floor loading.
Weight estimates for the maximum system configurations on the 3931 PDU-based system
are listed in Figure 11-10 on page 469.
A - 795 kg - - 795 kg
(1753 lbs) (1753 lbs)
The power and weight estimation tool for IBM Z servers on IBM Resource Link (log in
required) covers the estimated weight for your designated configuration.
On the Resource Link web page, click Tools → Power and weight estimation.
Note: On the IBM z16, all I/O cabling and power cords enter the rear of the machine;
therefore, all related features for Bottom and Top Exit cabling are in the rear of the frame.
Raised floor
The IBM z16 server can installed in a raised floor environment. You can select top exit
features to manage I/O cables from the top frame of the IBM z16.
The following top exit options are available for z16 servers:
Top Exit I/O cabling without Tophat feature code (FC 7816)
Bottom Exit cabling feature code (FC 7899)
Top Exit Cabling with Tophat feature code (FC 7898)
The Top Exit cabling feature can be placed as shown in Figure 11-4, with the exit area toward
the front of the frame, or with the exit area toward the rear of the frame.
Note: The feature provides the same extra hardware for every frame in the configuration.
If the Top Exit cabling feature is not ordered, two sliding plates are available on the top of the
frame (one on each side of the rear of the frame) that can be partially opened. By opening
these plates, I/O cabling and power cords can exit the frame. The plates should be removed
to install the Top Exit cabling feature as shown in Figure 11-5.
All external cabling enters the rear of the rack from under floor or from above the rack.
Different from previous IBM Z, no cabling access or cable plugging is available at the front of
the rack. The top view of the rack with and without FC 7917 is shown in Figure 11-6.
The Top Exit Cabling option can be used for routing power cables and IO cables out the top of
the machine.
Without the Top Exit Cabling feature, power and cables still can be run out the top of the rack
through two adjustable openings at the top rear of the rack, as shown on the left side of
Figure 11-6 on page 465.
The Bottom Exit Cabling feature provides tailgate hardware for routing power cables or IO
cables out the bottom of the machine.
For more information, see IBM 8561 Installation Manual for Physical Planning, GC28-7015,
and 11.4, “Physical planning” on page 464.
The frame tie-down kit can be used on a nonraised floor (FC 8015) where the frame is
secured directly to a concrete floor, or on a raised floor (FC 8014) where the frame is secured
to the concrete floor underneath the raised floor.
Raised floors 241.3 mm (9.5 inches) - 1270 mm (50 inches) are supported.
For more information, see IBM 3931 Installation Manual for Physical Planning, GC28-7015.
For more information, see IBM 3931 Installation Manual for Physical Planning, GC28-7015.
The energy management structure for the server is shown in Figure 11-9.
The hardware components in the IBM z16 are monitored and managed by the energy
management component in the Support Element (SE) and HMC. The graphical user
interfaces (GUIs) of the SE and HMC provide views, such as the Monitors Dashboard,
Environmental Efficiency Statistics, and Energy Optimization Advisor.
The following tools are available to plan and monitor the energy consumption of IBM z16
servers:
Power estimation tool on Resource Link
Energy Optimization Advisor task for maximum potential power on HMC and SE
Monitors Dashboard and Environmental Efficiency Statistics tasks on HMC and SE
Select the advice hyperlinks to provide specific recommendations for your system, as shown
in Figure 11-10.
The data is presented in table format and graphical “histogram” format. The data also can be
exported to a .csv-formatted file so that the data can be imported into a spreadsheet. For this
task, you must use a web browser to connect to an HMC.
Note: Throughout this chapter, IBM z16 refers to IBM z16 Model A01 (Machine Type 3931)
unless otherwise specified.
Uniprocessor performance also increased. On average, an IBM z16 model 701 offers
performance improvements almost 10% over the IBM z15 Model 701. Figure 12-1 shows a
system performance comparison of successive IBM Z servers.
228* engines
200-way
IBM z15
215 engines
190-way
IBM z14 196 engines
Maximum PCI
170-way
z13 168 engines
141-way
2055 2253
1695 1832
1202 1514
Minimum PCI 581 920
z9 EC z10 EC z196 zEC12 z13 IBM z14 IBM z15 IBM z16
z/OS 1.6 z/OS 1.8 z/OS 1.11 z/OS 1.13 z/OS 2.1 z/OS 2.2 z/OS 2.3 z/OS 2.5
PCI - Processor Capacity Index
228 engines*: IBM 16 Max200 has 200 configurable Pus, 24 SAPs, 2 IFPs and 2 Spares
Operating system support varies for the number of “engines” that are supported.
These numbers differ depending on the workload type and LPAR configuration.
The zEDC Express feature was adopted by enterprises because it helps with software costs
for compression and decompression operations (by offloading these operations), and
increases data encryption (compression before encryption) efficiency.
With IBM z15, the zEDC Express functions were moved off from the PCIe infrastructure into
the processor chip. By moving the compression and decompression into the processor
on-chip, IBM z15 processor provides a new level of performance for these tasks and
eliminates the need for the zEDC Express feature virtualization. It also brings new use cases
to the platform.
The IBM z16 continues to support IBM Integrated Accelerator for zEDC. For more
information, see Appendix C, “IBM Integrated Accelerator for zEnterprise Data Compression”
on page 507.
For z/OS studies, the capacity scaling factor that is commonly associated with the reference
processor is set to a 2094-701 with a Processor Capacity Index (PCI) value of 593. This value
is unchanged since z/OS V1R11 LSPR. The use of the same scaling factor across LSPR
releases minimizes the changes in capacity results for an older study and provides more
accurate capacity view for a new study.
Performance data for IBM z16 servers were obtained with z/OS V2R4 (running Db2 for z/OS
V12, CICS TS V5R3, IMS V14, Enterprise COBOL V6R2, and WebSphere Application Server
for z/OS V9.0.0.8). All IBM Z server generations are measured in the same environment with
the same workloads at high usage.
Note: If your software configuration is different from what is described here, the
performance results might vary.
On average, IBM z16 servers can deliver up to 17% more performance in a 200-way
configuration than an IBM z15 190 -way. However, the observed performance increase varies
depending on the workload type.
Consult the LSPR when you consider performance on the IBM z16. The range of
performance ratings across the individual LSPR workloads is likely to include a large spread.
Performance of the individual logical partitions (LPARs) varies depending on the fluctuating
resource requirements of other partitions and the availability of processor units (PUs).
Therefore, it is important to know which LSPR workload type suite your production
environment. For more information, see 12.8, “Workload performance variation” on page 483.
For more information about performance, see this web page of the IBM Resource Link
website.
For more information about millions of service units (MSU) ratings, see this IBM Z resources
web page.
This processor access to caches and memory is called Relative Nest Intensity (for more
information, see 12.4, “Relative Nest Intensity” on page 479). By using this data, LSPR
introduces three workload capacity categories that replace all older primitives and mixes.
LSPR contains the internal throughput rate ratios (ITRRs) for the IBM z16 and the previous
generation processor families. These ratios are based on measurements and projections that
use standard IBM benchmarks in a controlled environment.
Note: The throughput that any user experiences can vary depending on the amount of
multiprogramming in the user’s job stream, the I/O configuration, and the workload that is
processed. Therefore, no assurance can be given that an individual user can achieve
throughput improvements that are equivalent to the performance ratios that are stated.
However, the path length that is associated with the operating system or subsystem can vary
based on the following factors:
Competition with other tasks in the system for shared resources. As the total number of
tasks grows, more instructions are needed to manage the resources.
The number of logical processors (n-way) of the image or LPAR. As the number of logical
processors grows, more instructions are needed to manage resources that are serialized
by latches and locks.
3
Low I/O Content Mix Workload.
4 Transaction Intensive Mix Workload.
As workloads are moved between microprocessors with various designs, performance varies.
However, when on a processor, this component tends to be similar across all models of that
processor.
With IBM z16, physical L3/L4 caches no longer exist. L2 caches that are on each processor
core are virtual L3/L4 caches on IBM z16. For more information, see Chapter 3, “Central
processor complex design” on page 67.
Figure 12-2 IBM z16 physical and virtual single drawer memory hierarchy
Workload performance is sensitive to how deep into the memory hierarchy the processor
must go to retrieve the workload instructions and data for running. The best performance
occurs when the instructions and data are in the caches that are nearest the processor
because little time is spent waiting before running. If the instructions and data must be
retrieved from farther out in the hierarchy, the processor spends more time waiting for their
arrival.
As workloads are moved between processors with various memory hierarchy designs,
performance varies because the average time to retrieve instructions and data from within the
memory hierarchy varies. Also, when on a processor, this component continues to vary
because the location of a workload’s instructions and data within the memory hierarchy is
affected by several factors that include, but are not limited to, the following factors:
Locality of reference
I/O rate
Competition from other applications and LPARs
The term Relative Nest Intensity (RNI) indicates the level of activity to this part of the memory
hierarchy. By using data from CPU MF, the RNI of the workload that is running in an LPAR can
be calculated. The higher the RNI, the deeper into the memory hierarchy the processor must
go to retrieve the instructions and data for that workload.
The “Nest”
L2LP L2RP
L1 MEMP
L3P L4LP L4RP
Many factors influence the performance of a workload. However, these factors often are
influencing the RNI of the workload. The interaction of all these factors results in a net RNI for
the workload, which in turn directly relates to the performance of the workload.
These factors are tendencies, not absolutes. For example, a workload might have a low I/O
rate, intensive processor use, and a high locality of reference, which all suggest a low RNI.
However, it might be competing with many other applications within the same LPAR and many
other LPARs on the processor, which tends to create a higher RNI. It is the net effect of the
interaction of all these factors that determines the RNI.
The traditional factors that were used to categorize workloads in the past are shown with their
RNI tendency in Figure 12-4.
Little can be done to affect most of these factors. An application type is whatever is necessary
to do the job. The data reference pattern and processor usage tend to be inherent to the
nature of the application. The LPAR configuration and application mix are mostly a function of
what must be supported on a system. The I/O rate can be influenced somewhat through
buffer pool tuning.
However, one factor, software configuration tuning, is often overlooked but can have a direct
effect on RNI. This term refers to the number of address spaces (such as CICS
application-owning regions [AORs] or batch initiators) that are needed to support a workload.
Tuning to reduce the number of simultaneously active address spaces to the optimum number
that is needed to support a workload can reduce RNI and improve performance. In the LSPR,
the number of address spaces for each processor type and n-way configuration is tuned to be
consistent with what is needed to support the workload. Therefore, the LSPR workload
capacity ratios reflect a presumed level of software configuration tuning. Retuning the
software configuration of a production workload as it moves to a larger or faster processor
might be needed to achieve the published LSPR ratios.
The LSPR now runs various combinations of former workload primitives, such as CICS, Db2,
IMS, OSAM, VSAM, WebSphere, COBOL, and utilities, to produce capacity curves that span
the typical range of RNI.
These categories are based on the L1MP and the RNI. The RNI is influenced by many
variables, such as application type, I/O rate, application mix, processor usage, data reference
patterns, LPAR configuration, and the software configuration that is running. CPU MF data
can be collected by z/OS System Measurement Facility on SMF 113 records or by z/VM
Monitor starting with z/VM V5R4.
For more information about how z/VM Monitor captures CPU MF records visit the following
link: https://www.vm.ibm.com/perf/tips/cpumf.html
For more information about the no-charge IBM zPCR tool (which reflects the latest IBM LSPR
measurements), see the Getting Started with IBM z Processor Capacity Reference.
As described in 12.5, “LSPR workload categories based on L1MP and RNI” on page 481, the
underlying performance sensitive factor is how a workload interacts with the processor
hardware.
For more information about RNI, see 12.5, “LSPR workload categories based on L1MP and
RNI” on page 481.
The AVERAGE RNI LSPR workload is intended to match most client workloads. When no
other data is available, use the AVERAGE RNI LSPR workload for capacity analysis.
The CPU MF data can be used determine workload type. When available, this data allows the
RNI for a production workload to be calculated.
By using the RNI and another factor from CPU MF, the L1MP (Level 1 Miss Per 100
instructions or percentage of data and instruction references that miss the L1 cache), a
workload can be classified as LOW, AVERAGE, or HIGH RNI. This classification and resulting
hit are automated in the IBM zPCR tool. It is preferable to use IBM zPCR for capacity sizing.
Refer to Table 12-1 for the LSPR Workload Decision Table, based in L1MP and RNI.
Starting with z/OS V2R1 with APAR OA43366, zFS file is no longer required for CPU MF and
Hardware Instrumentation Services (HIS). HIS is a z/OS function that collects hardware event
data for processors in SMF records type 113, and a z/OS UNIX System Services output files.
Only SMF 113 record is required to know proper workload type by using CPU MF counter
data. CPU overhead of CPUMF is minimal. Also, the amount of SMF 113 record is 1% of
typical SMF 70 and 72 which RMF writes.
CPU MF and HIS can by used for deciding workload type and other purposes. For example,
starting with z/OS V2R1, you can record Instruction Counts in SMF type 30 record when you
activate CPU MF. Therefore, we strongly recommend that you always activate CPU MF.
For more information about getting CPUMF counter data, see CPU MF - 2022 Update and
WSC Experiences at the IBM Techdoc Library website.
A holistic performance approach is required when the performance gains are reduced
because of frequency. Therefore, hardware and software synergy becomes an absolute
requirement.
Starting with z13, Instructions Per Cycle (IPC) improvements in core and cache became the
driving factor for performance gains. As these microarchitectural features increase (which
contributes to instruction parallelism), overall workload performance variability also increases
because not all workloads react the same way to these enhancements.
Because of the nature of the IBM z16 multi-CPC drawer system and resource management
across those drawers, performance variability from application to application is expected.
Also, the memory and cache designs affect various workloads in many ways. All workloads
are improved, with cache-intensive loads benefiting the most. For example, having more PUs
per CPC drawer, each with higher capacity than IBM z15, more workload can fit on an IBM
z16 CPC drawer. This configuration can result in better performance. For example, IBM z15
two drawer system model T01 can populate maximum 71 PUs (Max71).
In contrast, IBM z16 two drawer system Max82 can populate maximum 82 PUs. Therefore,
five and six more PUs can share caches and memories within the first and second drawers
respectively, so the performance improvements is expected.
The workload variability for moving from IBM z14 and IBM z15 to IBM z16 is expected to be
stable. Workloads that are migrating from z10 EC, z196, and zEC12 to IBM z16 can expect to
see similar results with slightly less variability than the migration from IBM z14 and IBM z15.
Do not use MIPS or MSUs for capacity planning: Do not use “one number” capacity
comparisons, such as MIPS or MSUs. IBM does not officially announce the processor
performance as “MIPS”. MSU is only a number for software license charge and it does not
represent the processor's performance.
Note: You should create an EDF file for each z/OS SYSID and load all the EDFs for the
same CPC into IBM zPCR at the same time so to ensure that the correct LSPR Workload
is assigned to each LPAR. IBM zPCR supports using drag-n-drop for multiple EDF files.
CP3KEXTR is offered as a no-charge application. It can also create the EDF files for
IBMzCP3000. IBM zCP3000 is an IBM internal tool, but you can create the EDF files for it on
your system.
For more information about CP3KEXTR, see the IBM Techdoc z/OS Data Extraction Program
(CP3KEXTR) for IBM zPCR and IBM zBNA.
For additional information, see CP3KVMXT - VM Extract Utility for zCP3000 and zPCR
Capacity Planning Support Tools.
Figure 12-5 on page 486 shows a sample IBM zPCR window of a workload type. In this
example, the workload type displays in the “Assigned Workload” column. The example shows
only one partition, PX11, selected. Note that all active partitions can be selected on this
panel. When you load the EDF file to IBM zPCR, it automatically sets your LPAR
configuration. It also makes easy to define the LPAR configuration to the IBM zPCR.
Both windows shown in Figure 12-6 will remain visible. Changes to the Partition Detail Report
will reflect in the HiperDispatch window.
HiperDispatch supports logical CPs running z/OS v1.7 and later and z/VM v6.3 and later. For
z/OS partitions zIIPs and shared CPs are affected similarly. For z/VM partitions IFLs and
associated logical CPs are also affected similarly.
Figure 12-8 HiperDispatch Graph for the GP LCP HD Assignments Processor Topology Support
The HiperDispatch window shown in Figure 12-7 on page 488, contains most of the
information from the Partition Detail Report, plus:
Engines by Weight: Partition Weight % times the number of real CPs in the pool
VHs: Number of LCPs categorized as Vertical High
VMs: Number of LCPs categorized as Vertical Medium.
VM %: Percent of time the partition’s Vertical Medium LCPs are committed
VLs: Number of LCPs categorized as Vertical Low
VL Nvr Pk: Number of LCPs categorized as Vertical Low Never Parked
VL Nvr Pk%: Percent of time the partition’s Vertical Low Never Parked LCPs are
committed.
When input fields are modified on the Partition Detail Report window, results on the
HiperDispatch window will also be updated. Note that when exiting the HiperDispatch window,
any changes made to the Partition Detail Report window are not automatically reset.
Note: For GP or IFL partitions where HiperDispatch is not supported, only the VMs and
VM% columns apply. For ICF partitions, none of the HiperDispatch columns apply.
The IBM zPCR topology report is shown in Figure 12-10 on page 491, where:
LPARs are identified by row
z16 Drawer/DCM/CHIP appears at the top lines
Topology report displays warning messages
LPARs Totals by Pool table is displayed at the bottom with support to filter by partition
Report is accessed from the Partition Detail Window
Latest versions of extract are required:
– available here: IBM Support page.
Note: The Topology report in Figure 12-10 on page 491 is showing all active partitions.
Information about a specific partition can be obtained by clicking on the “Remove all”
button to the left of the Partition Totals by Pool table at the bottom right and then clicking
on the “View” check-box for a specific partition.
IBM zPCR Topology Report is based in the z/OS new Data Gatherer functionality delivered
with APAR A62064, PTF available for z/OS 2.5 and z/OS 2.4. z/OS data is in the SMF 70.1
record.
z/VM support is provided in the 7.3 version and APARs are available for z/VM 7.1 and 7.2.
Additionally, consider collecting the z/OS SMF 99 Subtype 14 records for all LPARs in the IBM
z16. Record has a single LPAR scope, so need all LPARs to get the total picture.
In the Topology Reports, IBM zPCR reports “measured” data and it shows the “what-is” and
not the “what-if” topology scenarios.
Note: To access HMC Partition Resource Assignments, you must use the “Service” logon
id or any other user id with the same authority.
Under “Tasks” at the bottom right click the Configuration (+) sign. Next, click “View Partition
Resource Assignments”. The panel shown in Figure 12-11 will display:
Figure 12-11 HMC - View Partition Resource Assignments for the IBM z167
Use the Partition Resource Assignments to view processor allocations to partitions in your
system. The active logical partitions are identified at the top of the table. The Node and Chip
numbers associated with each active logical partition are identified on the left. You can view
the Node and Chips assignments using the Expand All and Collapse All options under the
“Actions” pull down view or hide sections.
The physical processor types may have some of the following conditions:
Indicates the physical processor types are shared: (SH)
Indicates the physical processor is dedicated: (D)
7
The image has been reduced from its original size horizontally, due to the number of active partitions, and vertically due to the number
of installed drawers.
9
NUMA, Non-Uniform Memory Access is a multiprocessor design where memory access time depends on the memory location relative
to the processor.
up to 191 PUs 48 48 47 48
Figure 12-12 on page 494 shows a Partition Detail Report of an IBM z16, 3931-A01 (Max
200)/700 with 196 CPs (60 GPs, 60 zIIPs, 60 IFLs and 16 ICFs), and four active partitions (1
GP, 1 zIIP, 1 IFL and 1 ICF). Resources are allocated as shown in the Partition Identification
and Partition Configuration fields.
As this IBM z16 has 196 CPs, (Figure 12-12 on page 494), the PU resources are distributed
according to the Table 12-2, and are shown in the Estimated Distribution of RCPs Across
Drawers graph in Figure 12-13 on page 495.
Important:
Please pay special attention to the colors assigned to the Logical CPs column and relate
them to the “warnings” and “notes” at the bottom of the report. See “Partition Detail Report
warnings” on page 494.
8
Your currently installed version of IBM zPCR must be uninstalled before installing IBM zPCR 9.6.4. This step is
necessary to facilitate conversion to the latest IBM Java 17 Semeru 64-bit runtime environment that is included with
IBM zPCR.
Additionally, IBM zPCR version 9.6.4 implemented a new notice for partitions approaching,
within 10%, the maximum drawer size. This critical notice indicates that one or more partitions
partition are getting close to a drawer boundary. When that happens, capacity growth by
adding LCPs is very limited.
The new notice appears as a “Note” message in the Partition Detail Report. The “Note” and
the partition LCPs are shaded with the same violet color, as shown for partition IFL-01 in
Figure 12-12 above and for partition GP-02 in Figure 12-14 on page 496.
Figure 12-14 on page 496 shows another example of a Partition Detail Report this time for an
IBM z16, 3931-A01, (Max200)/700 with 165 CPS (45 GPs, 44 zIIPs, 60 IFLs and 16 ICFs),
and six active partitions (2 GP, 2 zIIP, 1 IFL and 1 ICF). Resources are allocated as shown in
the Partition Identification and Partition Configuration fields.
As the IBM z16 (Figure 12-14 on page 496) has 165 CPs, the PU resources are distributed
according to the Table 12-2 on page 493, and are shown in the Estimated Distribution of
RCPs Across Drawers graph in Figure 12-15 on page 497.
The Partition Detail Report in Figure 12-14 above highlights the partition GP-02 to indicate it
is within 10% of the maximum drawer size in the number of CPs. The GP-02 partition and the
“Note” at the bottom are shaded with the same violet color.
IBM z16 continues the IBM z14 and IBM z15 NUMA9 design. IBM z16 has two clusters and
four DCMs per drawer (refer to Table 12-2 on page 493 for the number of Configurable PUs
per drawer). In the case where a single partition spans from one drawer into a second, the
cross-drawer penalty has increased on IBM z16. However, this is offset by more cores per
drawer and higher capacity than IBM z15, which allows more work to “fit” on a single drawer.
As discussed in 3.5.9, “Processor unit assignment” on page 110, and under “Memory
allocation” on page 113, PR/SM memory and logical processor allocation goal is to place all
logical partition resources on a single CPC drawer, if possible. There can be negative impacts
on a logical partition’s performance when CPs are allocated in different drawers.
IBM zPCR implements a warning notice when logical partition’s number of Logical CPs
defined is larger than the number of Real CPs in a single drawer. When that scenario occurs,
it is advisable to contact IBM to review your configuration. See Figure 12-16 on page 498.
For all optical links, the connector type is LC Duplex (except for the zHyperLink) and the
ICA SR connections, which are established with multi-fiber push-on (MPO) connectors.
The MPO connector of the zHyperLink and the ICA connection feature two rows of 12 fibers
and are interchangeable.
The electrical Ethernet cable for the Open Systems Adapter (OSA) connectivity is connected
through an RJ45 jack.
The attributes of the channel options that are supported on IBM z16 are listed in Table A-1.
zHyperlink Express1.1 0451 OM4, See Table A-2 on page 501 New build,
8 GBps OM3 Carry forward
FICON Express32S SX 0462 8, 16, or 32 OM1, OM2, See Table A-3 on page 502. New build
OM3, or OM4
Open Systems Adapter (OSA) and Remote Direct Memory over Converged Ethernet (RoCE)
OSA-Express7S 1.2 GbE SX 0455 1.25 MM 62.5 µm 275 m (200) New build
MM 50 µm 550 m (500)
Parallel Sysplex
The unrepeated distances for different multimode (MM) fiber optic types for zHyperLink
Express are listed in Table A-2.
The on-chip accelerators provide support for and enable compliance with security policies
because the data is not leaving the platform to be processed. The hardware, firmware, and
software are vertically integrated to seamlessly deliver this function to the applications.
In August 2021, IBM announced a new generation of IBM Z processor, Telum with new
Artificial Intelligence (AI) accelerator (see Figure B-1). This innovation brings incredible value
to the applications and workloads that are running on IBM Z.
With the new IBM Z Integrated Accelerator for AI, customers can benefit from the acceleration
of AI operations, such as fraud detection, customer behavior predictions, and streamlining of
supply chain operations, all in real time. Customers can derive the valuable insights from their
data instantly.
AI accelerator delivers AI inference in real time, at large scale and rate, with no transaction
left behind without the need to offload data off the IBM Z for performing AI inference.
The AI capability is applied directly into the running transaction, shifting the traditional
paradigm of applying AI to the transactions that were completed. This innovative technology
can be used for intelligent IT workloads placement algorithms, which contributes to the better
overall system performance. The co-processor is driven by the new Neural Networks
Processing Assist (NNPA) instruction.
Figure B-2 shows the AI accelerator and its components: the data movers that surround the
compute arrays are composed of the Processor Tiles (PT), Processing Elements (PE), and
Special Function Processors (SFP).
Intelligent data movers and prefetchers are connected to the chip by way of a ring interface for
high-speed, low-latency, read/write cache operations (200+ GBps read/store bandwidth, and
600+GBps bandwidth between engines).
Compute Arrays consist of 128 processor tiles with 8-way FP-16 FMA SIMD, which are
optimized for matrix multiplication and convolution, and 32 processor tiles with 8-way
FP-16/FP-32 SIMD, which is optimized for activation functions and complex functions.
The integrated AI accelerator delivers more than 6 TFLOPs per chip and over 200 TFLOPs in
the 32 chip system (a fully configured IBM z16 with four CPC drawers).
The AI accelerator is shared by all cores on the chip. The firmware, which is running on the
cores and accelerator, orchestrates and synchronizes the execution on the accelerator.
Acknowledging the diverse AI training frameworks, customers can train their models on
platforms of their choice, including IBM Z (on-premises and in hybrid cloud) and then, deploy
it efficiently on IBM Z in colocation with the transactional workloads. No other development
effort is needed to enable this strategy.
To allow this flexible “Train anywhere, Deploy on IBM Z” approach, IBM invests in Open
Neural Network Exchange (ONNX) technology (see Figure B-3).
This standard format represents AI models, with which a data scientist can build and train a
model in the framework of choice without worrying about the downstream inference
implications. To enable deployment of ONNX models, IBM provides an ONNX model compiler
that is optimized for IBM Z. In addition, IBM is optimizing key open source frameworks, such
as TensorFlow (and TensorFlow Serving) for use on IBM Z.
IBM’s open source hzDNN library provides common APIs for the functions that enable
conversions from the Tensor format to the accelerator required format. Customers can run
zDNN under z/OS1 and Linux on IBM Z. A Deep Learning Compiler (DLC) for z/OS and Linux
also is available that provide the AI functions to the applications.
An optional PCIe feature that was available for IBM z14 servers, zEDC Express, addressed
customer requirements by providing hardware-based acceleration for data compression and
decompression. zEDC provided data compression with lower CPU consumption than
compression technology that was available on the IBM Z server.
Many customers increased their zEDC footprint to 8 GBps with up to 16 features per IBM z14
system at 1 GBps throughput per feature (redundancy reduces total throughput to 8 GBps).
Although the zEDC PCIe feature provided CPU savings by offloading the select compression
and decompression operations, it included the drawback of limited virtualization capabilities
(one zEDC PCIe feature can be shared across a maximum of 15 LPARs) and limited
bandwidth.
IBM z15 introduced an on-chip accelerator (implemented in the PU chip) for compression and
decompression operations, which was tied directly into processor’s L3 cache. As a result, it
provided much higher bandwidth and removed the virtualization limitations of a PCIe feature.
The IBM z16 further addresses the growth of data compression requirements with the
integrated on-chip compression unit (implemented in processor Nest, one per PU chip) that
significantly increases compression throughput and speed compared to previous zEDC
deployments.
The IBM z16 Integrated Accelerator for zEDC delivers industry-leading throughput and
replaces the zEDC Express PCIe adapter that is available on the IBM z14 and earlier servers.
IBM z16 compression and decompression are implemented in the Nest Accelerator Unit
(NXU, see Figure 3-15 on page 91) on each processor chip and replaces the existing zEDC
Express adapter in the PCIe+ I/O drawer.
One NXU is available per processor chip, which is shared by all cores on the chip and
features the following benefits:
New concept of sharing and operating an accelerator function in the nest
Supports DEFLATE-compliant compression and decompression and GZIP CRC/ZLIB
Adler
Low latency
High bandwidth
Problem state execution
Hardware and firmware interlocks to ensure system responsiveness
Designed instruction
Run in millicode
Moving the compression function from the (PCIe) I/O drawer to the processor chip means that
compression can operate directly in L3 cache and data does not need to be passed by using
I/O operations.
Compression modes
Compression is run in one of the following modes:
Synchronous
Execution occurs in problem state where the user application starts the instruction in its
virtual address space.
Asynchronous
Execution is optimized for Large Operations under z/OS for authorized applications (for
example, BSAM) and issues I/O by using EADMF for asynchronous execution.
1
Virtual L3 (shared victim) cache for IBM z16. For more information, see Chapter 2, “Central processor complex
hardware components” on page 15.
Note: The zEDC Express feature does not carry forward to IBM z16.
The IFAPRDxx feature is still required for authorized services. For problem state services,
such as zlib use of Java, it is not required.
Performance metrics
On-chip compression introduces the following system reporting changes:
No RMF PCIe reporting for zEDC
Synchronous executions are not recorded (just an instruction invocation)
Asynchronous executions are recorded:
– SMF30 information is captured for asynchronous usage
– RMF EADM reporting is enhanced (RMF 74.10)
– SAP use is updated to include the time that is spent compressing and decompressing
The current zlib and the new zlib function are available for IBM z14 and earlier servers and
IBM z16 hardware. It functions with or without the IBM z16 z/OS PTFs on IBM z14 and earlier
servers.
A specific fix category that is named IBM.Function.zEDC identifies the fixes that enable or use
the zEDC and On-Chip Compression function.
z/OS guests that run under z/VM V7.R1 with PTFs and later can use the zEDC Express
feature and IBM z16 On-Chip Compression.
IBM 31-bit and 64-bit SDK for z/OS Java Technology Edition, Version 7 Release 1 (5655-W43
and 5655-W44) and IBM SDK 7 for z/OS Java provides the use of the zEDC Express feature
and Shared Memory Communications-Remote Direct Memory Access (SMC-R), which is
used by the 10 GbE RoCE Express feature.
For more information about how to implement and use the IBM Z compression features, see
Reduce Storage Occupancy and Increase Operations Efficiency with IBM zEnterprise Data
Compression, SG24-8259.
zBNA is based on Microsoft Windows. It provides graphical and text reports, including Gantt
charts, and support for alternative processors.
zBNA can be used to analyze client-provided System Management Facilities (SMF) records
to identify jobs and data sets that are candidates for zEDC and IBM z16 On-Chip
Compression across a specified time window (often a batch window).
Therefore, zBNA can help you estimate the use of On-Chip Compression features and help
identify savings.
The IBM z16 On-Chip Compression accelerator solves these virtualization limitations
because the function is no longer an I/O device and is available as a problem state instruction
to all Linux on IBM Z guests without constraints.
IBM z16 On-Chip Compression is available to open source applications by way of zlib.
Although they are two different offerings, Tailored Fit Pricing for Software is prerequisite
requirement of Flexible Capacity for Cyber Resiliency (or short ‘Flexible Capacity’).
Note: For more information about Tailored Fit Pricing, see this IBM Z Resources web page.
The information that is included in this section about TFP is taken in part from the IBM
publication Tailored Fit Pricing - A sensible pricing model, 75034875USEN-01.
IBM Z platform users were used to paying for IBM Z software and hardware for peak capacity
and managing their software costs by capping machine usage. Traditionally, they capped
machine use by using the following methods:
Running batch workloads during off-shift hours
Reducing machine resources access to development and test
Not introducing new workloads or applications onto the platform, even when it was the
most logical technology for such workloads
Investing in tools and resources to manage subcapacity capping
IBM introduced Tailored Fit Pricing (originally for IBM Z software) as a simpler pricing model
to allow Z customers to better use their platform investments as their business demands, and
in a more cost competitive way. Building on the success of this program, a variable IBM Z
hardware model was introduced to extend the value of Tailored Fit Pricing for IBM Z.
With Tailored Fit Pricing models now available across hardware and software, customers can
gain more flexibility and control with pricing solutions that can be tailored for business
demands, which helps to balance costs while deriving even more value from hybrid cloud.
IBM’s Tailored Fit Pricing model can address the following key concerns:
Complexity of subcapacity pricing model that lead to IBM Z being managed as a cost
center
Difficulty in establishing the cost of new workload deployment and the effect on cost of
existing workloads
Investment in tools and resources to manage subcapacity that can inflate costs
Lack of development and test resources
Purchasing hardware for peak capacity to handle short term spikes
The software and hardware pricing models provide customers an opportunity to grow and
more fully use their IBM Z investment for new opportunities.
IBM originally introduced DevTest Solution and New Application Solution in 2017. These
solutions further evolved when in May 2019, IBM announced two significant other solutions:
Enterprise Consumption (since renamed to the Software Consumption Solution), and
Enterprise Capacity Solution.
In May 2021, IBM announced a new hardware solution, called Hardware Consumption
Solution. All of these options were gathered into a new family of IBM Z pricing called Tailored
Fit Pricing (TFP).
Customers typically transition onto TFP Consumption for the following reasons:
It is a software pricing model that is better suited to today’s workloads profiles (typically,
where they are increasingly spiky). Also, it is a pricing model that is better suited to future
uses; for example, inclusion in Hybrid Cloud architectures.
A customer on TFP Consumption can confidently remove all forms of capping and expose
all their workloads to all of the hardware infrastructure they own.
Any form of growth (from a new workload to a 30-year-old COBOL application that is used
more often) qualifies for a much-improved Price per MSU.
One key concept in a Software Consumption Model is the customer baseline. The IBM team
works with the customer to review their previous 12 months production MSU consumption
and billing to determine an effective price per MSU and establish a predictable price for all
growth at discounted rate (see Figure D-1).
With the Software Consumption model, no concept of peaks or white space is used that
previously were integral to the sub-capacity-based model. Therefore, customers are free to
remove capping and can use all of the owned capacity without worrying about penalties for
peaking or spiking.
Although a customer does commit to an MSU baseline, if the MSUs for a specific year are not
fully used, the customer can carry over any unused MSUs for use in the following year for the
life of the contract. TFP consumption encourages and rewards growth; that is, MSUs that are
processed above the baseline are charged at an aggressive Growth Price per MSU.
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 515
All workload processing can benefit by running with no capping and having all owned
infrastructure available. Batch processing also can take advantage and reduce batch
windows dramatically.
Without capping in place, customers can expect jobs to finish faster yet at the same cost.
Online processing can process more transactions simultaneously, which improves response
times. This result is a function of the new billing approach that is based on the amount of work
that is done rather than peaks that are reached.
To provide improved economics for growth, TFP Consumption customers pay preferential
pricing on the MSUs that are used above their baseline, regardless of whether that growth
came from existing or new workloads. No other approval, qualification, or processing is
required to use the growth pricing rate.
With coverage of both MLC and capacity-based IPLA products, the Software Consumption
Model offers a single and comprehensive software solution to IBM Z customers.
To meet the demands of modern workloads, IBM Z hardware can now include, in addition to
the base capacity, a subscription-based corridor of pay for use capacity. This always on
corridor of consumption-priced capacity helps alleviate the effect of short unpredictable
spikes in workload that are becoming more common in today’s digital world.
The use charges feature a granularity of 1 hour and are based on MSUs that are used as
measured by the subcapacity reporting Tool (SCRT), not full engine capacity.
Tailored Fit Pricing for IBM Z hardware enables customers to be ready for the unknown and
unexpected. The presence of the always on capacity contributes to better efficiency, reduced
overhead, and shorter response times. This offering is available for customers, starting with
IBM z15, z/OS general-purpose CPs, and who are using Tailored Fit Pricing for IBM Z
software.
One of the most important facts about TFP-Hardware is that a single system or central
processing complex (CPC) always is considered. It is on this level that use is measured. The
blue bars that are shown in Figure D-2 represent individual 15-minute intervals, and the use is
measured in each one.
In Figure D-2, the dark green line shows the average use of the machine over the entire
period that is measure (normally a month). The black line shows the Rolling 4 Hour Average
(R4HA), and it is displayed for information purposes only because the R4HA is not used for
TFP-Hardware calculations.
The red line is the customer owned capacity (or Base capacity). The light green line shows
the activated capacity of the machine, as reported by Sub-Capacity Reporting Tool (SCRT).
As shown in Figure D-2, some of the blue bars reach above the red line, into the corridor
between the red line and light green lines, which represents the TFP-Hardware capacity.
Therefore, use over the customer’s owned (Purchased) capacity is measured, so a Usage
charge is incurred. If no use is measured over the customer’s owned (Purchased) capacity, no
Usage charge is incurred.
In addition to the Usage charge, which might not be relevant, a Subscription charge also can
be assessed. The Subscription charge is a flat, per system, per month payment that is based
on the amount of TFP-Hardware capacity that is provided on the system. The Subscription
charge is invoiced prepaid for the contract term or postpaid monthly.
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 517
Because only entire engines can be activated, it is always full engine sizes that are based on
IBM Large System Performance Reference (LSPR) capacity levels. The Subscription charge
covers the value that the extra activated capacity brings, even if no measured usage exists
(see Figure D-3).
The green bar in Figure D-3 represents the capacity that the customer owns (the so-called
Base capacity). The transparent bars represent activated TFP-Hardware corridor.
The blue bars show the measured usage within the 15-minute intervals that are greater than
the customer-owned capacity. The yellow bar is the highest measured TFP-Hardware use
within the defined hour.
The red bar in Figure D-3 is the highest measured use during a calendar day and is only
relevant if the customer has more than 4 hours of TFP-Hardware use within one calendar day.
The left column that represents one calendar day shows that 10 spikes above the Base
capacity were measured. The spikes that are counted for invoicing are the yellow bars (the
highest within each hour).
The yellow bars hold a measured million of service units (MSU) value. If the total MSU use of
the yellow bars is 250 MSU, the customer receives an invoice of 250 times the hourly usage
charge per MSU.
Also, the right column in Figure D-3 represents one calendar day, where 18 spikes above the
Base capacity are measured. The spike that is counted for invoicing is the red bar (the highest
within the calendar day). The red bar is considered because it was measured that the
TFP-Hardware corridor was used for more than 4 hours during the calendar day.
Conclusion
The combination of TFP for the hardware and software is a powerful one for our customers to
maximize their investment in IBM Z. For eligible customers, it allows them to use their
hardware for their general capacity requirements with a consumption-based corridor on top.
When this TFP for hardware is combined with the TFP-SW solutions, it allows customers to
unleash the full power of their IBM Z machine.
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 519
IBM Z Flexible Capacity for Cyber Resiliency
Resiliency continues to be a hot topic for clients and a key value driver for IBM Z, especially
for the largest IBM Z clients in regulated industries.
IBM’s answer was the introduction of zDR Cloud. This offering enables active MIPS flexibility
between the production site and the DR site. It also offers greater flexibility over CBU tests for
length of test execution and frequency.
To overcome these issues, IBM developed the new IBM Z Flexible Capacity for Cyber
Resiliency offering.
Flexible Capacity is designed to provide increased flexibility and control to shift production
workloads between participating IBM z16 machines at different sites and stay for up to one
year.
It helps organizations shift capacity between participating IBM z16 machines at different sites
to improve Cyber Resiliency and offers increased flexibility to shift capacity on demand
between participating IBM z16 machines at different sites.
By using IBM GDPS, scripts can be run to fully automate Flexible Capacity for Cyber
Resiliency capacity shifts between participating IBM z16 machines at different sites.
Frictionless compliance
Meet the ever-evolving stringent requirements of global regulators, which allows a highly
automated and fast process to demonstrate a production site swap.
Pro-active avoidance
Protect your critical business services from natural disasters. Avoid rolling power outages.
Migrate your critical workloads to an alternative site before your business is affected and
remain in that state for up to one year.
Set up process
Flexible Capacity for Cyber Resiliency is facilitated through a new type of Temporary Capacity
record.
In the first step of the setup process (see Figure D-4), the active capacity of the participating
IBM z16 machines is changed to a base machine plus the temporary entitlement record
(TER) up to the High Water Mark (HWM) of the machine. The base capacity is defined by the
customer. The machines’ HWM remains unchanged.
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 521
In the next step (see Figure D-5), the now unassigned capacity is restored with a new Flexible
Capacity record. Another Flexible Capacity record is installed on System B to increase the
capacity to the amount of MIPS that the customer licensed.
The IBM Flex Capacity record on System A then is activated until the machine’s HWM. On
System B, the Flexible Capacity record is installed, but not activated. The setup process is
now complete (see Figure D-6).
Transferring capacity
After the set up is complete, it is possible to transfer capacity and workload from Site A to Site
B. This process can be automated by using GDPS technology.
During this time window, Flexible Capacity can remain active on both sites, as shown in
Figure D-7
After 24 hours, the Flexible Capacity in Site A is removed and System A is reduced to base
capacity (see Figure D-8).
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 523
Multi-system environment
A two-site, two-system environment is only the base infrastructure. Flexible Capacity for
Cyber Resiliency also is suited for complex multi-system environments.
Flexible Capacity from several systems can be transferred to single systems in Site B or
vice-versa (see Figure D-9).
Movement of capacity does not need to be “all to one”: the capacity also can be split.
However, the total capacity that is active on all systems after the swap cannot exceed the total
capacity that is active before the swap.
Flexibly Capacity is available on all engine types; exceeding the purchased capacity incurs
charged. Monitoring to ensure compliance is done by way of Call Home data that IBM
receives periodically.
The Flexible Capacity Transfer record always is considered to be the first record that was
activated, regardless of the order in which temporary records or TFP-Hardware were
activated.
The example that is shown in Figure D-10 shows an active Sysplex across Site A and Site B.
System A has a Base capacity of 401 and the Flexible Capacity Record is partially activated
up to a 710. On top of the active Flexible Capacity sits a TFP-Hardware capacity of two extra
CPs to a maximum of 712. The inactive part of the Flexible Capacity record floats on top of
the TFP-Hardware capacity.
Now, the data-center operator decides to perform maintenance on System A and transfers its
active Flexible Capacity to System B.
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 525
Figure D-11shows the configuration of both machines after the transfer.
System A transferred all of its Flexible Capacity to System B and is left with only the base
capacity of 401. On top of the base capacity sits the TFP-Hardware capacity.
System B has now the entire Flexible Capacity record activated and shows an active capacity
of 720. The TFP-Hardware capacity sits again on top of the active capacity and adds now
another 2 CPs to the 720.
This shows that the presence or activation of TFP-Hardware does not impact the amount of
capacity that can be activated by a Flexible Capacity Transfer record.
The TFP-Hardware capacity always “floats on top” of any other activated capacity for
TFP-Hardware usage charging. Therefore, no double charging can occur.
FC 9933 and 0376 must be active on each machine that is participating in Flexible Capacity
for Cyber Resiliency.
A new Capacity on-Demand contract attachment also must be signed by the customer.
Cross site movement Inter-site moves can be done, regardless of distance, mirroring, or
coupling technology.
Intra-site moves are not allowed (two machines in the same data
center cannot move capacity back and forth).
Entitlement The owner of the machine holds a title to the physical hardware.
The capacity of that machine is enabled and controlled by way of
the LIC of the machine, which is licensed, not sold
Overlap period A 24-hour period in which the temporary record can be active on
both systems.
Activation period Keep the flexible capacity record active on your alternative site for
up to one year.
License transfer LIC is licensed to one specific serial numbered machine only. It
cannot be transferred to another machine.
License expiration The LIC license is expired 5 years past WFM. An invalid LIC
license resumes if the IBM Z machine is upgraded or replaced with
an IBM Z machine that is not older than N-2.
TFP for software Offering requires TFP for software; CMP is grandfathered in.
Maintenance Continue with same pricing scheme as for zDR Cloud 1.0 (price
active capacity). Overlap time is determined by the customer.
Microcode only IBM Z Flexible Capacity for Cyber Resiliency is Microcode only.
More memory, I/O cards, drawers, and other infrastructure-related
components must be prepared by the client.
Call home Customer agrees to use Call Home data to monitor capacity
usage.
Charges for capacity Capacity that is used beyond the purchased capacity is charged at
exceeding the temp record previously defined OOCoD prices.
Appendix D. Tailored Fit Pricing and IBM Z Flexible Capacity for Cyber Resiliency 527
IBM Z Flexible Capacity for Cyber Resilience versus Capacity
Back Up
Flexible Capacity and Capacity Back Up (CBU) feature different purposes and limitations, as
listed in Table D-2.
Primary All CBU benefits at an alternative site, plus: Run full mainframe workload if DR event is
purpose and Run mainframe workload in alternative site declared:
benefits for up to 1 year Ability to test DR readiness up to 10 days.
Demonstrate ability to migrate entire Validate with nonproduction workload (up to
mainframe workload to alternative site (for 10 days per CBU test).
example, regulatory requirements) Validate with production workload (up to 10
Run entire workload in alternative site for days per CBU test). Requires primary site
extended period (for example, facilities capacity to be disabled.
requirements)
Proactively swap production to alternative
out-of-region site before a (natural) disaster
occurs
Limitations Alternative site can be used for maximum of Each CBU test is limited to a maximum of 10
1 year days
Allows up to 24-hour overlap period, where
full capacity is active on both sites for
transferring workload
Allows 12 swaps between sites per year
(one direction)
Contracts and LIC is licensed to only one specific Capacity Back Up Agreement
features serial-numbered machine, and its transfer CBU capacity provided in annual
to another machine is not permitted increments
Offering requires TFP for software; CMP is CBU tests provided in individual increments
grandfathered in.
The LIC license expires 5 years past WFM.
An invalid LIC license resumes if the IBM Z
machine is upgraded or replaced with an
IBM Z machine that is not older than N-2.
The common building blocks are displayed and range from 1 - 4 frames, with various numbers
of CPC drawers and PCIe+ I/O drawers.
Appendix E. Frame configurations Power Distribution Units and Bulk Power Assembly-based systems 531
Figure E-5 Three frames ZAB
Appendix E. Frame configurations Power Distribution Units and Bulk Power Assembly-based systems 533
Figure E-9 Two frames AB, four CPCs
Appendix E. Frame configurations Power Distribution Units and Bulk Power Assembly-based systems 535
Figure E-13 Three frames ZAB
Appendix E. Frame configurations Power Distribution Units and Bulk Power Assembly-based systems 537
Figure E-17 Four frames ZABC, four CPC
The publications that are listed in this section are considered particularly suitable for a more
detailed discussion of the topics that are covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide more information about the topic in this
document. Note that some publications that are referenced in this list might be available in
softcopy only:
IBM z16 Technical Introduction, SG24-8950
IBM Z Connectivity Handbook, SG24-5444
IBM z16 (3931) Configuration Setup, SG24-8960
You can search for, view, download, or order these documents and other Redbooks,
Redpapers, Web Docs, draft and additional materials, at the following website:
ibm.com/redbooks
Other publications
The publication IBM Z 8561 Installation Manual for Physical Planning, GC28-7002, also is
relevant as another information source.
Online resources
The following online resources are available:
The IBM Resource Link for documentation and tools website:
http://www.ibm.com/servers/resourcelink
IBM Telum Processor: The next-gen microprocessor for IBM Z and IBM LinuxONE:
https://www.ibm.com/blogs/systems/ibm-telum-processor-the-next-gen-microprocess
or-for-ibm-z-and-ibm-linuxone/
Leveraging ONNX Models on IBM Z and LinuxONE:
https://community.ibm.com/community/user/ibmz-and-linuxone/blogs/andrew-sica/20
21/10/29/leveraging-onnx-models-on-ibm-z-and-linuxone
Jump starting your experience with AI on IBM Z:
https://blog.share.org/Article/jump-starting-your-experience-with-ai-on-ibm-z
SG24-8951-00
ISBN 0738460788
Printed in U.S.A.
®
ibm.com/redbooks