CPUs_GPUs_accelerators_and_memory_v1.0
CPUs_GPUs_accelerators_and_memory_v1.0
Andrea Sciabà
On behalf of the Technology Watch WG
HOW Workshop
18-22 March 2019
Jefferson Lab, Newport News
Introduction
• The goal of the presentation is to give a broad overview of the
status and prospects of compute technologies
– Intentionally, with a HEP computing bias
• Focus on processors and accelerators and volatile memory
• The wider purpose of the working group is to provide information
that can be used to optimize investments
– Market trends, price evolution
• More detailed information is already available in a document
– Soon to be added to the WG website
2
Outline
• General market trends
• CPUs
– Intel, AMD
– ARM
– Other architectures
• GPUs
• FPGAs
• Supporting technologies
• Memory technologies
3
Semiconductor device market and trends
4
Semiconductor fabrication
6
Intel and AMD market share
• AMD server market share is rapidly Source: Passmark website
7
Internet and smart population growth and effects
8
CPUS AND ACCELERATORS
9
Intel server CPU line-up
• Intel Xeon Scalable Processors
– Currently based on Skylake-SP and
coming in four flavours, up to 28 cores
• Only minor improvements foreseen for
2019
– Adding support for Optane DC Persistent
Memory and hardware security patches
• New microarchitecture (Sunny Cove) to
become available late 2019
– Several improvements benefiting both
generic and specialised applications
10
Current and future Intel server architectures
Microarchitecture Technology Launch year Highlights
Skylake-SP 14nm 2017 Improved frontend and execution units
More load/store bandwidth
Improved hyperthreading
AVX-512
Cascade Lake 14nm++ 2019 Vector Neural Network Instructions (VNNI) to improve inference performance
Support 3D XPoint-based memory modules and Optane DC
Security mitigations
Cooper Lake 14nm++ 2020 bfloat16 (brain floating point format)
Sunny Cove 10nm+ 2019 Single threaded performance
(aka Ice Lake) New instructions
Improved scalability
Larger L1, L2, μop caches and 2nd level TLB
More execution ports
Willow Cove 10nm 2020? Cache redesign
New transistor optimization
Security Features
Golden Cove 7/10nm? 2021? Single threaded performance
AI Performance
Networking/5G Performance
Security Features
11
Other Intel x86 architectures
• Xeon Phi
– Features 4-way hyperthreading and AVX-512 support
– Elicited a lot of interest in the HEP community and for deep learning applications
– Announced to be discontinued in summer 2018
• Networking processors (Xeon D)
– SoC design
– Used to accelerate networking functionality or to process encrypted data streams
– Two families, D-500 for networking and D-100 for higher performance, based on
Skylake-SP with on-package chipset
– Hewitt Lake just announced, probably based on Cascade Lake
• Hybrid CPUs
– Will be enabled by Foveros, the 3D chip stacking technology recently demonstrated
12
AMD server CPU line-up
• EPYC 7000 line-up from 2017
– Resurgence after many years of
Bulldozer CPUs thanks to the
Zen microarchitecture
• +40% in IPC, almost on par with
Intel
• 2x power efficiency vs Piledriver
– Up to 32 cores
• Already being tested and
used at some WLCG sites
13
EPYC Naples
• EPYC Naples (Zen) consists of up to 4 separate dies,
interconnected via Infinity Fabric
– Chiplets allow a significant reduction in cost and
higher yield
• Main specifications
– up to 32 cores
– 4 dies per chip (14nm), each die embedding IO and
memory controllers
– 2.0-3.1 GHz of base frequency
– 8 DDR4 memory channels with hardware encryption
– up to 128 PCI gen3 lanes per processor (64 in dual )
– TDP range: 120W-200W
• Similar per-core and per-GHz HS06 performance to
Xeon
14
EPYC Rome
• Next AMD EPYC generation (Zen 2), embeds 9 dies, including
one for I/O and memory access
– Should compete with Ice Lake
• Main specs:
– 9 dies per chip : a 14nm single IO/memory die and 8 CPU 7nm
chiplets
• +300-400 MHz for low core count CPUs
– 8 DDR4 memory channels, up to 3200 MHz
– up to 64 cores
– up to 128 PCI Gen3/4 lanes per processor
– TDP range: 120W-225W (max 190W for SP3 compatibility)
– Claimed +20% performance per-core over Zen, +75% through
the whole chip with similar TDP over Naples
– To be released during 2019
15
Recent experiences in WLCG
1. LHCb
– Using some nodes with EPYC 7301 CPUs (16 cores)
– Performance of LHCb trigger application almost equal to Xeon Silver 4114 (10 cores)
– Need to populate all 8 DIMM slots for maximum performance
– Testing it as potential hypervisor platform
– Will competitively tender with Intel next year
2. NIKHEF
– Have 93 single-socket 32 core EPYC 7551P nodes in production
– A single EPYC 7371 node (single socket, 16 cores), available for tests
3. INFN
– All WLCG sites have installed in 2018 a number of systems (40 in total) with EPYC 7351 (16 cores) in Twin Square configuration
– Experience very positive
4. BNL
– Extensive tests with several EPYC CPUs presented at HEPiX Fall 2018
– Measured performance from mid/upper range EPYC similar to mid/upper range Xeon Gold
• Caltech
– Two servers with EPYC 7551P (32 cores), soon available for benchmarking
16
ARM in the data center
• ARM is ubiquitous in the mobile and
embedded CPU word
• Data center implementations have been
relatively unsuccessful so far
– Performance/power and performance/$ not
competitive with Intel and AMD
• LHC experiments are capable of using ARM
CPUs if needed
– Some do nightly builds on ARM since years
• Only a few implementations (potentially)
relevant to the data center
– Cavium ThunderX2
– Fujitsu A64FX
– ARM Neoverse
– Ampere eMAG, Graviton
17
Marvell ThunderX2 and Fujitsu A64FX
• ThunderX2 for mainstream cloud and HPC data centers, from
2018
– Enjoys the greatest market visibility and reasonable
performance/$
• Used e.g. at CRAY XC-50 at Los Alamos and HPE Apollo 70 based Astra
HPC system at Sandia National Laboratory
– ARM V8.1 architecture
• Up to 32 cores, 4-way SMT
• Up to 8 DDR4 memory channels
• Up to 56 PCIe Gen3 lanes
20
RISC-V and MIPS
• RISC-V is an open source ISA
– To be used by some companies for controllers (Nvidia and WD), for
FPGA (Microsemi), for fitness bands…
– For the time being, not targeting the data center
– Might compete with ARM in the mid term
– Completely eclipsed MIPS
• MIPS
– Considered dead
21
Discrete GPUs: current status
• GPU’s raw power follows the exponential trend on numbers of
transistors and cores
• New features appear unexpectedly, driven by market (e.g. tensor
cores)
– Tensor cores: programmable matrix-multiply-and-accumulate units
– Fast half precision multiplication and reduction in full precision
– Useful for accelerating deep learning training/inference
https://devblogs.nvidia.com/programming-tensor-cores-cuda-9/ 22
Nvidia and AMD
• Volta addressing the server market, • Vega 20
Turing the gaming market – Directly aimed at the server world (Instinct
MI50 and MI60)
Feature Volta (V100) Turing (2080 Ti) • Evolution of Vega 10 using a 7nm
Process 12nm 12nm process
CUDA cores yes yes
– more space for HBM2 memory, up to 32GB
Tensor cores yes yes
– 2x memory bandwidth
RT cores NA yes
FP16: 28 TFLOPS Same, but
– Massive FP64 gains
FP performance
FP32: 14 TFLOPS FP64: 1/32 of FP32 – PCIe Gen4
FP64: 7 TFLOPS
Tensor: 112 TFLOPS • Some improvements relevant for
Memory HBM2 GDDR6
inference scenarios
Memory bandwidth 900 GB/sec 616 GB/sec
– Support for INT8 and INT4 data types
Multi-GPU NVLink 2 NVLink 2/SLI
– Some new instructions
Applications AI, datacenter, AI, workstation,
workstation gaming
23
GPUs - Programmability
• NVIDIA CUDA:
– C++ based (supports C++14), de-facto standard
– New hardware features available with no delay in the API
• OpenCL:
– Can execute on CPUs, AMD GPUs and recently Intel FPGAs
– Overpromised in the past, with scarce popularity
• Compiler directives: OpenMP/OpenACC
– Latest GCC and LLVM include support for CUDA backend
• AMD HIP:
– Interfaces to both CUDA and AMD MIOpen, still supports only a subset of the CUDA
features
• GPU-enabled frameworks to hide complexity (Tensorflow)
• Issue is performance portability and code duplication
24
GPUs in LHC experiments software frameworks
• Alice, O2 • LHCb (online - standalone) Allen
– Tracking in TPC and ITS framework: HLT-1 reduces 5TB/s input
– Modern GPU can replace 40 CPU cores to 130GB/s:
• CMS, CMSSW – Track reconstruction, muon-id, two-tracks
vertex/mass reconstruction
– Demonstrated advantage of
heterogeneous reconstruction from RAW – GPUs can be used to accelerate the entire
to Pixel Vertices at the CMS HLT HLT-1 from RAW data
– ~10x both in speed-up and energy – Events too small, have to be batched:
efficiency wrt full Xeon socket makes the integration in Gaudi difficult
– Plans to run heterogeneous HLT during LHC • ATLAS
Run3 – Prototype for HLT track seed-finding,
calorimeter topological clustering and anti-
kt jet reconstruction
– No plans to deploy this in the trigger for
Run 3
25
FPGA
• Players: Xilinx (US), Intel (US), Lattice
Semiconductor (US), Microsemi (US),
and QuickLogic (US), TSMC (Taiwan),
Microchip Technology (US), United
Microelectronics (Taiwan),
GLOBALFOUNDRIES (US), Achronix (US),
and S2C Inc. (US)
• Market valued at USD 5 Billion in 2016 Process Technology
20 nm 16 nm 14 nm
and expected to be valued at 10 Billion Intel® Xilinx® Intel® Xilinx® Intel® Xilinx®
in 2023 Virtex®
Virtex®
UltraScale+®
Intel®
Stratix®
• Growing demand for advanced driver-
Top Performance Tier UltraScale
® Zynq®
10
UltraScale+®
assistance systems (ADAS), Intel® Arria®
Kintex
developments in IoT and reduction in Mid Performance Tier
10
UltraScale
®
Source: https://www.intel.com/content/www/us/en/programmable/documentation/mtr1422491996806.html#qom1512594527835__fn_soc_variab_avail_xlx
26
FPGA programming
• Used as an application acceleration device • In HEP
– Targeted at specific use cases
• Neural inference engine – High Level Triggers
• MATLAB
• LabVIEW FPGA • https://cds.cern.ch/record/2647951
• OpenCL – Deep Neural Networks
– Very high level abstraction • https://arxiv.org/abs/1804.06913
– Optimized for data parallelism
• https://indico.cern.ch/event/703881/
• C / C++ / System C
– High level synthesis (HLS) – High Throughput Data Processing
– Control with compiler switches and • https://indico.cern.ch/event/669298/
configurations
• VHDL / Verilog
– Low level programming
27
Other Machine Learning processors and accelerators
28
MEMORY TECHNOLOGIES
29
Static RAM (SRAM)
• On die memory on the CPU used for
L1/L2/L3 cache
– SRAM cell size not scaling with node
• SRAM cache constitutes large fraction of
area on modern CPUs
• Power consumption is an issue
• Applications driving larger caches
• No direct replacement in sight for
L1/L2
• Alternate L3 cache technologies
– eDRAM - Used in IBM Power CPUs
– STT-MRAM - proposed as possible
replacement
https://www.sigarch.org/whats-the-future-of-technology-scaling/
30
Dynamic RAM (DRAM)
• Dominant standards continue to
evolve
– DDR4 -> DDR5
• 3200MT/s -> 6400MT/s
• 16Gb -> 32Gb chips
– GDDR5 - > GDDR5X
• 14 Gbps/pin -> 16Gbps/pin
• 8Gb -> 16Gb chips
– HBM -> HBM2
• 1 Gbps/pin -> 2.4 Gbps/pin
• 4 die stack -> 12 die stack
• 2Gb die -> 8Gb die
• Note memory latency remains mostly (Youngwoo Kim, KAIST’s Terabyte Labs)
https://www.3dincites.com/2019/02/designcon-2019-shows-board-and-system-designers-
unchanged the-benefits-of-advanced-ic-packaging/
31
DRAM Outlook
• Major vendors showing next
generation chips (DDR5/GDDR6)
• Multiple technologies being
investigated for future DRAM
• EUV lithography not needed for at
least 3 more generations (Micron)
• Contract DRAM pricing fell ~30% in
Q1 2019
• Pressure expected on DRAM prices
thru 2019 due to additional
production capacity coming online
https://www.techinsights.com/technology-intelligence/overview/technology-roadmaps/
32
Performance gaps in memory hierarchy
https://www.opencompute.org/files/OCP-GenZ-March-2018-final.pdf https://www.eetimes.com/author.asp?section_id=36&doc_id=1334088#
33
Emerging technologies
• May eventually fill the gap
– STT-MRAM between SRAM and
DRAM (work in progress)
– “Persistent Memory” in NVDIMM
package for the DRAM/NAND gap
• Low latency NAND (e.g. Z-NAND)
• 3D XPoint (aka “Optane”)
– Technologies still in the lab
• MRAM
• NRAM
• FeRAM
• PCRAM https://www.snia.org/sites/default/files/PM-
Summit/2018/presentations/14_PM_Summit_18_Analysts_Session_Oros_Final_Post_UPD
ATED_R2.pdf
• ReRAM
34
SUPPORTING TECHNOLOGIES
35
Interconnect technology
• Increasing requirements on bandwidth
and latency driving the development
– E.g. moving data between CPU and GPU
is often a bottleneck
– Several standards competing (PCIe
Gen4/5, CCIX, Gen-Z, OpenCAPI, CXL…)
• Proprietary technologies
– NVLink (GPU-to-GPU, GPU-to-POWER9)
– Ultra Path (Intel), CPU-to-CPU
– Infinity Fabric (AMD), chiplet-to-chiplet
36
Packaging technology
• Traditionally a silicon die is individually packaged,
but more and more CPUs package together more
(sometimes different) dies
• Classified according to how dies are arranged and
connected
– 2D packaging (e.g. AMD EPYC): multiple dies on a
substrate
– 2.5D packaging (e.g. Intel Kaby Lake-G, CPU+GPU):
interposer between die and substrate for higher
speed
– Intel Foveros, a 2.5D with an interposer with active
logic (Intel “Lake Field” hybrid CPU)
– 3D packaging (e.g. stacked DRAM in HBM), for lower
power, higher bandwidth and smaller footprint
• Can alleviate scaling issues with monolithic CPU
dies but at a cost, both financial and in power and
latency
37
What next?
• We do not really know what will be there in the HL-LHC era
(2026-2037)
• Some “early indicators” of what might come next
– Several nanoelectronics projects might help in
• Increasing density of memory chips
• Reducing size of transistors in IC
– Nanocrystals, silicon nanophotonics, carbon nanotubes, single-atom
thick graphene film, etc.
– https://www.understandingnano.com/nanotechnology-electronics.html
38
Conclusions
• Market trends
– Server market is increasing, AMD share as well
– EUV lithography driving 7nm mass production
• CPU, GPUs and accelerators
– AMD EPYC promising from a cost perspective
– Nvidia GPUs still dominant due to the better software support
– Recent developments for GPUs greatly favor inference workloads
– FPGA market dominated by telecom, industry and automotive but there is also some HEP usage
• Memory technologies
– SDRAM still the on-chip memory of choice, DRAM still for the main memory, no improvements in
latency
– NVDIMM – emerging memory packaging for memory between DRAM and NAND flash (see next
talk)
– Other non-volatile memory technologies in development
39
Additional resources
• All subgroups
– https://gitlab.cern.ch/hepix-techwatch-wg
• CPUs, GPUs and accelerators
– Document (link)
• Memory technologies
– Document (link)
40
Acknowledgments
• Special thanks to Shigeki Misawa, Servesh Muralidharan, Peter
Wegner, Eric Yen, Andrea Chierici, Chris Hollowell, Charles
Leggett, Michele Michelotto, Niko Neufeld, Harvey Newman,
Felice Pantaleo, Bernd Panzer-Steindel, Mattieu Puel and
Tristan Suerink
41
BACKUP SLIDES
42
Market share of technology companies
Server companies PC companies
44