Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
GPU Programming with CUDA – Accelerated Architectures Mike Griffiths GPUComputing@Sheffield http://gpucomputing.sites.sheffield.ac.uk/ Overview • Why Accelerators? • Architectural Details • Latest Products • Accelerated Systems Overview • Why Accelerators? • Architectural Details • Latest Products • Accelerated Systems CPU Limitations • Power = Frequency x Voltage² • Performance Improvements traditionally realised by increasing frequency • Voltage decreased to maintain steady power • Voltage cannot be decreased any further • 1’s and 0’s represented by different voltages • Need to be able to distinguish between the two Moores Law Moores Law: A doubling of transistors every couple of years BUT Clock speeds are not increasing Longer more complex pipelines? Increase performance by adding parallelism Perform many operations per clock cycle More cores More operations per core Keep power per core low Accelerators • Much of the functionality of CPUs is unused for HPC • Branch prediction, out of order execution, etc. • Ideally for HPC we want: Simple, Low Power and Highly Parallel cores • Problem: Still need operating systems, I/O, scheduling • Solution: “Hybrid Systems” – CPUs provide management, “Accelerators” (or co-processors) provide compute power. Overview • Why Accelerators? • Architectural Details • Latest Products • Accelerated Systems Designing an Accelerator • Chip fabrication prohibitively expensive • HPC market relatively small • Graphics Processing Units (GPUs) have evolved from the desire from improved graphical realism in games • Significantly different architecture • Lots of number crunching cores • Highly parallel • Initially GPUs started to be used for general purpose use (GPGPU) • NVIDIA and AMD now tailor their architectures for HPC What are the alternatives • Intel Xeon Phi – Many Integrated Core (MIC) architecture • Lots of Pentium cores with wide vector units • Closer to traditional multi-core • Simplifies programming? • Codenamed “Larrabee, Knights Ferry, Knights Corner, Knights Landing” • Many simple-low power cores Accelerators in HPC 48,000 Xeon Phi boards Equal number of Opterons to GPUS Architecture of a Multi-Core CPU • AMD 12-core • Not much space is dedicated to compute = compute unit (core) Architecture of a NVIDIA GPU • NVIDIA Fermi GPU • Much more space dedicated to compute (at the cost of cache and advanced features) = streaming multiprocessors compute unit (each with 32 cores) Architecture of a Xeon Phi • Similarly has large amounts of dedicated compute space = compute unit (core) Memory • Accelerators use dedicated Graphics Memory • Separate to CPU “main” memory CPUs use DRAM GPUs and Xeon Phi use Graphics DRAM • Many HPC applications require high memory bandwidth Overview • Why Accelerators? • Architectural Details • Latest Products (with a focus on NVIDIA GPUs) • Accelerated Systems Latest Products • NVIDIA – Tesla GPUs, specifically for HPC (using same architecture as Ge-Force) • AMD – FirePro HPC, specifically for HPC (evolved from ATI Radeon) • Intel – Xeon Phi – recently emerged to compete with GPUs Tesla Series GPUs • Chip partitioned into Streaming Multiprocessors (SMs) • Multiple cores per SM • Not cache coherent. No communication possible across SMs. NVIDIA Streaming Multiprocessor • Less scheduling units than cores • Threads are scheduled in groups of 32, called a warp • Threads within a warp always execute the same instruction in lock-step (on different data elements) Tesla Range Specifications “Fermi” 2050 “Fermi” 2070 “Fermi” 2090 “Kepler” K20 “Kepler” K20X “Kepler” K40 CUDA cores 448 448 512 2496 2688 2880 DP Performance 515 GFlops 515 GFlops 665 GFlops 1.17 TFlops 1.31 TFlops 1.43 Tflops Memory Bandwidth 144 GB/s 144 GB/s 178 GB/s 208 GB/s 250 GB/s 288 GB/s Memory 3 GB 6 GB 6 GB 5 GB 6 GB 12 GB NVIDIA Roadmap Overview • Why Accelerators? • Architectural Details • Latest Products • Accelerated Systems Machine architectures P P P P P P P P C C C C C C C C M M M M Interconnect Memory Interconnect Distributed memory system Shared-memory system. e.g. Beowulf clusters. e.g. Sunfire, SGI Origin, Symetric Multiprocessors Architecture matches message passing paradigm. Can use interconnect + memory as a communications network P Processor C Cache Accelerated Systems • CPUs and Accelerators are used together • GPUs cannot be used instead of CPUs • GPUs perform compute heavy parts • Communication is via PCIe bus DRAM GDRAM CPU GPU/ Accelerator I/O PCIe I/O Larger Accelerated Systems • Can have multiple CPUs and Accelerators within each “Shared Memory Node” • CPUs share memory but accelerators do not! I/O PCIe I/O GPU/ Accelerator CPU GDRAM Interconnect DRAM GDRAM GPU/ Accelerator CPU PCIe I/O I/O Accelerated Supercomputers … … … … … … Multiple Accelerators in Parallel • (Normally) use one host CPU core (thread) per accelerator • Program manages communication between host CPUs • MPI for distributed memory • OpenMP for shared memory on the same node Simple Accelerated Workstation • Insert your accelerator into PCI-e • Make sure that • There is enough space • Your power supply unit (PSU)is up to the job • You install the latest drivers GPU Workstation Server • Multiple Servers can be connected via interconnect • Several vendors offer GPU servers • For example 2 multi core CPUs + 4 GPUS Compute Blades • Dedicated HPC Blades for scalable HPC • E.g. Cray XK7 • 4 CPUs + 4 GPUS + 2 interconnect chips (shared by 2 computer nodes The Iceberg GPU Nodes C410X with 8 Fermi GPU 2xC6100 with dual Intel westmere 6core CPU’s Programming Techniques • GPU Accelerated Libraries and Applications (MATLAB, Ansys, etc) • GPU mostly abstracted from end user • GPU Accelerated Directives (OpenACC) • Helps compiler auto generate code for the GPU • CUDA for NVIDIA GPUs • Extension to the C language (more to follow) • OpenCL • Similar to CUDA but cross-platform • No access to cutting edge NVIDIA functionaility Summary • Accelerators have higher compute and memory bandwidth capabilities than CPUs • Silicon dedicated to many simplistic cores • Use of graphics memory • Accelerators are typically not used alone, but work in tandem with CPUs • Most common are NVIDIA GPUs and Intel Xeon Phis. • Including current top 2 systems on top500 list • Architectures differ • GPU accelerated systems scale from simple workstations to large-scale supercomputers