Parallelism in Modern C++: From CPU to GPU [2019 Class Archive]

“Parallelism in Modern C++: From CPU to GPU” is a two-day training course with programming exercises taught by Gordon Brown and Michael Wong. It is offered at the Gaylord Rockies from 9AM to 5PM on Saturday and Sunday, September 21st and 22nd, 2019 (immediately following the conference). Lunch is included.

Course Description

Parallel programming can be used to take advance of multi-core and heterogeneous architectures and can significantly increase the performance of software. It has gained a reputation for being difficult, but is it really? Modern C++ has gone a long way to making parallel programming easier and more accessible; providing both high-level and low-level abstractions. C++11 introduced the C++ memory model and standard threading library which includes threads, futures, promises, mutexes, atomics and more. C++17 takes this further by providing high level parallel algorithms; parallel implementations of many standard algorithms; and much more is expected in C++20. The introduction of the parallel algorithms also opens C++ to supporting non-CPU architectures, such as GPU, FPGAs, APUs and other accelerators.

This course will teach you the fundamentals of parallelism; how to recognize when to use parallelism, how to make the best choices and common parallel patterns such as reduce, map and scan which can be used over and again. It will teach you how to make use of the C++ standard threading library, but it will take this further by teaching you how to extend parallelism to heterogeneous devices, using the SYCL programming model to implement these patterns on a GPU using standard C++.

Prerequisites

This course requires the following:

  • Working knowledge of C++11.
  • Git.

Familiarity with the following would be advantageous, but is not required:

  • Knowledge of C++14/17 (including templates and move semantics).
  • CMake.
  • Understanding of multi-thread programming.

Course Topics

  • Course aim
    • The aim of this course is to provide students with a solid understanding of the fundamentals and best practices of parallelism in C++, common parallel patterns and how to develop for heterogeneous architectures. Students will gain an understanding of both the fundamentals of parallelism as well as a practical knowledge of writing parallel algorithms for both CPU and GPU architectures using a variety of modern C++ features for parallelism and concurrency. Students will also gain a working understanding of the C++ standard threading library and the SYCL programming model.
  • Course objectives
    1. Understanding of why parallelism is important.
    2. Understanding of parallelism fundamentals.
    3. Understanding of parallel patterns
    4. Understanding of parallel algorithms.
    5. Understanding of heterogeneous system architectures.
    6. Understanding of asynchronous programming.
    7. Understanding of the challenges of using heterogeneous systems.
    8. Understanding of the C++ threading library.
    9. Understanding of the SYCL programming model.
  • Course outcomes
    1. Understanding of why parallelism is important.
      1. Understand the current landscape of computer architectures and their limitations.
      2. Understand the performance benefits of parallelism.
      3. Understand when and where parallelism is appropriate.
    2. Understanding of parallelism fundamentals.
      1. Understand the difference between parallelism and concurrency.
      2. Understand the difference between task parallelism and data parallelism.
      3. Understand the balance of productivity, efficiency and portability.
    3. Understanding of parallel patterns.
      1. Understand the importance of parallel patterns.
      2. Understand common parallel patterns such as map, reduce, scatter, gather and scan, among others.
    4. Understanding of parallel algorithms.
      1. Understand how to to apply parallelism to common algorithms.
      2. Understand how alter the behavior of an algorithm using execution policies.
    5. Understanding of heterogeneous system architectures.
      1. Understand the program execution and memory model of non CPU architectures, like GPUs.
      2. Understand SIMD execution and its benefits and limitations.
    6. Understanding of asynchronous programming.
      1. Understand how to execute a work asynchronously.
      2. Understand how to wait on the completion of asynchronous work.
      3. Understand how to execute both task and data parallel work.
    7. Understanding of the challenges of using heterogeneous systems
      1. Understand the challenges of executing code on a remote device.
      2. Understand how code can be offloaded to a remote device.
      3. Understand the effects of latency between different memory regions and important considerations for data movement.
      4. Understand the importance of coalesced data access and how to achieve it.
      5. Understand the importance of cache locality and how to achieve it.
    8. Understanding of the C++ threading library.
      1. Understand how to create threads and wait on threads.
      2. Understand how to share data between threads
    9. Understanding of the SYCL programming model.
      1. Understand how to write a variety of C++ applications for heterogeneous systems using the SYCL programming model.

Course Instructors

Gordon BrownGordon Brown is a senior software engineer at Codeplay Software specializing in heterogeneous programming models for C++. He has been involved in the standardization of the Khronos standard SYCL and the development of Codeplay’s implementation of the standard from its inception. More recently he has been involved in the efforts within SG1/SG14 to standardize execution and to bring heterogeneous computing to C++.
Michael WongMichael Wong is VP of R&D at Codeplay Software. He is a current Director and VP of ISOCPP , and a senior member of the C++ Standards Committee with more then 15 years of experience. He chairs the WG21 SG5 Transactional Memory and SG14 Games Development/Low Latency/Financials C++ groups and is the co-author of a number C++/OpenMP/Transactional memory features including generalized attributes, user-defined literals, inheriting constructors, weakly ordered memory models, and explicit conversion operators. He has published numerous research papers and is the author of a book on C++11. He has been an invited speaker and keynote at numerous conferences. He is currently the editor of SG1 Concurrency TS and SG5 Transactional Memory TS. He is also the Chair of the SYCL standard and all Programming Languages for Standards Council of Canada. Previously, he was CEO of OpenMP involved with taking OpenMP toward Acceelerator support and the Technical Strategy Architect responsible for moving IBM’s compilers to Clang/LLVM after leading IBM’s XL C++ compiler team.