CUDA Concurrency -- Part 7 of 9 CUDA Training Series, July 21, 2020
CUDA® is a parallel computing platform and programming model that extends C++ to allow developers to program GPUs with a familiar programming language and simple APIs.
NVIDIA will present a 9-part CUDA training series intended to help new and existing GPU programmers understand the main concepts of the CUDA platform and its programming model. Each part will include a 1-hour presentation and example exercises. The exercises are meant to reinforce the material from the presentation and can be completed during a 1-hour hands-on session following each lecture (for in-person participants) or on your own (for remote participants). OLCF and NERSC will both be holding in-person events for each part of the series. Please see the topics of each part and dates here.
Part 7 of 9: CUDA Concurrency
Date and Time: 10 am - 12 pm (Pacific time), Tuesday, July 21, 2020
The format of this event will be online only. Concurrency refers to the ability to perform multiple CUDA operations simultaneously. This allows the user to do things like run multiple kernels on a single GPU or overlap data transfer and computation. This module introduces CUDA Streams and asynchronous functions that can be used to achieve concurrency in your applications.
After the presentation, there will be a hands-on session where participants can complete example exercises meant to reinforce the presented concepts and get help from the experts.
Please register here. This event is online only.
Remote Connection Information
Remote Connection Information: TBA
- Slides: TBA (to be posted shortly before the event)
- Recording: TBA (to be posted shortly after the event)
- Exercises: The example exercises for this module can be found in the "exercises/hw7" folder of this GitHub repo. (to be available shortly before the event)
- Survey: Please help us to provide feedback by completing a 3-min short survey (access via the "Survey" tab located in the middle of this page)