Course details
Parallel Computations on GPU
PCG Acad. year 2020/2021 Winter semester 5 credits
The course covers the architecture and programming of graphics processing units by the NVidia and partially AMD. First, the architecture of GPUs is studied in detail. Then, the model of the program execution using hierarchical thread organisation and the SIMT model is discussed. Next, the memory hierarchy and synchronization techniques are described. After that, the course explains novel techniques of dynamic parallelism and data-flow processing concluded by practical usage of multi-GPU systems in environments with shared (NVLink) and distributed (MPI) memory. The second part of the course is devoted to high level programming techniques and libraries based on the OpenACC technology.
Guarantor
Course coordinator
Language of instruction
Completion
Time span
- 26 hrs lectures
- 12 hrs pc labs
- 14 hrs projects
Assessment points
- 60 pts final exam (written part)
- 15 pts mid-term test (written part)
- 25 pts projects
Department
Lecturer
Instructor
Subject specific learning outcomes and competences
Knowledge of the parallel programming on GPUs in the area of general purpose computing, orientation in the area of accelerated systems, libraries and tools.
Understanding of hardware limitations having impact on the efficiency of software solutions.
Learning objectives
To familiarize yourself with the architecture and programming of graphics processing unit in the area of general purpose computuing using the NVidia libraries and OpenACC standard. To learn how to design and implement accelerated programs exploiting the potential of GPUs. To gain knowledge about the available libraries for programming on GPUs.
Why is the course taught
The future of computation systems ranging from ordinary PC up to top supercomputers is seen in heterogeneous systems where the sequential parts and the logic is processed by CPUs while the computational parts are offload to accelerators, in this case GPUs. This course will teach you the architecture and software libraries for programming graphics processing units in the area of general purpose computations.
Prerequisite knowledge and skills
Knowledge gained in courses AVS and partially in PRL and PPP.
Study literature
- Current PPT slides for lectures
- Nvidia CUDA documentation: https://docs.nvidia.com/cuda/
- OpenACC documentation: https://www.openacc.org/
- Kirk, D., and Hwu, W.: Programming Massively Parallel Processors: A Hands-on Approach, Elsevier, 2010, s. 256, ISBN: 978-0-12-381472-2. download.
- Storti,D., and Yurtoglu, M.: CUDA for Engineers: An Introduction to High-Performance Parallel Computing, Addison-Wesley Professional; 1 edition, 2015. ISBN 978-0134177410. link.
Syllabus of lectures
- Architecture of graphics processing units.
- CUDA programming model, tread execution.
- CUDA memory hierarchy.
- Synchronization and reduction.
- Dynamic parallelism and unified memory.
- Design and optimization of GPU algorithms.
- Stream processing, computation-communication overlapping.
- Multi-GPU systems.
- Nvidia Thrust library.
- OpenACC basics.
- OpenACC memory management.
- Code optimization with OpenACC.
- Libraries and tools for GPU programming.
Syllabus of computer exercises
- CUDA: Memory transfers, simple kernels
- CUDA: Shared memory
- CUDA: Texture and constant memory
- CUDA: Dynamic parallelism and unified memory.
- OpenACC: basic techniques.
- OpenACC: advanced techniques.
Syllabus - others, projects and individual work of students
- Development of an application in Nvidia CUDA
- Development of an application in OpenACC
Progress assessment
Assessment of two projects, 14 hours in total and, computer laboratories and a midterm examination.
Exam prerequisites:
To get 20 out of 40 points for projects and midterm examination.
Controlled instruction
- Missed labs can be substituted in alternative dates.
- There will be a place for missed labs in the last week of the semester.
Exam prerequisites
To get 20 out of 40 points for projects and midterm examination.
Course inclusion in study plans