1. Demystifying the Placement Policies of the NVIDIA GPU Thread Block Scheduler for Concurrent Kernels
- Author
-
Robert J. Walls, Guin Gilman, Tian Guo, and Samuel S. Ogden
- Subjects
010302 applied physics ,Computer Networks and Communications ,Computer science ,02 engineering and technology ,Parallel computing ,Thread (computing) ,Pascal (programming language) ,01 natural sciences ,Execution time ,020202 computer hardware & architecture ,Scheduling (computing) ,Kernel (linear algebra) ,Resource (project management) ,Hardware and Architecture ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Turing ,computer ,Software ,Block (data storage) ,computer.programming_language - Abstract
In this work, we empirically derive the scheduler's behavior under concurrent workloads for NVIDIA's Pascal, Volta, and Turing microarchitectures. In contrast to past studies that suggest the scheduler uses a round-robin policy to assign thread blocks to streaming multiprocessors (SMs), we instead find that the scheduler chooses the next SM based on the SM's local resource availability. We show how this scheduling policy can lead to significant, and seemingly counter-intuitive, performance degradation; for example, a decrease of one thread per block resulted in a 3.58X increase in execution time for one kernel in our experiments. We hope that our work will be useful for improving the accuracy of GPU simulators and aid in the development of novel scheduling algorithms.
- Published
- 2021
- Full Text
- View/download PDF