Title: A tasks reordering model to reduce transfers overhead on GPUs
Abstract: The compute capabilities of current GPUs allow exploiting concurrency when several independent tasks are simultaneously launched. These tasks are typically composed by data transfer commands and kernel computation commands. In this paper we develop a run-time approach to optimize the concurrency between data transfers and kernel computation operations in a multithreaded scenario where each CPU thread is sending tasks to the GPU. Our solution is based on a temporal execution model for concurrent tasks that is able to establish the tasks execution order that minimizes the total execution time, including data transfers. Moreover, a heuristic to select the best order has been developed, which is able to improve the execution time achieved by the hardware scheduler of current NVIDIA cards. Our approach obtains performance improvements, under real workloads, of up to 19% with respect to the execution using multiple hardware queues managed by Hyper-Q.
Publication Year: 2017
Publication Date: 2017-06-29
Language: en
Type: article
Indexed In: ['crossref']
Access and Citation
Cited By Count: 6
AI Researcher Chatbot
Get quick answers to your questions about the article from our AI researcher chatbot