TY - JOUR
T1 - Scalpel : High Performance Contention-Aware Task Co-Scheduling for Shared Cache Hierarchy
AU - Liu, Song
AU - Ma, Jie
AU - Zhang, Zengyuan
AU - Wan, Xinhe
AU - Zhao, Bo
AU - Wu, Weiguo
PY - 2025/2/1
Y1 - 2025/2/1
N2 - For scientific computing applications that consist of many loosely coupled tasks, efficient scheduling is critical to achieve high performance and good quality of service (QoS). One of the challenges for co-running tasks is the frequent contention for shared cache hierarchy of multi-core processors. Such contention significantly increases cache miss rate and therefore, results in performance deterioration for computational tasks. This paper presents Scalpel, a contention-aware task grouping and co-scheduling approach for efficient task scheduling on shared cache hierarchy. Scalpel utilizes the shared cache access features of tasks to group them in a heuristic way, which reduces the contention within groups by achieving equal shared cache locality, while maintaining load balancing between groups. Based thereon, it proposes a two-level scheduling strategy to schedule groups to processors and assign tasks to available cores in a timely manner, while considering the impact of task scheduling on shared cache locality to minimize task execution time. Experiments show that Scalpel reduces the shared cache miss rate by up to 2.143 and optimizes the execution time by up to 1.533 for scientific computing benchmarks, compared to several baseline approaches.
AB - For scientific computing applications that consist of many loosely coupled tasks, efficient scheduling is critical to achieve high performance and good quality of service (QoS). One of the challenges for co-running tasks is the frequent contention for shared cache hierarchy of multi-core processors. Such contention significantly increases cache miss rate and therefore, results in performance deterioration for computational tasks. This paper presents Scalpel, a contention-aware task grouping and co-scheduling approach for efficient task scheduling on shared cache hierarchy. Scalpel utilizes the shared cache access features of tasks to group them in a heuristic way, which reduces the contention within groups by achieving equal shared cache locality, while maintaining load balancing between groups. Based thereon, it proposes a two-level scheduling strategy to schedule groups to processors and assign tasks to available cores in a timely manner, while considering the impact of task scheduling on shared cache locality to minimize task execution time. Experiments show that Scalpel reduces the shared cache miss rate by up to 2.143 and optimizes the execution time by up to 1.533 for scientific computing benchmarks, compared to several baseline approaches.
KW - Computers
KW - Dynamic scheduling
KW - Hardware
KW - Load management
KW - Multicore processing
KW - Processor scheduling
KW - Quality of service
KW - Resource management
KW - Scientific computing
KW - Software
UR - http://www.scopus.com/inward/record.url?scp=85209904006&partnerID=8YFLogxK
U2 - 10.1109/TC.2024.3500381
DO - 10.1109/TC.2024.3500381
M3 - Article
SN - 2326-3814
VL - 74
SP - 678
EP - 690
JO - IEEE Transactions on Computers
JF - IEEE Transactions on Computers
IS - 2
M1 - 10756745
ER -