A Slowdown Prediction Method to Improve Memory Aware Scheduling
Doktorsavhandling, 2016

Scientific and technological advances in the area of integrated circuits have allowed the performance of microprocessors to grow exponentially since the late 1960’s. However, the imbalance between processor and memory bus capacity has increased in recent years. The increasing on-chip-parallelism of multi-core processors has turned the memory subsystem into a key factor for achieving high performance. When two or more processes share the memory subsystem their execution times typically increase, even at relatively low levels of memory traffic. Current research shows that a throughput increase of up to 40% is possible if the job-scheduler can minimizes the slowdown caused by memory contention in industrial multi-core systems such as high performance clusters, datacenters or clouds. In order to optimize the throughput the job-scheduler has to know how much slower the process will execute when co-scheduled on the same server as other processes. Consequently, unless the slowdown is known, or can be fairly well estimated, the scheduling becomes pure guesswork and the performance suffers. The central question addressed in this thesis is how the slowdown caused by memory traffic interference between processes executing on the same server can be predicted and to what extent. This thesis presents and evaluates a new slowdown prediction method which estimates how much longer a program will execute when co-scheduled on the same multi-core server as another program. The method measures how external memory traffic affects a program by generating different levels of synthetic memory traffic while observing the change in execution time. Based on the observations it makes a first order prediction of how much slowdown the program will experience when exposed to external memory traffic. Experimental results show that the method’s predictions correlate well with the real measured slowdowns. Furthermore, it is shown that scheduling based on the new slowdown prediction method yields a higher throughput than three other techniques suggested for avoiding co-scheduling slowdowns caused by memory contention. Finally, a novel scheme is suggested to avoid some of the worst co-schedules, thus increasing the system throughput.

last level cache

memory bandwidth

performance evaluation

multi-core processor

slowdown aware scheduling

resource contention

co-scheduling

EC, Hörsalsvägen 11, Chalmers
Opponent: Prof. Nectarios Koziris, National Technical University of Athens, Greece

Författare

Andreas de Blanche

Chalmers, Data- och informationsteknik, Datorteknik

Styrkeområden

Informations- och kommunikationsteknik

Ämneskategorier

Datorsystem

ISBN

978-91-7597-369-2

Technical report - Department of Computer Science and Engineering, Chalmers University of Technology and Göteborg University: 123D

Doktorsavhandlingar vid Chalmers tekniska högskola. Ny serie: 4050

EC, Hörsalsvägen 11, Chalmers

Opponent: Prof. Nectarios Koziris, National Technical University of Athens, Greece

Mer information

Skapat

2017-10-07