Interval-Asynchrony: Delimited Intervals of Localised Asynchrony for Fast Parallel SGD
Paper in proceeding, 2026

Stochastic gradient descent (SGD) is a crucial optimisation algorithm due to its ubiquity in machine learning applications. Parallelism is a popular approach to scale SGD, but the standard synchronous formulation struggles due to significant synchronisation overhead. For this reason, asynchronous implementations are increasingly common. These provide an improvement in throughput at the expense of introducing stale gradients which reduce model accuracy. Previous approaches to mitigate the downsides of asynchronous processing include adaptively adjusting the number of worker threads or the learning rate, but at their core these are still fully asynchronous and hence still suffer from lower accuracy due to more staleness. We propose Interval-Asynchrony, a semi-asynchronous method which retains high throughput while reducing gradient staleness, both on average as well as with a hard upper bound. Our method achieves this by introducing periodic asynchronous intervals, within which SGD is executed asynchronously, but between which gradient computations may not cross. The size of these intervals determines the degree of asynchrony, providing us with an adjustable scale. Since the optimal interval size varies over time, we additionally provide two strategies for dynamic adjustment thereof. We evaluate our method against several baselines on the CIFAR-10 and CIFAR-100 datasets, and demonstrate a 32% decrease in training time as well as improved scalability up to 128 threads.

Parallel SGD

Asynchronous Data Processing

Staleness

Parallel Algorithms

Author

Jacob Stacey Garby

University of Gothenburg

Chalmers, Computer Science and Engineering (Chalmers), Computer and Network Systems

Philippas Tsigas

University of Gothenburg

Chalmers, Computer Science and Engineering (Chalmers), Computer and Network Systems

Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

03029743 (ISSN) 16113349 (eISSN)

Vol. 15901 LNCS 236-249
9783031998560 (ISBN)

31st International Conference on Parallel and Distributed Computing, Euro-Par 2025
Dresden, Germany,

Relaxed Semantics Across the Data Analytics Stack (RELAX-DN)

European Commission (EC) (EC/HE/101072456), 2023-03-01 -- 2027-03-01.

Subject Categories (SSIF 2025)

Computer Sciences

Computer Systems

Control Engineering

DOI

10.1007/978-3-031-99857-7_17

Related datasets

Artifact of the paper: Interval-Asynchrony: Delimited Intervals of Localised Asynchrony for Fast Parallel SGD [dataset]

URI: https://zenodo.org/records/15576941

More information

Latest update

9/5/2025 1