Interval-Asynchrony: Delimited Intervals of Localised Asynchrony for Fast Parallel SGD
Paper i proceeding, 2026

Stochastic gradient descent (SGD) is a crucial optimisation algorithm due to its ubiquity in machine learning applications. Parallelism is a popular approach to scale SGD, but the standard synchronous formulation struggles due to significant synchronisation overhead. For this reason, asynchronous implementations are increasingly common. These provide an improvement in throughput at the expense of introducing stale gradients which reduce model accuracy. Previous approaches to mitigate the downsides of asynchronous processing include adaptively adjusting the number of worker threads or the learning rate, but at their core these are still fully asynchronous and hence still suffer from lower accuracy due to more staleness. We propose Interval-Asynchrony, a semi-asynchronous method which retains high throughput while reducing gradient staleness, both on average as well as with a hard upper bound. Our method achieves this by introducing periodic asynchronous intervals, within which SGD is executed asynchronously, but between which gradient computations may not cross. The size of these intervals determines the degree of asynchrony, providing us with an adjustable scale. Since the optimal interval size varies over time, we additionally provide two strategies for dynamic adjustment thereof. We evaluate our method against several baselines on the CIFAR-10 and CIFAR-100 datasets, and demonstrate a 32% decrease in training time as well as improved scalability up to 128 threads.

Parallel SGD

Asynchronous Data Processing

Staleness

Parallel Algorithms

Författare

Jacob Stacey Garby

Göteborgs universitet

Chalmers, Data- och informationsteknik, Dator- och nätverkssystem

Philippas Tsigas

Göteborgs universitet

Chalmers, Data- och informationsteknik, Dator- och nätverkssystem

Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

03029743 (ISSN) 16113349 (eISSN)

Vol. 15901 LNCS 236-249
9783031998560 (ISBN)

31st International Conference on Parallel and Distributed Computing, Euro-Par 2025
Dresden, Germany,

Relaxed Semantics Across the Data Analytics Stack (RELAX-DN)

Europeiska kommissionen (EU) (EC/HE/101072456), 2023-03-01 -- 2027-03-01.

Ämneskategorier (SSIF 2025)

Datavetenskap (datalogi)

Datorsystem

Reglerteknik

DOI

10.1007/978-3-031-99857-7_17

Relaterade dataset

Artifact of the paper: Interval-Asynchrony: Delimited Intervals of Localised Asynchrony for Fast Parallel SGD [dataset]

URI: https://zenodo.org/records/15576941

Mer information

Senast uppdaterat

2025-09-05