Abstract
We present a parallel asynchronous Stochastic Gradient Descent algorithm for shared memory architectures. Different from previous asynchronous algorithms, we consider the case where the gradient updates are not particularly sparse. In the context of the MagmaDNN framework, we compare the parallel efficiency of the asynchronous implementation with that of the traditional synchronous implementation. Tests are performed for training deep neural networks on multicore CPUs and GPU devices.
Original language | English |
---|---|
Title of host publication | 2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) |
Pages | 1-4 |
DOIs | |
Publication status | E-pub ahead of print - 28 Jul 2020 |
Event | 2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) - New Orleans, LA, USA Duration: 18 May 2020 → 22 May 2020 |
Conference
Conference | 2020 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) |
---|---|
Period | 18/05/20 → 22/05/20 |