Revisiting Distributed Synchronous SGD

arXiv: Learning, Volume abs/1702.05800, 2017.

Cited by: 8|Bibtex|Views110
EI
Other Links: dblp.uni-trier.de|academic.microsoft.com

Abstract:

Distributed training of deep learning models on large-scale training data is typically conducted with asynchronous stochastic optimization to maximize the rate of updates, at the cost of additional noise introduced from asynchrony. In contrast, the synchronous approach is often thought to be impractical due to idle time wasted on waiting ...More

Code:

Data:

Your rating :
0

 

Tags
Comments