作者
Cong Xie, Shuai Zheng, Sanmi Koyejo, Indranil Gupta, Mu Li, Haibin Lin
发表日期
2020
期刊
Advances in Neural Information Processing Systems
卷号
33
页码范围
12593-12603
简介
The scalability of Distributed Stochastic Gradient Descent (SGD) is today limited by communication bottlenecks. We propose a novel SGD variant:\underline {C} ommunication-efficient\underline {S} GD with\underline {E} rror\underline {R} eset, or\underline {CSER}. The key idea in CSER is first a new technique called``error reset''that adapts arbitrary compressors for SGD, producing bifurcated local models with periodic reset of resulting local residual errors. Second we introduce partial synchronization for both the gradients and the models, leveraging advantages from them. We prove the convergence of CSER for smooth non-convex problems. Empirical results show that when combined with highly aggressive compressors, the CSER algorithms accelerate the distributed training by nearly for CIFAR-100, and by for ImageNet.
引用总数
2019202020212022202320241189128
学术搜索中的文章
C Xie, S Zheng, S Koyejo, I Gupta, M Li, H Lin - Advances in Neural Information Processing Systems, 2020