The following publications are possibly variants of this publication:
- GradSA: Gradient Sparsification and Accumulation for Communication-Efficient Distributed Deep LearningBo Liu, Wenbin Jiang, Shaofeng Zhao, Hai Jin 0001, Bingsheng He. GPC 2020: 77-91 [doi]
- Communication-Efficient Distributed Deep Learning with Merged Gradient Sparsification on GPUsShaohuai Shi, Qiang Wang, Xiaowen Chu, Bo Li, Yang Qin, Ruihao Liu, Xinxiao Zhao. infocom 2020: 406-415 [doi]
- Gradient Sparsification for Communication-Efficient Distributed OptimizationJianqiao Wangni, Jialei Wang, Ji Liu, Tong Zhang. nips 2018: 1306-1316 [doi]
- Dual-Way Gradient Sparsification for Asynchronous Distributed Deep LearningZijie Yan, Danyang Xiao, Mengqiang Chen, Jieying Zhou, Weigang Wu. icpp 2020: [doi]
- SSD-SGD: Communication Sparsification for Distributed Deep Learning TrainingYemao Xu, Dezun Dong, Dongsheng Wang, Shi Xu, Enda Yu, Weixia Xu, Xiangke Liao. taco, 20(1), March 2023. [doi]