The following publications are possibly variants of this publication:
- Accelerating Sparse Deep Neural Network Inference Using GPU Tensor CoresYufei Sun, Long Zheng 0003, Qinggang Wang, Xiangyu Ye, Yu Huang 0013, Pengcheng Yao, Xiaofei Liao, Hai Jin 0001. hpec 2022: 1-7 [doi]
- Boomerang: On-Demand Cooperative Deep Neural Network Inference for Edge Intelligence on the Industrial Internet of ThingsLiekang Zeng, En Li, Zhi Zhou, Xu Chen 0004. network, 33(5):96-103, 2019. [doi]
- Channel Pruning for Accelerating Very Deep Neural NetworksYihui He, Xiangyu Zhang, Jian Sun. iccv 2017: 1398-1406 [doi]
- Accelerating deep neural network filter pruning with mask-aware convolutional computations on modern CPUsXiu Ma, Guangli Li, Lei Liu, Huaxiao Liu, Xueying Wang. ijon, 505:375-387, 2022. [doi]
- Discrimination-aware Channel Pruning for Deep Neural NetworksZhuangwei Zhuang, Mingkui Tan, Bohan Zhuang, Jing Liu, Yong Guo, Qingyao Wu, JunZhou Huang, Jin-Hui Zhu. nips 2018: 883-894 [doi]
- Toward Decentralized and Collaborative Deep Learning Inference for Intelligent IoT DevicesYakun Huang, Xiuquan Qiao, Schahram Dustdar, Jianwei Zhang, Jiulin Li. network, 36(1):59-68, 2022. [doi]