Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism

Jing Li 0093, Weifa Liang, Yuchen Li 0003, Zichuan Xu, Xiaohua Jia, Song Guo 0001. Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism. IEEE Trans. Mob. Comput., 22(5):3017-3030, May 2023. [doi]

Authors

Jing Li 0093

This author has not been identified. Look up 'Jing Li 0093' in Google

Weifa Liang

This author has not been identified. Look up 'Weifa Liang' in Google

Yuchen Li 0003

This author has not been identified. Look up 'Yuchen Li 0003' in Google

Zichuan Xu

This author has not been identified. Look up 'Zichuan Xu' in Google

Xiaohua Jia

This author has not been identified. Look up 'Xiaohua Jia' in Google

Song Guo 0001

This author has not been identified. Look up 'Song Guo 0001' in Google