Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism

Jing Li 0093, Weifa Liang, Yuchen Li 0003, Zichuan Xu, Xiaohua Jia, Song Guo 0001. Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism. IEEE Trans. Mob. Comput., 22(5):3017-3030, May 2023. [doi]

Abstract

Abstract is missing.