Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism

Jing Li 0093, Weifa Liang, Yuchen Li 0003, Zichuan Xu, Xiaohua Jia, Song Guo 0001. Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism. IEEE Trans. Mob. Comput., 22(5):3017-3030, May 2023. [doi]

References

No references recorded for this publication.

Cited by

No citations of this publication recorded.