Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism

Jing Li 0093, Weifa Liang, Yuchen Li 0003, Zichuan Xu, Xiaohua Jia, Song Guo 0001. Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism. IEEE Trans. Mob. Comput., 22(5):3017-3030, May 2023. [doi]

@article{LiLLXJG23,
  title = {Throughput Maximization of Delay-Aware DNN Inference in Edge Computing by Exploring DNN Model Partitioning and Inference Parallelism},
  author = {Jing Li 0093 and Weifa Liang and Yuchen Li 0003 and Zichuan Xu and Xiaohua Jia and Song Guo 0001},
  year = {2023},
  month = {May},
  doi = {10.1109/TMC.2021.3125949},
  url = {https://doi.org/10.1109/TMC.2021.3125949},
  researchr = {https://researchr.org/publication/LiLLXJG23},
  cites = {0},
  citedby = {0},
  journal = {IEEE Trans. Mob. Comput.},
  volume = {22},
  number = {5},
  pages = {3017-3030},
}