The following publications are possibly variants of this publication:
- Exploiting High-Bandwidth Memory for FPGA-Acceleration of Inference on Sum-Product NetworksLukas Weber, Johannes Wirth, Lukas Sommer, Andreas Koch 0001. ipps 2022: 112-119 [doi]
- Output-Directed Dynamic Quantization for DNN AccelerationBeilei Jiang, Xianwei Cheng, Yuan Li, Jocelyn Zhang, Song Fu, Qing Yang, Mingxiong Liu, Alejandro Olvera. icpp 2023: 645-654 [doi]
- Dynamic Adaptive DNN Surgery for Inference Acceleration on the EdgeChuang Hu, Wei Bao, Dan Wang, Fengming Liu. infocom 2019: 1423-1431 [doi]
- A Latency-Optimized Reconfigurable NoC for In-Memory Acceleration of DNNsSumit K. Mandal, Gokul Krishnan, Chaitali Chakrabarti, Jae-sun Seo, Yu Cao 0001, Ümit Y. Ogras. esticas, 10(3):362-375, 2020. [doi]
- DP-Nets: Dynamic programming assisted quantization schemes for DNN compression and accelerationDingcheng Yang, Wenjian Yu, Xiangyun Ding, Ao Zhou, Xiaoyi Wang. integration, 82:147-154, 2022. [doi]
- Occamy: Memory-efficient GPU Compiler for DNN InferenceJaeho Lee, Shinnung Jeong, Seungbin Song, Kunwoo Kim, Heelim Choi, Youngsok Kim, Hanjun Kim 0001. dac 2023: 1-6 [doi]
- ADDA: Adaptive Distributed DNN Inference Acceleration in Edge Computing EnvironmentHuitian Wang, Guangxing Cai, Zhaowu Huang, Fang Dong. icpads 2019: 438-445 [doi]