The following publications are possibly variants of this publication:
- Imitation Learning for Non-Autoregressive Neural Machine TranslationBingzhen Wei, Mingxuan Wang, Hao Zhou, Junyang Lin, Xu Sun 0001. acl 2019: 1304-1312 [doi]
- I2D2: Inductive Knowledge Distillation with NeuroLogic and Self-ImitationChandra Bhagavatula, Jena D. Hwang, Doug Downey, Ronan Le Bras 0001, Ximing Lu, Lianhui Qin, Keisuke Sakaguchi, Swabha Swayamdipta, Peter West, Yejin Choi 0001. acl 2023: 9614-9630 [doi]
- Revisiting Knowledge Distillation for Autoregressive Language ModelsQihuang Zhong, Liang Ding 0006, Li Shen 0008, Juhua Liu, Bo Du 0001, Dacheng Tao. acl 2024: 10900-10913 [doi]
- Learning from Imperfect Data: Towards Efficient Knowledge Distillation of Autoregressive Language Models for Text-to-SQLQihuang Zhong, Kunfeng Chen, Liang Ding 0006, Juhua Liu, Bo Du 0001, Dacheng Tao. emnlp 2024: 6874-6885 [doi]