Autoregressive Knowledge Distillation through Imitation Learning

Alexander Lin, Jeremy Wohlwend, Howard Chen, Tao Lei 0001. Autoregressive Knowledge Distillation through Imitation Learning. In Bonnie Webber, Trevor Cohn, Yulan He, Yang Liu, editors, Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020. pages 6121-6133, Association for Computational Linguistics, 2020. [doi]

@inproceedings{LinWCL20,
  title = {Autoregressive Knowledge Distillation through Imitation Learning},
  author = {Alexander Lin and Jeremy Wohlwend and Howard Chen and Tao Lei 0001},
  year = {2020},
  url = {https://www.aclweb.org/anthology/2020.emnlp-main.494/},
  researchr = {https://researchr.org/publication/LinWCL20},
  cites = {0},
  citedby = {0},
  pages = {6121-6133},
  booktitle = {Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020},
  editor = {Bonnie Webber and Trevor Cohn and Yulan He and Yang Liu},
  publisher = {Association for Computational Linguistics},
  isbn = {978-1-952148-60-6},
}