Distill, Adapt, Distill: Training Small, In-Domain Models for Neural Machine Translation

Mitchell A. Gordon, Kevin Duh. Distill, Adapt, Distill: Training Small, In-Domain Models for Neural Machine Translation. In Alexandra Birch, Andrew M. Finch, Hiroaki Hayashi, Kenneth Heafield, Marcin Junczys-Dowmunt, Ioannis Konstas, Xian Li, Graham Neubig, Yusuke Oda, editors, Proceedings of the Fourth Workshop on Neural Generation and Translation, NGT@ACL 2020, Online, July 5-10, 2020. pages 110-118, Association for Computational Linguistics, 2020. [doi]

Abstract

Abstract is missing.