Multi-level Distillation of Semantic Knowledge for Pre-training Multilingual Language Model

Mingqi Li, Fei Ding, Dan Zhang, Long Cheng, Hongxin Hu, Feng Luo 0001. Multi-level Distillation of Semantic Knowledge for Pre-training Multilingual Language Model. In Yoav Goldberg, Zornitsa Kozareva, Yue Zhang, editors, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11. pages 3097-3106, Association for Computational Linguistics, 2022. [doi]

Abstract

Abstract is missing.