DOCmT5: Document-Level Pretraining of Multilingual Language Models

Chia-Hsuan Lee 0001, Aditya Siddhant, Viresh Ratnakar, Melvin Johnson. DOCmT5: Document-Level Pretraining of Multilingual Language Models. In Marine Carpuat, Marie-Catherine de Marneffe, Iván Vladimir Meza Ruíz, editors, Findings of the Association for Computational Linguistics: NAACL 2022, Seattle, WA, United States, July 10-15, 2022. pages 425-437, Association for Computational Linguistics, 2022. [doi]

@inproceedings{0001SRJ22,
  title = {DOCmT5: Document-Level Pretraining of Multilingual Language Models},
  author = {Chia-Hsuan Lee 0001 and Aditya Siddhant and Viresh Ratnakar and Melvin Johnson},
  year = {2022},
  url = {https://aclanthology.org/2022.findings-naacl.32},
  researchr = {https://researchr.org/publication/0001SRJ22},
  cites = {0},
  citedby = {0},
  pages = {425-437},
  booktitle = {Findings of the Association for Computational Linguistics: NAACL 2022, Seattle, WA, United States, July 10-15, 2022},
  editor = {Marine Carpuat and Marie-Catherine de Marneffe and Iván Vladimir Meza Ruíz},
  publisher = {Association for Computational Linguistics},
  isbn = {978-1-955917-76-6},
}