Sparsifying Transformer Models with Trainable Representation Pooling

Michal Pietruszka, Lukasz Borchmann, Lukasz Garncarek. Sparsifying Transformer Models with Trainable Representation Pooling. In Smaranda Muresan, Preslav Nakov, Aline Villavicencio, editors, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022. pages 8616-8633, Association for Computational Linguistics, 2022. [doi]

@inproceedings{PietruszkaBG22,
  title = {Sparsifying Transformer Models with Trainable Representation Pooling},
  author = {Michal Pietruszka and Lukasz Borchmann and Lukasz Garncarek},
  year = {2022},
  url = {https://aclanthology.org/2022.acl-long.590},
  researchr = {https://researchr.org/publication/PietruszkaBG22},
  cites = {0},
  citedby = {0},
  pages = {8616-8633},
  booktitle = {Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022},
  editor = {Smaranda Muresan and Preslav Nakov and Aline Villavicencio},
  publisher = {Association for Computational Linguistics},
  isbn = {978-1-955917-21-6},
}