Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning

Bingbing Li, Zhenglun Kong, Tianyun Zhang, Ji Li 0006, Zhengang Li, Hang Liu, Caiwen Ding. Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning. In Trevor Cohn, Yulan He, Yang Liu, editors, Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, EMNLP 2020, Online Event, 16-20 November 2020. pages 3187-3199, Association for Computational Linguistics, 2020. [doi]

@inproceedings{LiKZ0LLD20,
  title = {Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning},
  author = {Bingbing Li and Zhenglun Kong and Tianyun Zhang and Ji Li 0006 and Zhengang Li and Hang Liu and Caiwen Ding},
  year = {2020},
  url = {https://www.aclweb.org/anthology/2020.findings-emnlp.286/},
  researchr = {https://researchr.org/publication/LiKZ0LLD20},
  cites = {0},
  citedby = {0},
  pages = {3187-3199},
  booktitle = {Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, EMNLP 2020, Online Event, 16-20 November 2020},
  editor = {Trevor Cohn and Yulan He and Yang Liu},
  publisher = {Association for Computational Linguistics},
  isbn = {978-1-952148-90-3},
}