WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing

Sanyuan Chen, Chengyi Wang 0002, Zhengyang Chen, Yu Wu 0012, Shujie Liu 0001, Zhuo Chen 0006, Jinyu Li 0001, Naoyuki Kanda, Takuya Yoshioka, Xiong Xiao, Jian Wu 0027, Long Zhou, Shuo Ren, Yanmin Qian, Yao Qian, Jian Wu, Michael Zeng 0001, Xiangzhan Yu, Furu Wei. WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing. J. Sel. Topics Signal Processing, 16(6):1505-1518, 2022. [doi]

@article{ChenWCWLCLKYXWZ22,
  title = {WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing},
  author = {Sanyuan Chen and Chengyi Wang 0002 and Zhengyang Chen and Yu Wu 0012 and Shujie Liu 0001 and Zhuo Chen 0006 and Jinyu Li 0001 and Naoyuki Kanda and Takuya Yoshioka and Xiong Xiao and Jian Wu 0027 and Long Zhou and Shuo Ren and Yanmin Qian and Yao Qian and Jian Wu and Michael Zeng 0001 and Xiangzhan Yu and Furu Wei},
  year = {2022},
  doi = {10.1109/JSTSP.2022.3188113},
  url = {https://doi.org/10.1109/JSTSP.2022.3188113},
  researchr = {https://researchr.org/publication/ChenWCWLCLKYXWZ22},
  cites = {0},
  citedby = {0},
  journal = {J. Sel. Topics Signal Processing},
  volume = {16},
  number = {6},
  pages = {1505-1518},
}