E.T.: re-thinking self-attention for transformer models on GPUs

Shiyang Chen, Shaoyi Huang, Santosh Pandey, Bingbing Li, Guang R. Gao, Long Zheng 0001, Caiwen Ding, Hang Liu 0001. E.T.: re-thinking self-attention for transformer models on GPUs. In Bronis R. de Supinski, Mary W. Hall, Todd Gamblin, editors, SC '21: The International Conference for High Performance Computing, Networking, Storage and Analysis, St. Louis, Missouri, USA, November 14 - 19, 2021. pages 25, ACM, 2021. [doi]

Authors

Shiyang Chen

This author has not been identified. Look up 'Shiyang Chen' in Google

Shaoyi Huang

This author has not been identified. Look up 'Shaoyi Huang' in Google

Santosh Pandey

This author has not been identified. Look up 'Santosh Pandey' in Google

Bingbing Li

This author has not been identified. Look up 'Bingbing Li' in Google

Guang R. Gao

This author has not been identified. Look up 'Guang R. Gao' in Google

Long Zheng 0001

This author has not been identified. Look up 'Long Zheng 0001' in Google

Caiwen Ding

This author has not been identified. Look up 'Caiwen Ding' in Google

Hang Liu 0001

This author has not been identified. Look up 'Hang Liu 0001' in Google