The following publications are possibly variants of this publication:
- SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head PruningHanrui Wang 0002, Zhekai Zhang, Song Han 0003. hpca 2021: 97-110 [doi]
- Dynamic Head: Unifying Object Detection Heads With AttentionsXiyang Dai, Yinpeng Chen, Bin Xiao, Dongdong Chen 0001, Mengchen Liu, Lu Yuan, Lei Zhang 0001. cvpr 2021: 7373-7382 [doi]
- Classification of Heads in Multi-head Attention MechanismsFeihu Huang, Min Jiang, Fang Liu, Dian Xu, Zimeng Fan, Yonghao Wang. ksem 2022: 681-692 [doi]
- Hydra Attention: Efficient Attention with Many HeadsDaniel Bolya, Cheng-Yang Fu, Xiaoliang Dai, Peizhao Zhang, Judy Hoffman. eccv 2023: 35-49 [doi]
- Attention to Head Locations for Crowd CountingYoumei Zhang, Chunluan Zhou, Faliang Chang, Alex C. Kot, Wei Zhang 0066. icig 2019: 727-737 [doi]
- Spatial Attention Network for Head DetectionRongchun Li, Biao Zhang, Zhen Huang, Xiang Zhao, Peng Qiao, Yong Dou. pcm 2018: 547-557 [doi]