The following publications are possibly variants of this publication:
- Making Vision Transformers Efficient from A Token Sparsification ViewShuning Chang, Pichao Wang, Ming Lin, Fan Wang, David Junhao Zhang, Rong Jin 0001, Mike Zheng Shou. cvpr 2023: 6195-6205 [doi]
- Dynamic Spatial Sparsification for Efficient Vision Transformers and Convolutional Neural NetworksYongming Rao, Zuyan Liu, Wenliang Zhao, Jie Zhou 0001, Jiwen Lu. pami, 45(9):10883-10897, September 2023. [doi]
- No Token Left Behind: Efficient Vision Transformer via Dynamic Token IdlingXuwei Xu, Changlin Li, Yudong Chen 0002, Xiaojun Chang, Jiajun Liu, Sen Wang 0001. ausai 2024: 28-41 [doi]
- AdaFormer: Efficient Transformer with Adaptive Token Sparsification for Image Super-resolutionXiaotong Luo, Zekun Ai, Qiuyuan Liang, Ding Liu, Yuan Xie 0006, Yanyun Qu, Yun Fu 0001. AAAI 2024: 4009-4016 [doi]
- Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision TransformersSifan Long, Zhen Zhao, Jimin Pi, Shengsheng Wang 0001, Jingdong Wang 0001. cvpr 2023: 10334-10343 [doi]