MaskCLIP: Masked Self-Distillation Advances Contrastive Language-Image Pretraining

Xiaoyi Dong, Jianmin Bao, Yinglin Zheng, Ting Zhang 0002, Dongdong Chen, Hao Yang 0036, Ming Zeng 0008, Weiming Zhang 0001, Lu Yuan, Dong Chen, Fang Wen 0001, Nenghai Yu. MaskCLIP: Masked Self-Distillation Advances Contrastive Language-Image Pretraining. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023. pages 10995-11005, IEEE, 2023. [doi]

Abstract

Abstract is missing.