The following publications are possibly variants of this publication:
- A cross-modal fusion based approach with scale-aware deep representation for RGB-D crowd counting and density estimationShihui Zhang, He Li, Weihang Kong. eswa, 180:115071, 2021. [doi]
- RGB-T Multi-Modal Crowd Counting Based on TransformerZhengyi Liu, Wei Wu, Yacheng Tan, Guanghui Zhang. bmvc 2022: 427 [doi]
- CSA-Net: Cross-modal scale-aware attention-aggregated network for RGB-T crowd countingHe Li, Junge Zhang, Weihang Kong, Jienan Shen, Yuguang Shao. eswa, 213(Part):119038, 2023. [doi]
- Cross-Modal Collaborative Representation Learning and a Large-Scale RGBT Benchmark for Crowd CountingLingbo Liu, Jiaqi Chen, Hefeng Wu, Guanbin Li, Chenglong Li, Liang Lin. cvpr 2021: 4823-4833 [doi]
- Cross-modal collaborative representation and multi-level supervision for crowd countingShufang Li, Zhengping Hu, Mengyao Zhao, Shuai Bi, Zhe Sun. sivp, 17(3):601-608, April 2023. [doi]
- Learning with Balanced Criss-Cross Attention for Cross-Modality Crowd CountingXin Zeng, Wanjun Zhang, Huake Wang, Xiaoli Bian. itcc2 2023: 53-58 [doi]