3D Random Occlusion and Multi-Layer Projection for Deep Multi-Camera Pedestrian Localization

Rui Qiu, Ming Xu*, Yuyao Yan, Jeremy S. Smith, Xi Yang

*Corresponding author for this work

Research output: Contribution to journalArticle


Although deep-learning based methods for monocular pedestrian detection have made great progress, they are still vulnerable to heavy occlusions. Using multi-view information fusion is a potential solution but has limited applications, due to the lack of annotated training samples in existing multi-view datasets, which increases the risk of overfitting. To address this problem, a data augmentation method is proposed to randomly generate 3D cylinder occlusions, on the ground plane, which are of the average size of pedestrians and projected to multiple views, to relieve the impact of overfitting in the training. Moreover, the feature map of each view is projected to multiple parallel planes at different heights, by using homographies, which allows the CNNs to fully utilize the features across the height of each pedestrian to infer the locations of pedestrians on the ground plane. The proposed 3DROM method has a greatly improved performance in comparison with the state-of-the-art deep-learning based methods for multi-view pedestrian detection.
Original languageEnglish
Article numberarXiv:2207.10895
Number of pages16
JournalarXiv preprint
Publication statusPublished - 22 Jul 2022


Dive into the research topics of '3D Random Occlusion and Multi-Layer Projection for Deep Multi-Camera Pedestrian Localization'. Together they form a unique fingerprint.

Cite this