TY - GEN
T1 - RAV4D
T2 - 2024 IEEE Radar Conference, RadarConf 2024
AU - Zhou, Yi
AU - Song, Ningfei
AU - Ma, Jieming
AU - Man, Ka Lok
AU - López-Benítez, Miguel
AU - Yu, Limin
AU - Yue, Yutao
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/5
Y1 - 2024/5
N2 - Indoor multi-person tracking is a widely explored area of research. However, publicly available datasets are either oversimplified or provide only visual data. To fill this gap, our paper presents the RAV4D dataset, a novel multimodal dataset that includes data from radar, microphone arrays, and stereo cameras. This dataset is characterised by the provision of 3D positions, Euler angles and Doppler velocities. By integrating these different data types, RAV 4D aims to exploit the synergistic and complementary capabilities of these modalities to improve tracking performance. The development of RAV4D addresses two main challenges: sensor calibration and 3D annotation. A novel calibration target is designed to effectively calibrate the radar, stereo camera and microphone array. In addition, a visually guided annotation framework is proposed to address the challenge of annotating radar data. This framework uses head positions, heading orientation and depth information from stereo cameras and radar to establish accurate ground truth for multimodal tracking trajectories. The dataset is publicly available at https://zenodo.org/records/l0208199.
AB - Indoor multi-person tracking is a widely explored area of research. However, publicly available datasets are either oversimplified or provide only visual data. To fill this gap, our paper presents the RAV4D dataset, a novel multimodal dataset that includes data from radar, microphone arrays, and stereo cameras. This dataset is characterised by the provision of 3D positions, Euler angles and Doppler velocities. By integrating these different data types, RAV 4D aims to exploit the synergistic and complementary capabilities of these modalities to improve tracking performance. The development of RAV4D addresses two main challenges: sensor calibration and 3D annotation. A novel calibration target is designed to effectively calibrate the radar, stereo camera and microphone array. In addition, a visually guided annotation framework is proposed to address the challenge of annotating radar data. This framework uses head positions, heading orientation and depth information from stereo cameras and radar to establish accurate ground truth for multimodal tracking trajectories. The dataset is publicly available at https://zenodo.org/records/l0208199.
KW - Multiple Object Tracking
KW - Radar Tracking
KW - Sensor Fusion
KW - Speaker Tracking
UR - http://www.scopus.com/inward/record.url?scp=85196831863&partnerID=8YFLogxK
U2 - 10.1109/RadarConf2458775.2024.10549701
DO - 10.1109/RadarConf2458775.2024.10549701
M3 - Conference Proceeding
AN - SCOPUS:85196831863
T3 - Proceedings of the IEEE Radar Conference
BT - RadarConf 2024 - 2024 IEEE Radar Conference, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 6 May 2024 through 10 May 2024
ER -