TY - GEN
T1 - Fast Depth Estimation of Object via Neural Network Perspective Projection
AU - Han, Yu
AU - Chen, Yaran
AU - Li, Haoran
AU - Ma, Mingjun
AU - Zhao, Dongbin
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - In autonomous driving and mobile robotic systems, obtaining the depths of objects in real-time is crucial. The current network-based methods usually design complex network to achieve 3D object detection or monocular depth estimation for the whole image, resulting in too slow to be applied to mobile robots. The perspective projection-based method can achieve real-time, which calculates the object depth based on the camera parameters, the object sizes in the world coordinates and in image coordinates. While it relies heavily on the accuracy of object size in images coordinates, and the size is usually obtained with errors through detector network. Combining the perspective projection-based methods and network-based methods, we propose a fast object depth estimation method by designing a neural network to learn perspective projection, called Fast-Depth-NPP: 1) Instead of considering the whole image, we only consider the local depth of the image; 2) Using local image patches as network inputs avoids measurement errors of object size with detector; 3) the use of global information is enhanced by incorporating position encoding. Our method is validated on the mobile robot public dataset Neurons Perception dataset, achieving excellent results and meeting the real-time requirements.
AB - In autonomous driving and mobile robotic systems, obtaining the depths of objects in real-time is crucial. The current network-based methods usually design complex network to achieve 3D object detection or monocular depth estimation for the whole image, resulting in too slow to be applied to mobile robots. The perspective projection-based method can achieve real-time, which calculates the object depth based on the camera parameters, the object sizes in the world coordinates and in image coordinates. While it relies heavily on the accuracy of object size in images coordinates, and the size is usually obtained with errors through detector network. Combining the perspective projection-based methods and network-based methods, we propose a fast object depth estimation method by designing a neural network to learn perspective projection, called Fast-Depth-NPP: 1) Instead of considering the whole image, we only consider the local depth of the image; 2) Using local image patches as network inputs avoids measurement errors of object size with detector; 3) the use of global information is enhanced by incorporating position encoding. Our method is validated on the mobile robot public dataset Neurons Perception dataset, achieving excellent results and meeting the real-time requirements.
KW - Convolutional Neural Network
KW - Depth Estimation
KW - Object Detection
UR - http://www.scopus.com/inward/record.url?scp=85137750834&partnerID=8YFLogxK
U2 - 10.1109/DDCLS55054.2022.9858358
DO - 10.1109/DDCLS55054.2022.9858358
M3 - Conference Proceeding
AN - SCOPUS:85137750834
T3 - Proceedings of 2022 IEEE 11th Data Driven Control and Learning Systems Conference, DDCLS 2022
SP - 788
EP - 794
BT - Proceedings of 2022 IEEE 11th Data Driven Control and Learning Systems Conference, DDCLS 2022
A2 - Sun, Mingxuan
A2 - Chen, Zengqiang
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 11th IEEE Data Driven Control and Learning Systems Conference, DDCLS 2022
Y2 - 3 August 2022 through 5 August 2022
ER -