TY - JOUR
T1 - Structerf-SLAM
T2 - Neural implicit representation SLAM for structural environments
AU - Wang, Haocheng
AU - Cao, Yanlong
AU - Wei, Xiaoyao
AU - Shou, Yejun
AU - Shen, Lingfeng
AU - Xu, Zhijie
AU - Ren, Kai
N1 - Publisher Copyright:
© 2024
PY - 2024/4
Y1 - 2024/4
N2 - In recent years, research on simultaneous localization and mapping (SLAM) using neural implicit representation has shown promising outcomes due to its smooth mapping and low memory consumption, particularly suitable for structured environments with limited boundaries. However, there is currently no implicit SLAM that can effectively utilize prior structural constraints to accurately build 3D maps. In this study, we propose an RGB-D dense tracking and mapping approach, Structerf-SLAM, that combines visual odometry with neural implicit representation. Our scene representation consists of dual-layer feature grids and pre-trained decoders that decode the interpolated features into RGB and depth values. Moreover, structured planar constraints are integrated. In the tracking stage, utilizing the three-dimensional plane features under the Manhattan assumption achieves more stable and rapid data association, consequently resolving the tracking misalignment issue in textureless regions (e.g., floor, wall, etc.). In the mapping stage, by enforcing planar consistency, the depth predicted by the neural radiation field is well-fitted by a plane, resulting in smoother and more realistic map reconstruction. Experiments on synthetic and real scene datasets demonstrate competitive results of Structerf-SLAM in both mapping and tracking quality.
AB - In recent years, research on simultaneous localization and mapping (SLAM) using neural implicit representation has shown promising outcomes due to its smooth mapping and low memory consumption, particularly suitable for structured environments with limited boundaries. However, there is currently no implicit SLAM that can effectively utilize prior structural constraints to accurately build 3D maps. In this study, we propose an RGB-D dense tracking and mapping approach, Structerf-SLAM, that combines visual odometry with neural implicit representation. Our scene representation consists of dual-layer feature grids and pre-trained decoders that decode the interpolated features into RGB and depth values. Moreover, structured planar constraints are integrated. In the tracking stage, utilizing the three-dimensional plane features under the Manhattan assumption achieves more stable and rapid data association, consequently resolving the tracking misalignment issue in textureless regions (e.g., floor, wall, etc.). In the mapping stage, by enforcing planar consistency, the depth predicted by the neural radiation field is well-fitted by a plane, resulting in smoother and more realistic map reconstruction. Experiments on synthetic and real scene datasets demonstrate competitive results of Structerf-SLAM in both mapping and tracking quality.
KW - Implicit scene reconstruction
KW - Neural implicit representation
KW - Structural constraints
KW - Visual SLAM
UR - http://www.scopus.com/inward/record.url?scp=85186330095&partnerID=8YFLogxK
U2 - 10.1016/j.cag.2024.103893
DO - 10.1016/j.cag.2024.103893
M3 - Article
AN - SCOPUS:85186330095
SN - 0097-8493
VL - 119
JO - Computers and Graphics (Pergamon)
JF - Computers and Graphics (Pergamon)
M1 - 103893
ER -