Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Sep 23;21(19):6355.
doi: 10.3390/s21196355.

Semantics Aware Dynamic SLAM Based on 3D MODT

Affiliations

Semantics Aware Dynamic SLAM Based on 3D MODT

Muhammad Sualeh et al. Sensors (Basel). .

Abstract

The idea of SLAM (Simultaneous Localization and Mapping) being a solved problem revolves around the static world assumption, even though autonomous systems are gaining environmental perception capabilities by exploiting the advances in computer vision and data-driven approaches. The computational demands and time complexities remain the main impediment in the effective fusion of the paradigms. In this paper, a framework to solve the dynamic SLAM problem is proposed. The dynamic regions of the scene are handled by making use of Visual-LiDAR based MODT (Multiple Object Detection and Tracking). Furthermore, minimal computational demands and real-time performance are ensured. The framework is tested on the KITTI Datasets and evaluated against the publicly available evaluation tools for a fair comparison with state-of-the-art SLAM algorithms. The results suggest that the proposed dynamic SLAM framework can perform in real-time with budgeted computational resources. In addition, the fused MODT provides rich semantic information that can be readily integrated into SLAM.

Keywords: 3D multiple object detection; dynamic SLAM; multiple object tracking; semantics.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
Framework for joint MODT and SLAM.
Figure 2
Figure 2
Proposed MODT integrated Dynamic SLAM framework.
Figure 3
Figure 3
3D MODT module in the proposed framework.
Figure 4
Figure 4
Camera-LiDAR Fusion Module.
Figure 5
Figure 5
Visual-LiDAR based Mask generated in comparison with Mask R CNN.
Figure 6
Figure 6
Dynamic SLAM with Visual-LiDAR based dynamic object mask.
Figure 7
Figure 7
Estimated trajectories of sequence 20 from KITTI Tracking Dataset.

References

    1. Walcott-Bryant A., Kaess M., Johannsson H., Leonard J.J. Dynamic Pose Graph SLAM Long-Term Mapping in Low Dynamic Environments; Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Vilamoura-Algarve, Portugal. 7–12 October 2012; pp. 1871–1878.
    1. Qiping C., Yinfei X., Shifeng G., Jie B., Qiang S. Sensing system of environmental perception technologies for driverless vehicle: A review of state of the art and challenges. Sens. Actuators A Phys. 2021;102:112566.
    1. He K., Gkioxari G., Dollár P., Girshick R. Mask R-CNN; Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV); Venice, Italy. 22–29 October 2017; pp. 2980–2988.
    1. Davison A.J. SLAM++: Simultaneous Localisation in addition, Mapping at the Level of Objects; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Portland, OR, USA. 23–28 June 2013; pp. 1352–1359.
    1. Gálvez-López D., Salas M., Tardós J.D., Montiel J.M.M. Real-time monocular object SLAM. Robot. Auton. Syst. 2016;75:435–449. doi: 10.1016/j.robot.2015.08.009. - DOI

LinkOut - more resources