Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2020 Jul 29;20(15):4220.
doi: 10.3390/s20154220.

Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review

Affiliations
Review

Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review

Jamil Fayyad et al. Sensors (Basel). .

Abstract

Autonomous vehicles (AV) are expected to improve, reshape, and revolutionize the future of ground transportation. It is anticipated that ordinary vehicles will one day be replaced with smart vehicles that are able to make decisions and perform driving tasks on their own. In order to achieve this objective, self-driving vehicles are equipped with sensors that are used to sense and perceive both their surroundings and the faraway environment, using further advances in communication technologies, such as 5G. In the meantime, local perception, as with human beings, will continue to be an effective means for controlling the vehicle at short range. In the other hand, extended perception allows for anticipation of distant events and produces smarter behavior to guide the vehicle to its destination while respecting a set of criteria (safety, energy management, traffic optimization, comfort). In spite of the remarkable advancements of sensor technologies in terms of their effectiveness and applicability for AV systems in recent years, sensors can still fail because of noise, ambient conditions, or manufacturing defects, among other factors; hence, it is not advisable to rely on a single sensor for any of the autonomous driving tasks. The practical solution is to incorporate multiple competitive and complementary sensors that work synergistically to overcome their individual shortcomings. This article provides a comprehensive review of the state-of-the-art methods utilized to improve the performance of AV systems in short-range or local vehicle environments. Specifically, it focuses on recent studies that use deep learning sensor fusion algorithms for perception, localization, and mapping. The article concludes by highlighting some of the current trends and possible future research directions.

Keywords: autonomous vehicles; deep learning; localization and mapping; perception; self-driving cars; sensor fusion.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
The six levels of autonomous vehicles as described by the Society of Automobile Engineers (SAE) [3], their definitions, and the features in each level.
Figure 2
Figure 2
Full autonomous navigation system. Sensor technology and sensor fusion overview. V2V, vehicle-to-vehicle; V2I, vehicle-to-infrastructure.
Figure 3
Figure 3
The different stages in the perception and the decision process.
Figure 4
Figure 4
Sensor fusion architecture described in terms of the three different levels. Level one represents early fusion, level two represents halfway fusion, and level three represents late fusion.
Figure 5
Figure 5
Theories of uncertainty for modeling and processing of “imperfect” data.
Figure 6
Figure 6
Classical approaches for sensor fusion algorithms.
Figure 7
Figure 7
Common deep learning sensor fusion algorithms used in autonomous vehicle applications. R-CNN: Region-Based CNN; SPP-Net: Spatial Pyramid Pooling network; YOLO: You only look once; SSD: Single-Shot Multibox Detector; DSSD: Deconvolutional Single-Shot Multibox Detector; LSTM: Long-Short Term Memory; GRU: Gated Recurrent Unit.
Figure 8
Figure 8
The different processing layers in a convolutional neural network (CNN) for object detection and identification.
Figure 9
Figure 9
Timeline development of CNN-based detectors. R-FCN: Region-Based Fully Connected Convolution Network.
Figure 10
Figure 10
The architecture of the region-based CNN (R-CNN) algorithm.
Figure 11
Figure 11
The architecture of spatial pyramid pooling (SPP-Net) algorithm.
Figure 12
Figure 12
The architecture of fast R-CNN.
Figure 13
Figure 13
The architecture of faster R-CNN.
Figure 14
Figure 14
The architecture of “you only look once” (YOLO) algorithm.
Figure 15
Figure 15
The architecture of the SSD algorithm. The CNN Network is VGG-16.

References

    1. Singh S. Critical Reasons for Crashes Investigated in the National Motor Vehicle Crash Causation Survey. Traffic Safety Facts Crash Stats. Report No. DOT HS 812 115. National Center for Statistics and Analysis; Washington, DC, USA: 2015.
    1. Olia A., Abdelgawad H., Abdulhai B., Razavi S.N. Assessing the Potential Impacts of Connected Vehicles: Mobility, Environmental, and Safety Perspectives. J. Intell. Transp. Syst. 2016;20:229–243. doi: 10.1080/15472450.2015.1062728. - DOI
    1. Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles (J3016 Ground Vehicle Standard)—SAE Mobilus. [(accessed on 23 October 2019)]; Available online: https://saemobilus.sae.org/content/j3016_201806.
    1. Learn More About General Motors’ Approach to Safely Putting Self-Driving Cars on the Roads in 2019. [(accessed on 23 October 2019)]; Available online: https://www.gm.com/our-stories/self-driving-cars.html.
    1. Autopilot. [(accessed on 23 October 2019)]; Available online: https://www.tesla.com/autopilot.

LinkOut - more resources