Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2022 Sep 14;22(18):6969.
doi: 10.3390/s22186969.

A Comprehensive Survey of Depth Completion Approaches

Affiliations
Review

A Comprehensive Survey of Depth Completion Approaches

Muhammad Ahmed Ullah Khan et al. Sensors (Basel). .

Abstract

Depth maps produced by LiDAR-based approaches are sparse. Even high-end LiDAR sensors produce highly sparse depth maps, which are also noisy around the object boundaries. Depth completion is the task of generating a dense depth map from a sparse depth map. While the earlier approaches focused on directly completing this sparsity from the sparse depth maps, modern techniques use RGB images as a guidance tool to resolve this problem. Whilst many others rely on affinity matrices for depth completion. Based on these approaches, we have divided the literature into two major categories; unguided methods and image-guided methods. The latter is further subdivided into multi-branch and spatial propagation networks. The multi-branch networks further have a sub-category named image-guided filtering. In this paper, for the first time ever we present a comprehensive survey of depth completion methods. We present a novel taxonomy of depth completion approaches, review in detail different state-of-the-art techniques within each category for depth completion of LiDAR data, and provide quantitative results for the approaches on KITTI and NYUv2 depth completion benchmark datasets.

Keywords: depth completion; depth maps; image-guidance.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 9
Figure 9
Qualitative comparison of the top three reported methods on KITTI depth completion test set, including (b) RigNet [33], (c) GuideNet [51], and (d) DySPN [60]. Given sparse depth maps and the input guidance color images (a), the methods output dense depth predictions (1st row). The corresponding error maps (2nd row) are taken from the KITTI leaderboard for comparison. Warmer color represents higher error.
Figure 1
Figure 1
First Column shows the RGB images from two different scenes, the middle column contains the sparse depth maps produced from LiDAR. The last column shows the predicted dense depth maps for the corresponding scenes. (a) RGB Image. (b) LiDAR sparse Depth Map. (c) Prediction.
Figure 2
Figure 2
Approaches to depth completion problem. Unguided approaches utilize either only LiDAR information or confidence maps and LiDAR information for dense depth completion. The image-guided methods (multi-branch and spatial propagation networks) employ guidance images (RGB, semantic maps, surface normals) to guide the process of depth completion. The multi-branch networks can be further divided into guided image filtering methods, which aim to learn useful kernels from one modality and apply it to other modalities.
Figure 3
Figure 3
Early fusion between RGB image and LiDAR sparse depth. At first, both modalities are fused and then sent to the Deep Neural Network for dense depth completion.
Figure 4
Figure 4
Sequential fusion between RGB image and LiDAR sparse depth map. The RGB branch produces color depth, which along with LiDAR sparse depth map, is sent to the depth branch to estimate the final dense depth map.
Figure 5
Figure 5
Late fusion between RGB image and LiDAR sparse depth map. It consists of two separate branches to process RGB images and LiDAR sparse depth maps. Both of the branches produce dense depth maps, which are fused to produce a final dense depth map.
Figure 6
Figure 6
Deep Fusion between RGB image and LiDAR sparse depth map. Each modality is passed from a dedicated branch. The features from the decoder of the RGB branch are fused into the encoder of the depth branch. The symbol “F” represents the fusion operation. Common choices for fusion operation include addition or concatenation. However, complex fusion schemes can also be employed. By the guidance of the RGB branch, the depth branch produces a final dense depth map.
Figure 7
Figure 7
KITTI depth completion benchmark. Part (a) shows the aligned RGB images. Part (b) depicts the sparse LiDAR depth maps, whereas Part (c) represents the dense ground-truth depth maps. Colorization is applied on LiDAR sparse depth maps and corresponding ground-truth to generate visualizations. The highlighted areas are used to show the sparsity in KITTI depth completion benchmark.
Figure 8
Figure 8
Nyu-v2 depth dataset. Part (a) shows the aligned RGB images. Part (b) depicts the sparse Kinect depth maps, which are generated by randomly sampling only 500 points from the ground truth. Part (c) represents the fully dense ground-truth depth maps. Colorization is applied on Kinect sparse depth maps and corresponding ground-truth to generate visualizations. (a) RGB Image. (b) Kinect sparse Depth Map. (c) Ground-truth.

References

    1. Cui Z., Heng L., Yeo Y.C., Geiger A., Pollefeys M., Sattler T. Real-time dense mapping for self-driving vehicles using fisheye cameras; Proceedings of the 2019 International Conference on Robotics and Automation (ICRA); Montreal, QC, Canada. 20–24 May 2019; pp. 6087–6093.
    1. Häne C., Heng L., Lee G.H., Fraundorfer F., Furgale P., Sattler T., Pollefeys M. 3D visual perception for self-driving cars using a multi-camera system: Calibration, mapping, localization, and obstacle detection. Image Vis. Comput. 2017;68:14–27. doi: 10.1016/j.imavis.2017.07.003. - DOI
    1. Wang K., Zhang Z., Yan Z., Li X., Xu B., Li J., Yang J. Regularizing Nighttime Weirdness: Efficient Self-supervised Monocular Depth Estimation in the Dark; Proceedings of the IEEE/CVF International Conference on Computer Vision; Virtual. 11–17 October 2021; pp. 16055–16064.
    1. Song X., Wang P., Zhou D., Zhu R., Guan C., Dai Y., Su H., Li H., Yang R. Apollocar3d: A large 3d car instance understanding benchmark for autonomous driving; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Long Beach, CA, USA. 16–17 June 2019; pp. 5452–5462.
    1. Liao Y., Huang L., Wang Y., Kodagoda S., Yu Y., Liu Y. Parse geometry from a line: Monocular depth estimation with partial laser observation; Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA); Singapore. 29 May–3 June 2017; pp. 5059–5066.

LinkOut - more resources