Potential Obstacle Detection Using RGB to Depth Image Encoder-Decoder Network: Application to Unmanned Aerial Vehicles
- PMID: 36081162
- PMCID: PMC9460703
- DOI: 10.3390/s22176703
Potential Obstacle Detection Using RGB to Depth Image Encoder-Decoder Network: Application to Unmanned Aerial Vehicles
Abstract
In this work, a new method is proposed that allows the use of a single RGB camera for the real-time detection of objects that could be potential collision sources for Unmanned Aerial Vehicles. For this purpose, a new network with an encoder-decoder architecture has been developed, which allows rapid distance estimation from a single image by performing RGB to depth mapping. Based on a comparison with other existing RGB to depth mapping methods, the proposed network achieved a satisfactory trade-off between complexity and accuracy. With only 6.3 million parameters, it achieved efficiency close to models with more than five times the number of parameters. This allows the proposed network to operate in real time. A special algorithm makes use of the distance predictions made by the network, compensating for measurement inaccuracies. The entire solution has been implemented and tested in practice in an indoor environment using a micro-drone equipped with a front-facing RGB camera. All data and source codes and pretrained network weights are available to download. Thus, one can easily reproduce the results, and the resulting solution can be tested and quickly deployed in practice.
Keywords: RGB to depth mapping; Unmanned Aerial Vehicles; deep neural network; depth prediction; encoder–decoder network; obstacle detection.
Conflict of interest statement
The author declares no conflict of interest.
Figures






References
-
- Ronneberger O., Fischer P., Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N., Hornegger J., Wells W.M., Frangi A.F., editors. Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015. Springer International Publishing; Cham, Switzerland: 2015. pp. 234–241.
-
- Wang J., Li B., Zhou Y., Meng Q., Rende S.F., Rocco E. Real-time and Embedded Compact Deep Neural Networks for Seagrass Monitoring; Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC); Toronto, ON, Canada. 11–14 October 2020; pp. 3570–3575. - DOI
-
- Levinshtein A., Chang C., Phung E., Kezele I., Guo W., Aarabi P. Real-Time Deep Hair Matting on Mobile Devices; Proceedings of the 2018 15th Conference on Computer and Robot Vision (CRV); Toronto, ON, Canada. 8–10 May 2018; pp. 1–7. - DOI
-
- Yao Z., He K., Zhou H., Zhang Z., Zhu G., Xing C., Zhang J., Zhang Z., Shao B., Tao Y., et al. Eye3DVas: Three-dimensional reconstruction of retinal vascular structures by integrating fundus image features; Proceedings of the Frontiers in Optics/Laser Science; Washington, DC, USA. 14–17 September 2020; Washington, DC, USA: Optica Publishing Group; 2020. p. JTu1B.22.
-
- Hachaj T., Stolińska A., Andrzejewska M., Czerski P. Deep Convolutional Symmetric Encoder-Decoder Neural Networks to Predict Students’ Visual Attention. Symmetry. 2021;13:2246. doi: 10.3390/sym13122246. - DOI
LinkOut - more resources
Full Text Sources