Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Mar 20;15(1):9688.
doi: 10.1038/s41598-025-92239-7.

A hybrid object detection approach for visually impaired persons using pigeon-inspired optimization and deep learning models

Affiliations

A hybrid object detection approach for visually impaired persons using pigeon-inspired optimization and deep learning models

Abdullah M Alashjaee et al. Sci Rep. .

Abstract

Visually challenged persons include a significant part of the population, and they exist all over the globe. Recently, technology has demonstrated its occurrence in each field, and state-of-the-art devices aid humans in their everyday lives. However, visually impaired people cannot view things around their atmospheres; they can only imagine the roaming surroundings. Furthermore, web-based applications are advanced to certify their protection. Using the application, the consumer can spin the requested task to share her/his position with the family members while threatening confidentiality. Through this application, visually challenged people's family members can follow their actions (acquire snapshots and position) while staying at their residences. A deep learning (DL) technique is trained with manifold images of entities highly related to the VIPs. Training images are amplified and physically interpreted to bring more strength to the trained method. This study proposes a Hybrid Approach to Object Detection for Visually Impaired Persons Using Attention-Driven Deep Learning (HAODVIP-ADL) technique. The major intention of the HAODVIP-ADL technique is to deliver a reliable and precise object detection system that supports the visually impaired person in navigating their surroundings safely and effectively. The presented HAODVIP-ADL method initially utilizes bilateral filtering (BF) for the image pre-processing stage to reduce noise while preserving edges for clarity. For object detection, the HAODVIP-ADL method employs the YOLOv10 framework. In addition, the backbone fusion of feature extraction models such as CapsNet and InceptionV3 is implemented to capture diverse spatial and contextual information. The bi-directional long short-term memory and multi-head attention (MHA-BiLSTM) approach is utilized to classify the object detection process. Finally, the hyperparameter tuning process is performed using the pigeon-inspired optimization (PIO) approach to advance the classification performance of the MHA-BiLSTM approach. The experimental results of the HAODVIP-ADL method are analyzed, and the outcomes are evaluated using the Indoor Objects Detection dataset. The experimental validation of the HAODVIP-ADL method portrayed a superior accuracy value of 99.74% over the existing methods.

Keywords: Deep learning; Feature extraction; Object detection; Pigeon-inspired optimization; Visually impaired persons.

PubMed Disclaimer

Conflict of interest statement

Declarations. Competing interests: The authors declare no competing interests.

Figures

Fig. 1
Fig. 1
Overall workflow of the HAODVIP-ADL model.
Fig. 2
Fig. 2
Structure of BF model.
Fig. 3
Fig. 3
Architecture of YOLOv10 technique.
Fig. 4
Fig. 4
Structure of CapsNet approach.
Fig. 5
Fig. 5
Overall structure of the MHA-BiLSTM approach.
Fig. 6
Fig. 6
Steps involved in the PIO model.
Fig. 7
Fig. 7
Sample images.
Fig. 8
Fig. 8
Object detection images.
Fig. 9
Fig. 9
Confusion matrix of HAODVIP-ADL methodology (a–f) Epochs 500–3000.
Fig. 10
Fig. 10
Average of HAODVIP-ADL technique under distinct epochs.
Fig. 11
Fig. 11
formula image curve of HAODVIP-ADL approach (a–f) Epochs 500–3000.
Fig. 12
Fig. 12
Loss analysis of HAODVIP-ADL approach (a–f) Epochs 500–3000.
Fig. 13
Fig. 13
Comparative analysis of HAODVIP-ADL approach with existing techniques.
Fig. 14
Fig. 14
mAP outcome of HAODVIP-ADL technique with recent models.
Fig. 15
Fig. 15
Mean IoU outcome of HAODVIP-ADL technique with recent models.
Fig. 16
Fig. 16
CT analysis of HAODVIP-ADL approach with existing techniques.

References

    1. Bashiri, F. S. et al. Object detection to assist visually impaired people: A deep neural network adventure. In Advances in Visual Computing: 13th International Symposium, ISVC 2018, Las Vegas, NV, USA, November 19–21, 2018, Proceedings 13 (Springer, 2018).
    1. Fadhel, Z., Attia, H. & Ali, Y. H. Optimized and comprehensive fake review detection based on Harris Hawks optimization integrated with machine learning techniques. J. Cybersecur. Inform. Manag.15 (1), 1 (2025).
    1. Ashiq, F. et al. CNN-based object recognition and tracking system to assist visually impaired people. IEEE Access.10, 14819–14834 (2022).
    1. Masud, U., Saeed, T., Malaikah, H. M., Islam, F. U. & Abbas, G. Smart assistive system for visually impaired people obstruction avoidance through object detection and classification. IEEE Access.10, 13428–13441 (2022).
    1. Kumar, N. & Jain, A. A deep learning-based model to assist blind people in their navigation. J. Inf. Technol. Educ. Innov. Pract.21, 95–114 (2022).

MeSH terms

LinkOut - more resources