Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2025 Apr 20;25(8):2611.
doi: 10.3390/s25082611.

A Review of Vision-Based Multi-Task Perception Research Methods for Autonomous Vehicles

Affiliations
Review

A Review of Vision-Based Multi-Task Perception Research Methods for Autonomous Vehicles

Hai Wang et al. Sensors (Basel). .

Abstract

Multi-task perception technology for autonomous driving significantly improves the ability of autonomous vehicles to understand complex traffic environments by integrating multiple perception tasks, such as traffic object detection, drivable area segmentation, and lane detection. The collaborative processing of these tasks not only improves the overall performance of the perception system but also enhances the robustness and real-time performance of the system. In this paper, we review the research progress in the field of vision-based multi-task perception for autonomous driving and introduce the methods of traffic object detection, drivable area segmentation, and lane detection in detail. Moreover, we discuss the definition, role, and classification of multi-task learning. In addition, we analyze the design of classical network architectures and loss functions for multi-task perception, introduce commonly used datasets and evaluation metrics, and discuss the current challenges and development prospects of multi-task perception. By analyzing these contents, this paper aims to provide a comprehensive reference framework for researchers in the field of autonomous driving and encourage more research work on multi-task perception for autonomous driving.

Keywords: autonomous driving; deep learning; detection; drivable area segmentation; multi-task learning.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflicts of interest.

Figures

Figure 1
Figure 1
Common algorithms for object detection.
Figure 2
Figure 2
Common algorithms for semantic segmentation.
Figure 3
Figure 3
Classification of multi-task learning sharing mechanisms: (a) hard parameter sharing; (b) soft parameter sharing; and (c) hybrid sharing.
Figure 4
Figure 4
A partial image of the BDD100K dataset: (a) daytime; (b) night; (c) rain; (d) snow; (e) city road; (f) motorway; (g) sunny; and (h) overcast.
Figure 5
Figure 5
Schematic of IoU definition.

Similar articles

References

    1. Liang J., Li Y., Yin G., Xu L., Lu Y., Feng J., Shen T., Cai G. A MAS-Based Hierarchical Architecture for the Cooperation Control of Connected and Automated Vehicles. IEEE Trans. Veh. Technol. 2023;72:1559–1573. doi: 10.1109/TVT.2022.3211733. - DOI
    1. Liu H., Yan S., Shen Y., Li C., Zhang Y., Hussain F. Model Predictive Control System Based on Direct Yaw Moment Control for 4WID Self-Steering Agriculture Vehicle. Int. J. Agric. Biol. Eng. 2021;14:175–181. doi: 10.25165/j.ijabe.20211402.5283. - DOI
    1. Kiran B.R., Sobh I., Talpaert V., Mannion P., Sallab A.A.A., Yogamani S., Perez P. Deep Reinforcement Learning for Autonomous Driving: A Survey. IEEE Trans. Intell. Transport. Syst. 2022;23:4909–4926. doi: 10.1109/TITS.2021.3054625. - DOI
    1. Wang H., Gu J., Wang M. A Review on the Application of Computer Vision and Machine Learning in the Tea Industry. Front. Sustain. Food Syst. 2023;7:1172543. doi: 10.3389/fsufs.2023.1172543. - DOI
    1. Wei L., Jianping H., Jiaxin L., Rencai Y., Tengfei Z., Mengjiao Y., Jing L. Method for the Navigation Line Recognition of the Ridge without Crops via Machine Vision. Int. J. Agric. Biol. Eng. 2024;17:230–239. doi: 10.25165/j.ijabe.20241702.7480. - DOI

LinkOut - more resources