Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Jul 26;25(15):4628.
doi: 10.3390/s25154628.

Image Alignment Based on Deep Learning to Extract Deep Feature Information from Images

Affiliations

Image Alignment Based on Deep Learning to Extract Deep Feature Information from Images

Lin Zhu et al. Sensors (Basel). .

Abstract

To overcome the limitations of traditional image alignment methods in capturing deep semantic features, a deep feature information image alignment network (DFA-Net) is proposed. This network aims to enhance image alignment performance through multi-level feature learning. DFA-Net is based on the deep residual architecture and introduces spatial pyramid pooling to achieve cross-scalar feature fusion, effectively enhancing the feature's adaptability to scale. A feature enhancement module based on the self-attention mechanism is designed, with key features that exhibit geometric invariance and high discriminative power, achieved through a dynamic weight allocation strategy. This improves the network's robustness to multimodal image deformation. Experiments on two public datasets, MSRS and RoadScene, show that the method performs well in terms of alignment accuracy, with the RMSE metrics being reduced by 0.661 and 0.473, and the SSIM, MI, and NCC improved by 0.155, 0.163, and 0.211; and 0.108, 0.226, and 0.114, respectively, compared with the benchmark model. The visualization results validate the significant improvement in the features' visual quality and confirm the method's advantages in terms of stability and discriminative properties of deep feature extraction.

Keywords: deep learning; feature extraction; image alignment; infrared and visible images.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflicts of interest.

Figures

Figure 1
Figure 1
DFA-Net network structure.
Figure 2
Figure 2
Structure of the deep feature information extraction network.
Figure 3
Figure 3
Structure of the spatial information fusion module.
Figure 4
Figure 4
Structure of the feature enhancement module.
Figure 5
Figure 5
Visualization of the ablation experiment results. Red boxes indicate people. Green boxes indicate objects such as streetlights, vehicles, etc.
Figure 6
Figure 6
Quantitative comparison with SOTA methods in the MSRS dataset.
Figure 7
Figure 7
Quantitative comparison with SOTA methods in the RoadScene dataset.
Figure 8
Figure 8
Image alignment visualization results on the MSRS dataset. Red boxes indicate people, green boxes indicate environmental references such as vehicles, streetlights, etc. (ac) highlight the spatial alignment effect of pedestrian targets; (df) verify the accurate alignment of multi-level targets in composite scenes of people and backgrounds; and (g) shows the alignment effect of vehicle targets.
Figure 9
Figure 9
Image alignment visualization results on the RoadScene dataset. Red boxes indicate people, green boxes indicate vehicles, foliage, buildings and other objects. (ac) highlight the registration between vehicles and backgrounds; (df) demonstrate the consistency of multi-object registration involving people, vehicles, and backgrounds; and (g) shows the registration results for people and backgrounds.

Similar articles

References

    1. Ma J., Jiang X., Fan A., Jiang J., Yan J. Image matching from handcrafted to deep features: A survey. Int. J. Comput. Vis. 2021;129:23–79. doi: 10.1007/s11263-020-01359-2. - DOI
    1. Jhan J.P., Rau J.Y. A generalized tool for accurate and efficient image registration of UAV multi-lens multispectral cameras by N-SURF matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021;14:6353–6362. doi: 10.1109/JSTARS.2021.3079404. - DOI
    1. Ma W., Wang K., Li J., Yang S.X., Li J., Song L., Li Q. Infrared and visible image fusion technology and application: A review. Sensors. 2023;23:599. doi: 10.3390/s23020599. - DOI - PMC - PubMed
    1. Li H., Wu X.J., Kittler J. RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf. Fusion. 2021;73:72–86. doi: 10.1016/j.inffus.2021.02.023. - DOI
    1. Wang Z., Feng X., Xu G., Wu Y. A robust visible and infrared image matching algorithm for power equipment based on phase congruency and scale-invariant feature. Opt. Lasers Eng. 2023;164:107517. doi: 10.1016/j.optlaseng.2023.107517. - DOI

LinkOut - more resources