Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Jun 27;25(7):985.
doi: 10.3390/e25070985.

SCFusion: Infrared and Visible Fusion Based on Salient Compensation

Affiliations

SCFusion: Infrared and Visible Fusion Based on Salient Compensation

Haipeng Liu et al. Entropy (Basel). .

Abstract

The aim of infrared and visible image fusion is to integrate the complementary information of the two modalities for high-quality fused images. However, many deep learning fusion algorithms have not considered the characteristics of infrared images in low-light scenes, leading to the problems of weak texture details, low contrast of infrared targets and poor visual perception in the existing methods. Therefore, in this paper, we propose a salient compensation-based fusion method that makes sufficient use of the characteristics of infrared and visible images to generate high-quality fused images under low-light conditions. First, we design a multi-scale edge gradient module (MEGB) in the texture mainstream to adequately extract the texture information of the dual input of infrared and visible images; on the other hand, the salient tributary is pre-trained by salient loss to obtain the saliency map based on the salient dense residual module (SRDB) to extract salient features, which is supplemented in the process of overall network training. We propose the spatial bias module (SBM) to fuse global information with local information. Finally, extensive comparison experiments with existing methods show that our method has significant advantages in describing target features and global scenes, the effectiveness of the proposed module is demonstrated by ablation experiments. In addition, we also verify the facilitation of this paper's method for high-level vision on a semantic segmentation task.

Keywords: deep learning; image fusion; infrared and visible images; salient compensation.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
Overall framework for SCFusion. It consists of multiscale edge gradient block (MEGB),salient dense residual block (SRDB), and spatial bias block (SBM). The saliency map generated by the saliency tributary is pre-trained by saliency loss, which is then sent to the main network to generate the fused image with the texture features obtained by MEGB under the joint training of structural similarity loss and content loss.
Figure 2
Figure 2
Multiscale edge gradient block (MEGB). It accomplishes texture detail enhancement by combining the output of the multi-scale with the output of the residual gradient flow.
Figure 3
Figure 3
Salient dense residual block (SRDB). It achieves contrast enhancement by combining attentional features with residual flow features.
Figure 4
Figure 4
Spatial bias block (SBM). It allows the network to learn both local and global information by connecting spatially biased features with texture features in channel cascades.
Figure 5
Figure 5
Vision quality comparison on the MSRS dataset. Areas with large differences are highlighted by RED and GREEN boxes, and enlarged images of RED boxes are in the lower right or left corner.
Figure 6
Figure 6
Vision quality comparison on the TNO dataset. Areas with large differences are highlighted by RED and GREEN boxes, and enlarged images of RED boxes are in the lower right or left corner.
Figure 7
Figure 7
Vision quality comparison on the M3FD dataset. Areas with large differences are highlighted by RED and GREEN boxes, and enlarged images of RED boxes are in the lower right or left corner.
Figure 8
Figure 8
Vision quality comparison of the segmentation results.
Figure 9
Figure 9
Vision quality comparison of the ablation study on important loss functions and modules.
Figure 9
Figure 9
Vision quality comparison of the ablation study on important loss functions and modules.

Similar articles

Cited by

References

    1. Zhang H., Xu H., Tian X., Jiang J., Ma J. Image fusion meets deep learning: A survey and perspective. Inf. Fusion. 2021;76:323–336. doi: 10.1016/j.inffus.2021.06.008. - DOI
    1. Chen J., Li X., Luo L., Ma J. Multi-focus image fusion based on multi-scale gradients and image matting. IEEE Trans. Multimed. 2021;24:655–667. doi: 10.1109/TMM.2021.3057493. - DOI
    1. Parihar A.S., Singh K., Rohilla H., Asnani G. Fusion-based simultaneous estimation of reflectance and illumination for low-light image enhancement. IET Image Process. 2021;15:1410–1423. doi: 10.1049/ipr2.12114. - DOI
    1. Shi Z., Guo B., Zhao M., Zhang C. Nighttime low illumination image enhancement with single image using bright/dark channel prior. EURASIP J. Image Video Process. 2018;2018:13. doi: 10.1186/s13640-018-0251-4. - DOI
    1. Zhang X. Benchmarking and comparing multi-exposure image fusion algorithms. Inf. Fusion. 2021;74:111–131. doi: 10.1016/j.inffus.2021.02.005. - DOI

LinkOut - more resources