Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 May 14;24(10):3130.
doi: 10.3390/s24103130.

Fine-Grained Cross-Modal Semantic Consistency in Natural Conservation Image Data from a Multi-Task Perspective

Affiliations

Fine-Grained Cross-Modal Semantic Consistency in Natural Conservation Image Data from a Multi-Task Perspective

Rui Tao et al. Sensors (Basel). .

Abstract

Fine-grained representation is fundamental to species classification based on deep learning, and in this context, cross-modal contrastive learning is an effective method. The diversity of species coupled with the inherent contextual ambiguity of natural language poses a primary challenge in the cross-modal representation alignment of conservation area image data. Integrating cross-modal retrieval tasks with generation tasks contributes to cross-modal representation alignment based on contextual understanding. However, during the contrastive learning process, apart from learning the differences in the data itself, a pair of encoders inevitably learns the differences caused by encoder fluctuations. The latter leads to convergence shortcuts, resulting in poor representation quality and an inaccurate reflection of the similarity relationships between samples in the original dataset within the shared space of features. To achieve fine-grained cross-modal representation alignment, we first propose a residual attention network to enhance consistency during momentum updates in cross-modal encoders. Building upon this, we propose momentum encoding from a multi-task perspective as a bridge for cross-modal information, effectively improving cross-modal mutual information, representation quality, and optimizing the distribution of feature points within the cross-modal shared semantic space. By acquiring momentum encoding queues for cross-modal semantic understanding through multi-tasking, we align ambiguous natural language representations around the invariant image features of factual information, alleviating contextual ambiguity and enhancing model robustness. Experimental validation shows that our proposed multi-task perspective of cross-modal momentum encoders outperforms similar models on standardized image classification tasks and image-text cross-modal retrieval tasks on public datasets by up to 8% on the leaderboard, demonstrating the effectiveness of the proposed method. Qualitative experiments on our self-built conservation area image-text paired dataset show that our proposed method accurately performs cross-modal retrieval and generation tasks among 8142 species, proving its effectiveness on fine-grained cross-modal image-text conservation area image datasets.

Keywords: cross-modal; cross-modal alignment; cross-modal retrieval; image captioning; multi-task.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

Figures

Figure A1
Figure A1
An example of knowledge distillation from a pre-trained model.
Figure A2
Figure A2
The structural details of the distilled encoder module.
Figure 1
Figure 1
An application instance of the ReCap model.
Figure 2
Figure 2
An overview of training ReCap for cross-modal semantics consistency.
Figure 3
Figure 3
Residual Attention Network Architecture.
Figure 4
Figure 4
Initial visual encoding.
Figure 5
Figure 5
Redundant disambiguation momentum encoding.
Figure 6
Figure 6
Unimodal encoding momentum update.
Figure 7
Figure 7
Examples of text-to-image retrieval on validation dataset.

Similar articles

Cited by

References

    1. Matin M., Shrestha T., Chitale V., Thomas S. Exploring the potential of deep learning for classifying camera trap data of wildlife: A case study from Nepal; Proceedings of the AGU Fall Meeting Abstracts; New Orleans, LA, USA. 13–17 December 2021; p. GC45I-0923.
    1. Norouzzadeh M.S., Nguyen A., Kosmala M., Swanson A., Palmer M.S., Packer C., Clune J. Automatically identifying, counting, and describing wild animals in camera-trap images with deep learning. Proc. Natl. Acad. Sci. USA. 2018;115:E5716–E5725. doi: 10.1073/pnas.1719367115. - DOI - PMC - PubMed
    1. Zett T., Stratford K.J., Weise F. Inter-observer variance and agreement of wildlife information extracted from camera trap images. Biodivers. Conserv. 2022;31:3019–3037. doi: 10.1007/s10531-022-02472-z. - DOI
    1. Swanson A., Kosmala M., Lintott C., Simpson R., Smith A., Packer C. Snapshot Serengeti, high-frequency annotated camera trap images of 40 mammalian species in an African savanna. Sci. Data. 2015;2:1–14. doi: 10.1038/sdata.2015.26. - DOI - PMC - PubMed
    1. McShea W.J., Forrester T., Costello R., He Z., Kays R. Volunteer-run cameras as distributed sensors for macrosystem mammal research. Landsc. Ecol. 2016;31:55–66. doi: 10.1007/s10980-015-0262-9. - DOI

LinkOut - more resources