Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Jan 2;40(1):btae004.
doi: 10.1093/bioinformatics/btae004.

CircSI-SSL: circRNA-binding site identification based on self-supervised learning

Affiliations

CircSI-SSL: circRNA-binding site identification based on self-supervised learning

Chao Cao et al. Bioinformatics. .

Abstract

Motivation: In recent years, circular RNAs (circRNAs), the particular form of RNA with a closed-loop structure, have attracted widespread attention due to their physiological significance (they can directly bind proteins), leading to the development of numerous protein site identification algorithms. Unfortunately, these studies are supervised and require the vast majority of labeled samples in training to produce superior performance. But the acquisition of sample labels requires a large number of biological experiments and is difficult to obtain.

Results: To resolve this matter that a great deal of tags need to be trained in the circRNA-binding site prediction task, a self-supervised learning binding site identification algorithm named CircSI-SSL is proposed in this article. According to the survey, this is unprecedented in the research field. Specifically, CircSI-SSL initially combines multiple feature coding schemes and employs RNA_Transformer for cross-view sequence prediction (self-supervised task) to learn mutual information from the multi-view data, and then fine-tuning with only a few sample labels. Comprehensive experiments on six widely used circRNA datasets indicate that our CircSI-SSL algorithm achieves excellent performance in comparison to previous algorithms, even in the extreme case where the ratio of training data to test data is 1:9. In addition, the transplantation experiment of six linRNA datasets without network modification and hyperparameter adjustment shows that CircSI-SSL has good scalability. In summary, the prediction algorithm based on self-supervised learning proposed in this article is expected to replace previous supervised algorithms and has more extensive application value.

Availability and implementation: The source code and data are available at https://github.com/cc646201081/CircSI-SSL.

PubMed Disclaimer

Conflict of interest statement

None declared.

Figures

Figure 1.
Figure 1.
CircSI-SSL framework.
Figure 2.
Figure 2.
RNA_Transformer structure.
Figure 3.
Figure 3.
AUC discrimination performance obtained by eight existing supervised algorithms.
Figure 4.
Figure 4.
AUC performance obtained by the latest three supervised learning algorithms on six datasets.
Figure 5.
Figure 5.
Performance comparison between CircSI-SSL and the latest three supervised algorithms in four indicators.
Figure 6.
Figure 6.
Average AUC performance comparison between CircSI-SSL and the latest three supervised algorithms on six datasets.
Figure 7.
Figure 7.
Average AUC performance with and without SSL across six datasets.
Figure 8.
Figure 8.
Comparison of transplant performance on linRNA datasets.

Similar articles

Cited by

References

    1. Alipanahi B, Delong A, Weirauch MT. et al. Predicting the sequence specificities of DNA-and RNA-binding proteins by deep learning. Nat Biotechnol 2015;33:831–8. - PubMed
    1. Bogard B, Francastel C, Hubé F.. A new method for the identification of thousands of circular RNAs. Non-Coding RNA Investig 2018;2:5.
    1. Cao C, Yang S, Li M. et al. CircSSNN: circRNA-binding site prediction via sequence self-attention neural networks with pre-normalization. BMC Bioinformatics 2023;24:220. - PMC - PubMed
    1. Chen L-L. The biogenesis and emerging roles of circular RNAs. Nat Rev Mol Cell Biol 2016;17:205–11. - PubMed
    1. Chen T, Kornblith S, Norouzi M. et al. A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning. Vienna, Austria, 13 July, 2020. 1597–1607. PMLR, 2020.

Publication types