Enhancing Structure-Aware Protein Language Models with Efficient Fine-Tuning for Various Protein Prediction Tasks
- PMID: 40601249
- DOI: 10.1007/978-1-0716-4623-6_2
Enhancing Structure-Aware Protein Language Models with Efficient Fine-Tuning for Various Protein Prediction Tasks
Abstract
Proteins are crucial in a wide range of biological and engineering processes. Large protein language models (PLMs) can significantly advance our understanding and engineering of proteins. However, the effectiveness of PLMs in prediction and design is largely based on the representations derived from protein sequences. Without incorporating the three-dimensional (3D) structures of proteins, PLMs would overlook crucial aspects of how proteins interact with other molecules, thereby limiting their predictive accuracy. To address this issue, we present S-PLM, a 3D structure-aware PLM, that employs multi-view contrastive learning to align protein sequences with their 3D structures in a unified latent space. Previously, we utilized a contact map-based approach to encode structural information, applying the Swin-Transformer to contact maps derived from AlphaFold-predicted protein structures. This work introduces a new approach that leverages a geometric vector perceptron (GVP) model to process 3D coordinates and obtain structural embeddings. We focus on the application of structure-aware models for protein-related tasks by utilizing efficient fine-tuning methods to achieve optimal performance without significant computational costs. Our results show that S-PLM outperforms sequence-only PLMs across all protein clustering and classification tasks, achieving performance on par with state-of-the-art methods that require both sequence and structure inputs. S-PLM and its tuning tools are available at https://github.com/duolinwang/S-PLM/ .
Keywords: 3D structure-aware PLM; ESM 2; Efficient fine-tuning; Large protein language models (PLMs); Multi-view contrastive learning.
© 2025. The Author(s), under exclusive license to Springer Science+Business Media, LLC, part of Springer Nature.
Similar articles
-
Boost Protein Language Model with Injected Structure Information Through Parameter Efficient Fine-tuning.Comput Biol Med. 2025 Sep;195:110607. doi: 10.1016/j.compbiomed.2025.110607. Epub 2025 Jun 30. Comput Biol Med. 2025. PMID: 40592174
-
S-PLM: Structure-Aware Protein Language Model via Contrastive Learning Between Sequence and Structure.Adv Sci (Weinh). 2025 Feb;12(5):e2404212. doi: 10.1002/advs.202404212. Epub 2024 Dec 12. Adv Sci (Weinh). 2025. PMID: 39665266 Free PMC article.
-
Large Language Model (LLM)-Based Advances in Prediction of Post-translational Modification Sites in Proteins.Methods Mol Biol. 2025;2941:313-355. doi: 10.1007/978-1-0716-4623-6_19. Methods Mol Biol. 2025. PMID: 40601266 Review.
-
Advancing the accuracy of clathrin protein prediction through multi-source protein language models.Sci Rep. 2025 Jul 8;15(1):24403. doi: 10.1038/s41598-025-08510-4. Sci Rep. 2025. PMID: 40628826 Free PMC article.
-
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3. Cochrane Database Syst Rev. 2022. PMID: 35593186 Free PMC article.
References
-
- Ofer D, Brandes N, Linial M (2021) The language of proteins: NLP, machine learning & protein sequences. Comput Struct Biotechnol J 19:1750–1758. https://doi.org/10.1016/j.csbj.2021.03.022 - DOI - PubMed - PMC
-
- Devlin J, Chang M-W, Lee K, Toutanova K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805. Accessed: Sep. 27, 2024. [Online]. Available: http://arxiv.org/abs/1810.04805
-
- Elnaggar A et al (2022) ProtTrans: toward understanding the language of life through self-supervised learning. IEEE Trans Pattern Anal Mach Intell 44(10):7112–7127. https://doi.org/10.1109/TPAMI.2021.3095381 - DOI - PubMed
-
- Lin Z et al (2023) Evolutionary-scale prediction of atomic-level protein structure with a language model. Science 379(6637):1123–1130. https://doi.org/10.1126/science.ade2574 - DOI - PubMed
-
- Rives A et al (2021) Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. Proc Natl Acad Sci USA 118(15):e2016239118. https://doi.org/10.1073/pnas.2016239118 - DOI - PubMed - PMC
MeSH terms
Substances
LinkOut - more resources
Full Text Sources