Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2025:2941:313-355.
doi: 10.1007/978-1-0716-4623-6_19.

Large Language Model (LLM)-Based Advances in Prediction of Post-translational Modification Sites in Proteins

Affiliations
Review

Large Language Model (LLM)-Based Advances in Prediction of Post-translational Modification Sites in Proteins

Pawel Pratyush et al. Methods Mol Biol. 2025.

Abstract

Post-translational modifications (PTMs) are vital regulators of protein function, influencing a myriad of cellular processes and disease mechanisms. Traditional experimental methods for PTM identification are both costly and labor-intensive, underlining the pressing need for efficient computational approaches. Early computational strategies predominantly relied on primary amino acid sequences and handcrafted features, which often lacked the contextual and structural understanding necessary for precise PTM site prediction. The emergence of transformer-based large language models (LLMs), particularly protein language models (pLMs), has revolutionized PTM prediction by producing context-aware embeddings that capture functional and structural intra-sequence dependencies. In this chapter, we provide a comprehensive review of recent advancements in leveraging LLMs (or, pLMs) for PTM site prediction, an important residue-level task in protein research. We identify emerging trends in the field, including the application of fine-tuning techniques, the integration of embeddings from multiple pLMs, and the incorporation of multiple modalities such as codon-aware embeddings, 3D structural data, and conventional representations. Additionally, we discuss tools that employ graph-based representations, the mamba architecture, and contrastive learning paradigms to further refine pLM-powered PTM site prediction models. We finally explore the interpretability and explainability aspects of the embeddings used in various tools. Despite the significant progress made, persistent limitations remain, and we outline these challenges while proposing directions for future research.

Keywords: AlphaFold; Contrastive learning; Explainability; Fine-tuning; GPT; Graph; Large language model; Mamba; Post-translational modification; Protein language model.

PubMed Disclaimer

Similar articles

References

    1. Keenan EK, Zachman DK, Hirschey MD (2021) Discovering the landscape of protein modifications. Mol Cell 81:1868–1878. https://doi.org/10.1016/j.molcel.2021.03.015 - DOI - PubMed - PMC
    1. Aebersold R, Agar JN, Amster IJ et al (2018) How many human proteoforms are there? Nat Chem Biol 14:206–214. https://doi.org/10.1038/nchembio.2576 - DOI - PubMed - PMC
    1. Hong X, Li N, Lv J et al (2023) PTMint database of experimentally verified PTM regulation on protein–protein interaction. Bioinformatics 39:btac823. https://doi.org/10.1093/bioinformatics/btac823 - DOI - PubMed
    1. Lee JM, Hammarén HM, Savitski MM, Baek SH (2023) Control of protein stability by post-translational modifications. Nat Commun 14:201. https://doi.org/10.1038/s41467-023-35795-8 - DOI - PubMed - PMC
    1. Ryšlavá H, Doubnerová V, Kavan D, Vaněk O (2013) Effect of posttranslational modifications on enzyme function and assembly. J Proteome 92:80–109. https://doi.org/10.1016/j.jprot.2013.03.025 - DOI

LinkOut - more resources