Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2023 Jul 22;12(7):1033.
doi: 10.3390/biology12071033.

Transformer Architecture and Attention Mechanisms in Genome Data Analysis: A Comprehensive Review

Affiliations
Review

Transformer Architecture and Attention Mechanisms in Genome Data Analysis: A Comprehensive Review

Sanghyuk Roy Choi et al. Biology (Basel). .

Abstract

The emergence and rapid development of deep learning, specifically transformer-based architectures and attention mechanisms, have had transformative implications across several domains, including bioinformatics and genome data analysis. The analogous nature of genome sequences to language texts has enabled the application of techniques that have exhibited success in fields ranging from natural language processing to genomic data. This review provides a comprehensive analysis of the most recent advancements in the application of transformer architectures and attention mechanisms to genome and transcriptome data. The focus of this review is on the critical evaluation of these techniques, discussing their advantages and limitations in the context of genome data analysis. With the swift pace of development in deep learning methodologies, it becomes vital to continually assess and reflect on the current standing and future direction of the research. Therefore, this review aims to serve as a timely resource for both seasoned researchers and newcomers, offering a panoramic view of the recent advancements and elucidating the state-of-the-art applications in the field. Furthermore, this review paper serves to highlight potential areas of future investigation by critically evaluating studies from 2019 to 2023, thereby acting as a stepping-stone for further research endeavors.

Keywords: attention mechanism; bioinformatics; deep learning; genome data; genomics; natural language processing; sequence analysis; transcriptome data; transformer model.

PubMed Disclaimer

Conflict of interest statement

The authors claim no conflict of interest.

Figures

Figure 1
Figure 1
Illustration of the transformer architecture and the attention mechanism. (A) Transformer structure; (B) Attention mechanism.
Figure 2
Figure 2
Distribution Patterns of Publication Years and Citation Frequencies (A) Distribution of Publication Years. (B) Distribution of Citation Frequencies. (C) Relationship between Citations and Publication Year.

Similar articles

Cited by

References

    1. Auslander N., Gussow A.B., Koonin E.V. Incorporating Machine Learning into Established Bioinformatics Frameworks. Int. J. Mol. Sci. 2021;22:2903. doi: 10.3390/ijms22062903. - DOI - PMC - PubMed
    1. Lee M. Deep Learning Techniques with Genomic Data in Cancer Prognosis: A Comprehensive Review of the 2021–2023 Literature. Biology. 2023;12:893. doi: 10.3390/biology12070893. - DOI - PMC - PubMed
    1. Gomes R., Paul N., He N., Huber A.F., Jansen R.J. Application of Feature Selection and Deep Learning for Cancer Prediction Using DNA Methylation Markers. Genes. 2022;13:1557. doi: 10.3390/genes13091557. - DOI - PMC - PubMed
    1. Sadad T., Aurangzeb R.A., Safran M., Imran, Alfarhood S., Kim J. Classification of Highly Divergent Viruses from DNA/RNA Sequence Using Transformer-Based Models. Biomedicines. 2023;11:1323. doi: 10.3390/biomedicines11051323. - DOI - PMC - PubMed
    1. Lee M. Recent Advances in Deep Learning for Protein-Protein Interaction Analysis: A Comprehensive Review. Molecules. 2023;28:5169. doi: 10.3390/molecules28135169. - DOI - PMC - PubMed

LinkOut - more resources