Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
[Preprint]. 2025 Jul 8:2025.07.04.663250.
doi: 10.1101/2025.07.04.663250.

GAME: Genomic API for Model Evaluation

Affiliations

GAME: Genomic API for Model Evaluation

Ishika Luthra et al. bioRxiv. .

Abstract

The rapid expansion of genomics datasets and the application of machine learning has produced sequence-to-activity genomics models with ever-expanding capabilities. However, benchmarking these models on practical applications has been challenging because individual projects evaluate their models in ad hoc ways, and there is substantial heterogeneity of both model architectures and benchmarking tasks. To address this challenge, we have created GAME, a system for large-scale, community-led standardized model benchmarking on user-defined evaluation tasks. We borrow concepts from the Application Programming Interface (API) paradigm to allow for seamless communication between pre-trained models and benchmarking tasks, ensuring consistent evaluation protocols. Because all models and benchmarks are inherently compatible in this framework, the continual addition of new models and new benchmarks is easy. We also developed a Matcher module powered by a large language model (LLM) to automate ambiguous task alignment between benchmarks and models. Containerization of these modules enhances reproducibility and facilitates the deployment of models and benchmarks across computing platforms. By focusing on predicting underlying biochemical phenomena (e.g. gene expression, open chromatin, DNA binding), we ensure that tasks remain technology-independent. We provide examples of benchmarks and models implementing this framework, and anticipate that the community will contribute their own, leading to an ever-expanding and evolving set of models and evaluation tasks. This resource will accelerate genomics research by illuminating the best models for a given task, motivating novel functional genomic benchmarks, and providing a more nuanced understanding of model abilities.

PubMed Disclaimer

Conflict of interest statement

Competing Interests V.A. is an employee of Sanofi. A.L. is an employee of Genentech, inc. The remaining authors declare no competing interests.

Figures

Figure 1:
Figure 1:. GAME framework.
GAME includes three modules: The Evaluator, containing a benchmark dataset; the Predictor, encompassing a sequence-to-activity model; and the Matcher, capturing relationships between tasks. All GAME modules are inherently interoperable by communicating in the GAME API protocol over TCP. For each benchmark, the Evaluator requests a prediction from the Predictor, which consults the Matcher to determine the closest task the Predictor can complete. Once the Matcher returns the best match, the Predictor will complete its prediction and return it to the Evaluator, which will evaluate performance. Members of the genomics community will contribute modules to enable continual evaluation of more models across more benchmarks.
Figure 2:
Figure 2:. Sample benchmarking done with GAME.
a, Expression evaluation tasks. Models (x axis) were evaluated for their correlation to measured expression levels (colours) across a variety of tasks (y axis). b, Chromatin conformation tasks. Correlation of Orca predictions vs measured chromatin contact frequencies (colours) for two Orca test-set chromosomes and one validation-set chromosome (y axis). c, Consistency evaluation for accessibility. Point and track based accessibility consistency evaluators (y axis) were used to evaluate the correlation between predictions for forward and reverse complement sequences (colours) in models of DNA accessibility (x axis). *Correlation values from down sampled datasets. Grey cells mark values that were not calculated due models that could not complete the Evaluator’s requested task.

References

    1. Guerrini M. M., Oguchi A., Suzuki A. & Murakawa Y. Cap analysis of gene expression (CAGE) and noncoding regulatory elements. Semin Immunopathol 44, 127–136 (2022). - PubMed
    1. Tang Z., Toneyan S. & Koo P. K. Current approaches to genomic deep learning struggle to fully capture human genetic variation. Nat Genet 55, 2021–2022 (2023). - PubMed
    1. Sasse A. et al. Benchmarking of deep neural networks for predicting personal gene expression from DNA sequence highlights shortcomings. Nat Genet 55, 2060–2064 (2023). - PubMed
    1. Huang C. et al. Personal transcriptome variation is poorly explained by current genomic deep learning models. Nat Genet 55, 2056–2059 (2023). - PMC - PubMed
    1. robson eyes s. & Ioannidis N. M. GUANinE v1.0: Benchmark Datasets for Genomic AI Sequence-to-Function Models. bioRxiv 2023.10.12.562113 (2024) doi: 10.1101/2023.10.12.562113. - DOI

Publication types

LinkOut - more resources