Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

The rapid expansion of genomics datasets and the application of machine learning has produced sequence-to-activity genomics models with ever-expanding capabilities. However, benchmarking these models on practical applications has been challenging because individual projects evaluate their models in ad hoc ways, and there is substantial heterogeneity of both model architectures and benchmarking tasks. To address this challenge, we have created GAME, a system for large-scale, community-led standardized model benchmarking on user-defined evaluation tasks. We borrow concepts from the Application Programming Interface (API) paradigm to allow for seamless communication between pre-trained models and benchmarking tasks, ensuring consistent evaluation protocols. Because all models and benchmarks are inherently compatible in this framework, the continual addition of new models and new benchmarks is easy. We also developed a Matcher module powered by a large language model (LLM) to automate ambiguous task alignment between benchmarks and models. Containerization of these modules enhances reproducibility and facilitates the deployment of models and benchmarks across computing platforms. By focusing on predicting underlying biochemical phenomena (e.g. gene expression, open chromatin, DNA binding), we ensure that tasks remain technology-independent. We provide examples of benchmarks and models implementing this framework, and anticipate that the community will contribute their own, leading to an ever-expanding and evolving set of models and evaluation tasks. This resource will accelerate genomics research by illuminating the best models for a given task, motivating novel functional genomic benchmarks, and providing a more nuanced understanding of model abilities.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC12265512PMC
http://dx.doi.org/10.1101/2025.07.04.663250DOI Listing

Publication Analysis

Top Keywords

models benchmarks
12
models
11
benchmarking tasks
8
evaluation tasks
8
benchmarks models
8
benchmarks
6
model
5
tasks
5
game genomic
4
genomic api
4

Similar Publications

Engaging residents with the support available at community-based residential mental health rehabilitation facilities is an ongoing challenge for health services. This study explored factors associated with residential rehabilitation engagement across Queensland, Australia through regression modelling of cross-sectional data from a statewide benchmarking activity completed in 2023 (n = 208). The Residential Rehabilitation Engagement Scale (RRES) assessed each resident's rehabilitation engagement.

View Article and Find Full Text PDF

This review article, developed by the EASD Global Council, addresses the growing global challenges in diabetes research and care, highlighting the rising prevalence of diabetes, the increasing complexity of its management and the need for a coordinated international response. With regard to research, disparities in funding and infrastructure between high-income countries and low- and middle-income countries (LMICs) are discussed. The under-representation of LMIC populations in clinical trials, challenges in conducting large-scale research projects, and the ethical and legal complexities of artificial intelligence integration are also considered as specific issues.

View Article and Find Full Text PDF

Improving Door-In-Door-Out Times for STEMI Transfer Patients: Impact of a Protocolized Autolaunch Process.

JACC Case Rep

July 2025

Department of Emergency Medicine, McGovern Medical School at the University of Texas Health Science Center at Houston (UTHealth Houston), Houston, Texas, USA; Texas Emergency Medicine Research Center, Houston, Texas, USA.

Background: The timely transfer of patients with ST-segment elevation myocardial infarction (STEMI) to percutaneous coronary intervention-capable centers is critical for improving outcomes. Although the American Heart Association recommends a door-in-door-out (DIDO) time of ≤30 minutes, national compliance remains low.

Project Rationale: At Harris Health, no patients with STEMI met this benchmark before 2022.

View Article and Find Full Text PDF

Predicting career trajectories is a complex yet impactful task, offering significant benefits for personalized career counseling, recruitment optimization, and workforce planning. However, effective career path prediction (CPP) modeling faces challenges including highly variable career trajectories, free-text resume data, and limited publicly available benchmark datasets. In this study, we present a comprehensive comparative evaluation of CPP models-linear projection, multilayer perceptron (MLP), LSTM, and large language models (LLMs)-across multiple input settings and two recently introduced public datasets.

View Article and Find Full Text PDF

OpenML is an open-source platform that democratizes machine-learning evaluation by enabling anyone to share datasets in uniform standards, define precise machine-learning tasks, and automatically share detailed workflows and model evaluations. More than just a platform, OpenML fosters a collaborative ecosystem where scientists create new tools, launch initiatives, and establish standards to advance machine learning. Over the past decade, OpenML has inspired over 1,500 publications across diverse fields, from scientists releasing new datasets and benchmarking new models to educators teaching reproducible science.

View Article and Find Full Text PDF