Generative Artificial Intelligence for Biology:
Toward Unifying Models, Algorithms, and Modalities

Xiner Li1,7,*,†, Xingyu Su1,*, Yuchao Lin1,11, Chenyu Wang2, Yijia Xiao3, Tianyu Liu4, Chi Han5, Michael Sun2, Montgomery Bohde1, Anna Hart5, Wendi Yu1, Masatoshi Uehara6, Gabriele Scalia7, Xiao Luo8, Carl Edwards7, Wengong Jin9,10, Jianwen Xie11, Ehsan Hajiramezanali7, Edward De Brouwer7, Qing Sun12, Byung-Jun Yoon13,16, Xiaoning Qian1,13,16, Marinka Zitnik14, Heng Ji5, Hongyu Zhao4, Wei Wang3, Shuiwang Ji1,15,17,†
1Department of Computer Science and Engineering, Texas A&M University, 2MIT CSAIL, 3Department of Computer Science, UCLA, 4Interdepartmental Program of Computational Biology and Bioinformatics, Yale University, 5Siebel School of Computing and Data Science, UIUC, 6Chan Zuckerberg Initiative, 7Genentech, Inc., 8Department of Statistics, University of Wisconsin–Madison, 9Broad Institute of MIT and Harvard, 10Khoury College of Computer Sciences, Northeastern University, 11Lambda, Inc., 12Artie McFerrin Department of Chemical Engineering, Texas A&M University, 13Department of Electrical and Computer Engineering, Texas A&M University, 14Department of Biomedical Informatics, Harvard Medical School, 15Department of Materials Science and Engineering, Texas A&M University, 16Computing and Data Sciences, Brookhaven National Laboratory, 17J. Mike Walker '66 Department of Mechanical Engineering, Texas A&M University
*Equal Contribution, Correspondence to: Xiner Li <lxe@tamu.edu>, Shuiwang Ji <sji@tamu.edu>
Layered taxonomy of generative AI for biomolecule

Figure 1. A layered taxonomy of generative AI for biology. Our perspective organizes the field from biological problem spaces and core architectural building blocks, through major generative paradigms such as autoregressive, masked/bidirectional, and diffusion/flow matching, to unified cross-paradigm hybrids, multimodal biological foundation models, and tool-augmented agentic systems for closed-loop discovery.

Abstract

Rapid advances in generative artificial intelligence have revolutionized biological modeling across domains such as protein, genetics, and single-cell. However, existing works often organize applications by molecule types or specific research tasks, overlooking the methodological convergence and cross-modal innovations. This paper aims to present a unified methodological perspective that highlights the fundamental technical commonalities across biological modalities. We systematically organize recent advances in generative modeling for biology through the lens of core machine learning paradigms, from language models (LMs) and diffusion models to their emerging hybrid architectures. Our work reveals how techniques initially developed for one molecular type (e.g., protein design) can be effectively transferred to others (e.g., RNA engineering), and identifies the convergence trend where discrete diffusion models and iterative language models represent different facets of a unified generative framework. We cover the evolution from domain-specific models to multi-modal biological foundation models and agent-based systems. By emphasizing methodological connections rather than applications, this paper aims to accelerate cross-domain innovation and make the field more accessible to the broader machine learning community. We conclude by identifying promising research directions where successful techniques in one biological domain remain unexplored in others, offering a roadmap for future advances in generative biology.

Survey Overview

Language Modeling for Biological Generation

Language-modeling pipelines for biomolecular generation

Figure 2. An overview of language-modeling pipelines for biological generation, illustrating how diverse biological inputs are tokenized, processed by a Transformer backbone, and trained under either masked bidirectional-context modeling or autoregressive causal modeling objectives.

Diffusion & Flow Matching Models

Forward-reverse diffusion pipeline for biomolecular generation

Figure 3. Overview of the forward-reverse diffusion pipeline for biological generation, showing the forward noising process and the learned reverse denoising process, in both discrete (token-level) and continuous (geometry-level) formulations.

Toward Unified Models

Unified generative-model view

Figure 4. A unified generative-model view integrating classic architectures, language models, and diffusion models. This illustrates the perspective that iterative language modeling can be viewed as tokenized diffusion, where training and inference correspond to iterative denoising, supported by shared backbones and modality-specific (discrete/continuous) output heads.

Multi-Modal Foundation Models

Multi-modal foundation model for biomolecules

Figure 5. A multi-modal foundation model for biomolecules, where heterogeneous inputs are embedded and fused into a backbone to support diverse downstream design tasks such as binder/vaccine design, motif scaffolding, structure generation, and inverse folding.

Biological Agent Systems

Biological agent paradigm

Figure 6. The biological agent paradigm that couples hypothesis generation with tool use/creation and laboratory automation, enabling iterative refinement from computational reasoning to experimental validation.

BibTeX

@article{
GenAI4bio:Li,
author = {Xiner Li and Xingyu Su and Yuchao Lin and Chenyu Wang and Yijia Xiao and Tianyu Liu and Chi Han and Michael Sun and Montgomery Bohde and Anna Hart and Wendi Yu and Masatoshi Uehara and Gabriele Scalia and Xiao Luo and Carl Edwards and Wengong Jin and Jianwen Xie and Ehsan Hajiramezanali and Edward De Brouwer and Qing Sun and Byung-Jun Yoon and Xiaoning Qian and Marinka Zitnik and Heng Ji and Hongyu Zhao and Wei Wang and Shuiwang Ji},
title = {Generative Artificial Intelligence for Biology: Toward Unifying Models, Algorithms, and Modalities},
journal = {ChemRxiv},
volume = {2026},
number = {0212},
pages = {},
year = {2026},
}