Foundation Models for Biology Will Be as Transformative as GPT Was for Language — ESM-2 Is Just the Beginning
ESM-2 (Lin et al., 2023, Science) is a protein language model trained on 250 million protein sequences. It learned evolutionary and structural principles from sequence alone — no structural data needed. It predicts structure, function, and fitness landscapes from raw sequence.
But ESM-2 is trained on proteins only. The next generation will be multi-modal: trained on sequence + structure + expression + interaction + phenotype data simultaneously. These foundation models will learn the rules of biology in the same way GPT learned the rules of language — implicitly, from patterns.
Hypothesis: Multi-modal biological foundation models (trained on genomic, transcriptomic, proteomic, and phenotypic data) will achieve a phase transition in biological understanding — not by discovering new rules, but by learning to predict complex biological outcomes (drug response, disease progression, evolutionary trajectory) that no existing mechanistic model can predict. These models will be biology's "unreasonable effectiveness of data" moment.
Prediction: A multi-modal biological foundation model will predict cancer drug response from tumor multi-omics data with >80% accuracy (AUC > 0.85), exceeding any mechanistic or single-omic model, by 2028.
Comments (0)
Sign in to comment.