A multimodal and temporal foundation model for virtual patient representations at healthcare system scale

2026-04-20Machine Learning

Machine LearningArtificial IntelligenceComputation and Language
AI summary

The authors created Apollo, a large computer model that learns from 30 years of hospital records to understand patient health journeys by combining many different types of medical data, like images and text. Apollo creates a single summary for each patient by integrating over 100,000 medical events and various data types. The authors tested Apollo on many medical prediction tasks, such as forecasting new diseases years ahead and predicting treatment outcomes, and found its results matched real clinical markers. They also showed Apollo can help search medical records using both text and images. This work aims to make complex patient histories easier for computers to interpret and use in healthcare.

multimodal datalongitudinal recordsfoundation modelpatient representationmedical modalitiesclinical vocabularyprognosis tasksfeature attributionsemantic search
Authors
Andrew Zhang, Tong Ding, Sophia J. Wagner, Caiwei Tian, Ming Y. Lu, Rowland Pettit, Joshua E. Lewis, Alexandre Misrahi, Dandan Mo, Long Phi Le, Faisal Mahmood
Abstract
Modern medicine generates vast multimodal data across siloed systems, yet no existing model integrates the full breadth and temporal depth of the clinical record into a unified patient representation. We introduce Apollo, a multimodal temporal foundation model trained and evaluated on over three decades of longitudinal hospital records from a major US hospital system, composed of 25 billion records from 7.2 million patients, representing 28 distinct medical modalities and 12 major medical specialties. Apollo learns a unified representation space integrating over 100 thousand unique medical events in our clinical vocabulary as well as images and clinical text. This "atlas of medical concepts" forms a computational substrate for modeling entire patient care journeys comprised of sequences of structured and unstructured events, which are compressed by Apollo into virtual patient representations. To assess the potential of these whole-patient representations, we created 322 prognosis and retrieval tasks from a held-out test set of 1.4 million patients. We demonstrate the generalized clinical forecasting potential of Apollo embeddings, including predicting new disease onset risk up to five years in advance (95 tasks), disease progression (78 tasks), treatment response (59 tasks), risk of treatment-related adverse events (17 tasks), and hospital operations endpoints (12 tasks). Using feature attribution techniques, we show that model predictions align with clinically-interpretable multimodal biomarkers. We evaluate semantic similarity search on 61 retrieval tasks, and moreover demonstrate the potential of Apollo as a multimodal medical search engine using text and image queries. Together, these modeling capabilities establish the foundation for computable medicine, where the full context of patient care becomes accessible to computational reasoning.