One AI Model That Does Everything: BrainIAC Predicts Brain Age, Dementia Risk, and Cancer Survival from Unlabeled MRIs

Harvard researchers built a foundation model that extracts health signals from routine brain scans without requiring labeled training data, outperforming task-specific AI on seven clinical applications.

Medical AI has a data problem. Building a tool that detects brain tumors requires thousands of labeled brain tumor scans. Building one that predicts dementia requires a separate dataset of labeled dementia patients. Building one that estimates brain age requires yet another curated collection with verified patient ages. Each application demands its own expensive, time-consuming annotation effort - which is why most AI tools do exactly one thing.

Researchers at Mass General Brigham just built a system that sidesteps this bottleneck entirely. Their model, called BrainIAC (Brain Imaging Adaptive Core), learns from unlabeled brain MRIs and then adapts to multiple clinical tasks without needing extensive labeled data for each one. The research, published this month in Nature Neuroscience, suggests a different path forward for medical AI: foundation models that learn general representations first and specialize later.

One Model, Seven Tasks

BrainIAC isn’t specialized. It’s a general-purpose brain MRI analyzer that handles a range of clinical applications with the same underlying architecture. The team at Mass General Brigham, led by Dr. Benjamin Kann from the Artificial Intelligence in Medicine Program, validated the model across seven distinct tasks:

  • Brain age estimation – predicting a patient’s biological brain age from their scan
  • Dementia risk prediction – identifying patients likely to develop cognitive decline
  • Brain tumor mutation detection – determining genetic alterations from imaging alone
  • Brain cancer survival prediction – forecasting outcomes for cancer patients
  • MRI scan type classification – distinguishing between different imaging protocols
  • Abnormal vs. normal classification – flagging scans that need attention

Across all seven tasks, BrainIAC outperformed three conventional, task-specific AI frameworks. The performance gap was particularly striking when training data was scarce or task complexity was high - exactly the conditions that typically break specialized models.

The Self-Supervised Trick

What makes BrainIAC different is how it learns. Traditional medical AI requires humans to label every scan: this one shows a tumor, this one shows atrophy, this one is from a 67-year-old. That labeling process is expensive, slow, and creates bottlenecks that limit what AI can do in medicine.

BrainIAC uses self-supervised learning, a technique where the model learns to extract features from unlabeled data by predicting relationships within the images themselves. The team trained it on nearly 49,000 brain MRI scans without telling it anything about what those scans showed. The model learned to identify patterns, structures, and variations on its own.

Once pre-trained on this massive unlabeled dataset, BrainIAC can be “fine-tuned” for specific tasks with far less labeled data than a model trained from scratch would require. This is the foundation model paradigm that has transformed language AI - models like GPT-4 learn general language understanding first, then adapt to specific applications. BrainIAC applies the same principle to medical imaging.

When Data Is Scarce

The most significant finding isn’t that BrainIAC works. It’s when it works best.

Medical AI researchers consistently hit the same wall: rare conditions don’t have enough labeled examples to train reliable models. Brain tumor mutations are uncommon. Specific dementia subtypes are uncommon. The datasets that exist are small, and small datasets produce brittle AI.

BrainIAC’s self-supervised pre-training changes that equation. Because it learned general brain anatomy and structure from nearly 49,000 unlabeled scans, it doesn’t need as many labeled examples to learn a specific task. The model already understands what brains look like; it just needs a few examples to learn what distinguishes, say, an aggressive tumor mutation from a less dangerous one.

This matters for conditions where large labeled datasets will never exist. If a foundation model can learn from hundreds of examples what a specialized model needs thousands to learn, it opens up medical AI to the long tail of rare diseases.

Generalization Across Pathology

The researchers were particularly interested in whether BrainIAC could transfer knowledge between healthy and diseased brains. Many AI models trained primarily on healthy subjects struggle when confronted with pathology - the tumors, lesions, and structural changes that are often the whole point of clinical imaging.

BrainIAC successfully generalized across healthy and abnormal images, applying what it learned from one domain to tasks in the other. It could handle straightforward tasks like classifying scan types and challenging ones like detecting tumor mutations with the same underlying model.

This cross-domain transfer is crucial for clinical deployment. A system that only works on clean, textbook-quality scans isn’t useful in hospitals where scans come with artifacts, patient movement, and pathology that wasn’t anticipated when the imaging protocol was chosen.

The Collaboration Question

BrainIAC’s development involved a large team spanning multiple institutions. Beyond the Mass General Brigham core group, researchers from institutions across Germany, the UK, and the US contributed to the work. Data from the Children’s Brain Tumor Network helped expand the model’s exposure to pediatric cases.

This collaborative model reflects an emerging reality in medical AI: foundation models require more data than any single institution can provide. The most robust systems will be those trained on diverse datasets from multiple hospitals, imaging centers, and patient populations. BrainIAC’s validation across nearly 49,000 scans is substantial, but it still represents a limited slice of human brain variation.

What It Doesn’t Do Yet

Kann and his team are clear about the limitations. BrainIAC has been validated on brain MRIs specifically, and “further research is needed to test this framework on additional brain imaging methods and larger datasets.”

That’s not a minor caveat. MRI is one imaging modality among many. CT scans, PET scans, and diffusion tensor imaging all capture different aspects of brain structure and function. Whether BrainIAC’s self-supervised learning approach transfers to these other modalities - or whether each would require its own foundation model - remains an open question.

The study also doesn’t include prospective clinical trials. BrainIAC was evaluated retrospectively, meaning researchers tested it on existing scans rather than using it to influence patient care in real time. Before any foundation model enters clinical practice, it needs to be tested in deployment conditions where its outputs actually affect treatment decisions.

The Bigger Picture

BrainIAC joins a growing class of medical foundation models that are fundamentally changing how AI is applied to clinical data. Rather than building narrow tools for narrow tasks, researchers are increasingly pre-training general-purpose models on massive unlabeled datasets and then adapting them to specific applications.

The approach has clear advantages: better performance with less labeled data, stronger generalization to new tasks, and a path toward handling rare conditions that lack sufficient training examples. It also raises questions about accountability. When a model makes a prediction based on patterns it learned unsupervised from 49,000 scans, explaining why it reached that conclusion becomes harder.

For now, Kann frames BrainIAC as a tool for “accelerating biomarker discovery, enhancing diagnostic tools, and speeding the adoption of AI in clinical practice.” The foundation model approach doesn’t replace the need for clinical validation - it accelerates how quickly new applications can be developed and tested.

The code and pre-trained model weights have been released on GitHub, which means other research groups can build on BrainIAC rather than starting from scratch. If the foundation model paradigm takes hold in medical imaging the way it has in language AI, we may be seeing the beginning of a shift from hundreds of specialized tools to a smaller number of general-purpose systems that adapt to whatever task is needed.

Whether that shift improves patient care depends entirely on how these models are validated, deployed, and monitored. The technical capability is now demonstrably real. The clinical integration remains the hard part.