| | --- |
| | license: cc-by-nc-nd-4.0 |
| | --- |
| | |
| | # BrainLM model |
| |
|
| | <!-- Provide a quick summary of what the model is/does. --> |
| |
|
| | The pretrained model of Brain Language Model (BrainLM) aims to achieve a general understanding of brain dynamics through self-supervised masked prediction. It is introduced in [this paper](https://www.biorxiv.org/content/10.1101/2023.09.12.557460v1) and its code is available at [this repository](https://github.com/vandijklab/BrainLM) |
| |
|
| | ## Model Details |
| |
|
| | ### Model Description |
| |
|
| | We introduce the Brain Language Model (BrainLM), a foundation model for brain activity dynamics trained on 6,700 hours of fMRI recordings. Utilizing self-supervised masked-prediction training, BrainLM demonstrates proficiency in both fine-tuning and zero-shot inference tasks. Fine-tuning allows for the prediction of clinical variables and future brain states. In zero-shot inference, the model identifies functional networks and generates interpretable latent representations of neural activity. Furthermore, we introduce a novel prompting technique, allowing BrainLM to function as an in silico simulator of brain activity responses to perturbations. BrainLM offers a novel framework for the analysis and understanding of large-scale brain activity data, serving as a “lens” through which new data can be more effectively interpreted. |
| |
|
| | - **Developed by:** [van Dijk Lab](https://www.vandijklab.org/) at Yale University |
| | - **Model type:** ViTMAE |
| | - **License:** [](https://creativecommons.org/licenses/by-nc-nd/4.0/) |
| |
|
| | ### Model Sources [optional] |
| |
|
| | <!-- Provide the basic links for the model. --> |
| |
|
| | - **Repository:** https://github.com/vandijklab/BrainLM |
| | - **Paper:** https://www.biorxiv.org/content/10.1101/2023.09.12.557460v1 |
| | - **Demo [optional]:** [More Information Needed] |
| |
|
| | ## Uses |
| |
|
| | BrainLM is a versatile foundation model for fMRI analysis. It can be used for: |
| |
|
| | - Decoding cognitive variables and mental health biomarkers from brain activity patterns |
| | - Predicting future brain states by learning spatiotemporal fMRI dynamics |
| | - Discovering intrinsic functional networks in the brain without supervision |
| | - Perturbation analysis to simulate the effect of interventions on brain activity |
| |
|
| | ### Out-of-Scope Use |
| |
|
| | Currently, this model has been trained and tested only on fMRI data. There are no guarantees regarding its performance on different modalities of brain recordings. |
| |
|
| | ## Bias, Risks, and Limitations |
| |
|
| | - The model was trained only on healthy adults, so may not generalize to other populations |
| | - The fMRI data has limited spatial-temporal resolution and BOLD signals are an indirect measure of neural activity |
| | - The model has only been evaluated on reconstruction and simple regression/classification tasks so far |
| | - Attention weights provide one method of interpretation but have known limitations |
| |
|
| | ### Recommendations |
| |
|
| | - Downstream applications of the model should undergo careful testing and validation before clinical deployment. |
| | - Like any AI system, model predictions should be carefully reviewed by domain experts before informing decision-making. |
| |
|
| | ## How to Get Started with the Model |
| |
|
| | Use the code below to get started with the model. |
| |
|
| |
|
| | ## Training Details |
| |
|
| | ### Data |
| |
|
| | Data stats: |
| | - UK Biobank (UKB): 76,296 recordings (~6450 hours) |
| | - Human Connectome Project (HCP): 1002 recordings (~250 hours) |
| |
|
| | Preprocessing Steps: |
| | - Motion Correction |
| | - Normalization |
| | - Temporal Filtering |
| | - ICA Denoising |
| |
|
| | Feature Extraction: |
| | - Brain Parcellation: AAL-424 atlas is used to divide the brain into 424 regions. |
| | - Temporal Resolution: ~1 Hz with 0.735s for UKB and 0.72s for HCP. |
| | - Dimensionality: 424-dimensional time series per scan. |
| |
|
| | Data Scaling |
| | - Robust scaling was applied, involving the subtraction of the median and division by the interquartile range across subjects for each parcel. |
| |
|
| | Data split: |
| | - Training data: 80% of the UKB dataset |
| | - Validation data: 10% of the UKB dataset |
| | - Test data: 10% of the UKB dataset and HCP dataset |
| |
|
| | ### Training Procedure |
| |
|
| | BrainLM was pretrained on fMRI recordings from the UK Biobank and HCP datasets. Recordings were parcellated, embedded, masked, and reconstructed via a Transformer autoencoder. The model was evaluated on held-out test partitions of both datasets. |
| |
|
| | Objective: Mean squared error loss between original and predicted parcels |
| |
|
| | Pretraining: |
| | - 100 epochs |
| | - Batch size 512 |
| | - Adam optimizer |
| | - Masking ratios: 20%, 75% and 90% |
| |
|
| | Downstream training: Fine-tuning on future state prediction and regression/classification clinical variables |
| |
|
| |
|
| | #### Metrics |
| |
|
| | In this work, we use the following metrics to evaluate the model's performance: |
| | - Reconstruction error (MSE between predicted and original parcel timeseries) |
| | - Clinical variable regression error (e.g. age, neuroticism scores) |
| | - Functional network classification accuracy |
| |
|
| | [More Information Needed] |
| |
|
| | ### Results |
| |
|
| | [More Information Needed] |
| |
|
| | #### Summary |
| |
|
| |
|
| |
|
| | ## Model Examination [optional] |
| |
|
| | <!-- Relevant interpretability work for the model goes here --> |
| |
|
| | [More Information Needed] |
| |
|
| | ## Technical Specifications [optional] |
| |
|
| | ### Model Architecture and Objective |
| |
|
| | [More Information Needed] |
| |
|
| |
|
| | **BibTeX:** |
| |
|
| | ```bibtex |
| | @article{ortega2023brainlm, |
| | title={BrainLM: A foundation model for brain activity recordings}, |
| | author={Ortega Caro, Josue and Oliveira Fonseca, Antonio Henrique and Averill, Christopher and Rizvi, Syed A and Rosati, Matteo and Cross, James L and Mittal, Prateek and Zappala, Emanuele and Levine, Daniel and Dhodapkar, Rahul M and others}, |
| | journal={bioRxiv}, |
| | pages={2023--09}, |
| | year={2023}, |
| | publisher={Cold Spring Harbor Laboratory} |
| | } |
| | ``` |