|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- vit |
|
|
- ecg |
|
|
- self-supervised |
|
|
- contrastive-learning |
|
|
--- |
|
|
|
|
|
# ViT-MAE ECG Encoder |
|
|
|
|
|
This repository contains a Vision Transformer (ViT-MAE) model pretrained on ECG signals using masked autoencoding. It is a forked and modified version of the Hugging Face `transformers` library. |
|
|
|
|
|
## Model Files |
|
|
|
|
|
- `model.safetensors`: Model weights (~77MB) saved in `safetensors` format. |
|
|
- `config.json`: Model architecture and configuration. |
|
|
|
|
|
## Usage |
|
|
|
|
|
To use this model, make sure to install the forked version of `transformers` (see below), which includes modifications for original implementations of ViTMAE model from HuggingFace. |
|
|
|
|
|
```bash |
|
|
git clone git@github.com:Alsalivan/ecgcmr.git |
|
|
cd ecgcmr/external/transformers |
|
|
pip install -e . |
|
|
``` |
|
|
|
|
|
## Load model |
|
|
|
|
|
``` |
|
|
from transformers import ViTMAEModel |
|
|
model = ViTMAEModel.from_pretrained("alsalivan/vitmae_ecg") |
|
|
``` |