Add Arxiv ID to metadata and improve model card
Browse filesHi! I'm Niels from the Hugging Face community science team.
This PR improves the model card by:
- Adding the `arxiv` ID to the metadata to link it to the [official paper](https://huggingface.co/papers/2602.04279).
- Adding relevant tags such as `ecg`, `multimodal`, and `reasoning` to enhance searchability.
- Linking the repository to the primary dataset used (`ECG-Protocol-Guided-Grounding-CoT`).
Great work on this reasoning MLLM for ECG interpretation!
README.md
CHANGED
|
@@ -1,12 +1,18 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
language:
|
| 4 |
-
- en
|
| 5 |
base_model:
|
| 6 |
- Qwen/Qwen3-VL-8B-Instruct
|
|
|
|
|
|
|
|
|
|
| 7 |
pipeline_tag: image-text-to-text
|
| 8 |
tags:
|
| 9 |
- medical
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
<div align="center">
|
|
@@ -20,6 +26,7 @@ tags:
|
|
| 20 |
<a href='https://arxiv.org/pdf/2602.04279'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
|
| 21 |
<a href='https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-blue'>
|
| 22 |
<a href='https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT'><img src='https://img.shields.io/badge/Dataset-Huggingface-yellow'>
|
|
|
|
| 23 |
|
| 24 |
<p align="center">
|
| 25 |
Jiarui Jin, Haoyu Wang, Xingliang Wu, Xiaocheng Fang, Xiang Lan, Zihan Wang<br/>
|
|
@@ -32,19 +39,15 @@ tags:
|
|
| 32 |
|
| 33 |
## Introduction
|
| 34 |
|
| 35 |
-
Electrocardiography (ECG) serves as an indispensable diagnostic tool in clinical practice, yet existing multimodal large language models (MLLMs) remain unreliable for ECG interpretation, often producing plausible but clinically incorrect analyses. To address this, we propose ECG-R1, the first reasoning MLLM designed for reliable ECG interpretation via three innovations. First, we construct the interpretation corpus using Protocol-Guided Instruction Data Generation, grounding interpretation in measurable ECG features and monograph-defined quantitative thresholds and diagnostic logic. Second, we present a modality-decoupled architecture with Interleaved Modality Dropout to improve robustness and cross-modal consistency when either the ECG signal or ECG image is missing. Third, we present Reinforcement Learning with ECG Diagnostic Evidence Rewards to strengthen evidence-grounded ECG interpretation. Additionally, we systematically evaluate the ECG interpretation capabilities of proprietary, open-source, and medical MLLMs, and provide the first quantitative evidence that severe hallucinations are widespread, suggesting that the public should not directly trust these outputs without independent verification.
|
| 36 |
-
|
| 37 |
-
## Resource
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
#### Paper: 📄 [Arxiv](https://arxiv.org/pdf/2602.04279)
|
| 41 |
-
|
| 42 |
-
### Github: ⌨ [Github](https://github.com/PKUDigitalHealth/ECG-R1)
|
| 43 |
-
|
| 44 |
-
#### Model: 🤗 [ECG-R1-8B](https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL)
|
| 45 |
|
| 46 |
-
##
|
| 47 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
|
| 49 |
## Citation
|
| 50 |
|
|
@@ -63,4 +66,4 @@ If you find ECG-R1 helpful for your research and applications, please cite our p
|
|
| 63 |
```
|
| 64 |
|
| 65 |
## Acknowledgement
|
| 66 |
-
We thank the authors of [PULSE](https://github.com/AIMedLab/PULSE/tree/dev), [ECG-Chat](https://github.com/YubaoZhao/ECG-Chat), [GEM](https://github.com/lanxiang1017/GEM), and [Swift](https://github.com/modelscope/ms-swift) for their publicly released models, datasets, and training codes.
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen3-VL-8B-Instruct
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
license: apache-2.0
|
| 7 |
pipeline_tag: image-text-to-text
|
| 8 |
tags:
|
| 9 |
- medical
|
| 10 |
+
- ecg
|
| 11 |
+
- multimodal
|
| 12 |
+
- reasoning
|
| 13 |
+
arxiv: 2602.04279
|
| 14 |
+
datasets:
|
| 15 |
+
- PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT
|
| 16 |
---
|
| 17 |
|
| 18 |
<div align="center">
|
|
|
|
| 26 |
<a href='https://arxiv.org/pdf/2602.04279'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
|
| 27 |
<a href='https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-blue'>
|
| 28 |
<a href='https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT'><img src='https://img.shields.io/badge/Dataset-Huggingface-yellow'>
|
| 29 |
+
<a href='http://ai.heartvoice.com.cn/ECG-R1/'><img src='https://img.shields.io/badge/Project-Page-green'></a>
|
| 30 |
|
| 31 |
<p align="center">
|
| 32 |
Jiarui Jin, Haoyu Wang, Xingliang Wu, Xiaocheng Fang, Xiang Lan, Zihan Wang<br/>
|
|
|
|
| 39 |
|
| 40 |
## Introduction
|
| 41 |
|
| 42 |
+
Electrocardiography (ECG) serves as an indispensable diagnostic tool in clinical practice, yet existing multimodal large language models (MLLMs) remain unreliable for ECG interpretation, often producing plausible but clinically incorrect analyses. To address this, we propose ECG-R1, the first reasoning MLLM designed for reliable ECG interpretation via three innovations. First, we construct the interpretation corpus using Protocol-Guided Instruction Data Generation, grounding interpretation in measurable ECG features and monograph-defined quantitative thresholds and diagnostic logic. Second, we present a modality-decoupled architecture with Interleaved Modality Dropout to improve robustness and cross-modal consistency when either the ECG signal or ECG image is missing. Third, we present Reinforcement Learning with ECG Diagnostic Evidence Rewards to strengthen evidence-grounded ECG interpretation. Additionally, we systematically evaluate the ECG interpretation capabilities of proprietary, open-source, and medical MLLMs, and provide the first quantitative evidence that severe hallucinations are widespread, suggesting that the public should not directly trust these outputs without independent verification.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 43 |
|
| 44 |
+
## Resources
|
| 45 |
|
| 46 |
+
- **Paper:** 📄 [Arxiv](https://arxiv.org/abs/2602.04279)
|
| 47 |
+
- **Github:** ⌨ [Github](https://github.com/PKUDigitalHealth/ECG-R1)
|
| 48 |
+
- **Online Platform:** 🌐 [ECG-R1 Platform](http://ai.heartvoice.com.cn/ECG-R1/)
|
| 49 |
+
- **Model:** 🤗 [ECG-R1-8B](https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL)
|
| 50 |
+
- **Data:** 🤗 [ECG-Protocol-Guided-Grounding-CoT](https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT)
|
| 51 |
|
| 52 |
## Citation
|
| 53 |
|
|
|
|
| 66 |
```
|
| 67 |
|
| 68 |
## Acknowledgement
|
| 69 |
+
We thank the authors of [PULSE](https://github.com/AIMedLab/PULSE/tree/dev), [ECG-Chat](https://github.com/YubaoZhao/ECG-Chat), [GEM](https://github.com/lanxiang1017/GEM), and [Swift](https://github.com/modelscope/ms-swift) for their publicly released models, datasets, and training codes.
|