Add Arxiv ID to metadata and improve model card
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,12 +1,18 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
language:
|
| 4 |
-
- en
|
| 5 |
base_model:
|
| 6 |
- Qwen/Qwen3-VL-8B-Instruct
|
|
|
|
|
|
|
|
|
|
| 7 |
pipeline_tag: image-text-to-text
|
| 8 |
tags:
|
| 9 |
- medical
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
| 12 |
<div align="center">
|
|
@@ -20,6 +26,7 @@ tags:
|
|
| 20 |
<a href='https://arxiv.org/pdf/2602.04279'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
|
| 21 |
<a href='https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-blue'>
|
| 22 |
<a href='https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT'><img src='https://img.shields.io/badge/Dataset-Huggingface-yellow'>
|
|
|
|
| 23 |
|
| 24 |
<p align="center">
|
| 25 |
Jiarui Jin, Haoyu Wang, Xingliang Wu, Xiaocheng Fang, Xiang Lan, Zihan Wang<br/>
|
|
@@ -32,19 +39,15 @@ tags:
|
|
| 32 |
|
| 33 |
## Introduction
|
| 34 |
|
| 35 |
-
Electrocardiography (ECG) serves as an indispensable diagnostic tool in clinical practice, yet existing multimodal large language models (MLLMs) remain unreliable for ECG interpretation, often producing plausible but clinically incorrect analyses. To address this, we propose ECG-R1, the first reasoning MLLM designed for reliable ECG interpretation via three innovations. First, we construct the interpretation corpus using Protocol-Guided Instruction Data Generation, grounding interpretation in measurable ECG features and monograph-defined quantitative thresholds and diagnostic logic. Second, we present a modality-decoupled architecture with Interleaved Modality Dropout to improve robustness and cross-modal consistency when either the ECG signal or ECG image is missing. Third, we present Reinforcement Learning with ECG Diagnostic Evidence Rewards to strengthen evidence-grounded ECG interpretation. Additionally, we systematically evaluate the ECG interpretation capabilities of proprietary, open-source, and medical MLLMs, and provide the first quantitative evidence that severe hallucinations are widespread, suggesting that the public should not directly trust these outputs without independent verification.
|
| 36 |
-
|
| 37 |
-
## Resource
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
#### Paper: π [Arxiv](https://arxiv.org/pdf/2602.04279)
|
| 41 |
-
|
| 42 |
-
### Github: β¨ [Github](https://github.com/PKUDigitalHealth/ECG-R1)
|
| 43 |
-
|
| 44 |
-
#### Model: π€ [ECG-R1-8B](https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL)
|
| 45 |
|
| 46 |
-
|
| 47 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
|
| 49 |
## Citation
|
| 50 |
|
|
@@ -63,4 +66,4 @@ If you find ECG-R1 helpful for your research and applications, please cite our p
|
|
| 63 |
```
|
| 64 |
|
| 65 |
## Acknowledgement
|
| 66 |
-
We thank the authors of [PULSE](https://github.com/AIMedLab/PULSE/tree/dev), [ECG-Chat](https://github.com/YubaoZhao/ECG-Chat), [GEM](https://github.com/lanxiang1017/GEM), and [Swift](https://github.com/modelscope/ms-swift) for their publicly released models, datasets, and training codes.
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen3-VL-8B-Instruct
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
license: apache-2.0
|
| 7 |
pipeline_tag: image-text-to-text
|
| 8 |
tags:
|
| 9 |
- medical
|
| 10 |
+
- ecg
|
| 11 |
+
- multimodal
|
| 12 |
+
- reasoning
|
| 13 |
+
arxiv: 2602.04279
|
| 14 |
+
datasets:
|
| 15 |
+
- PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT
|
| 16 |
---
|
| 17 |
|
| 18 |
<div align="center">
|
|
|
|
| 26 |
<a href='https://arxiv.org/pdf/2602.04279'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
|
| 27 |
<a href='https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-blue'>
|
| 28 |
<a href='https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT'><img src='https://img.shields.io/badge/Dataset-Huggingface-yellow'>
|
| 29 |
+
<a href='http://ai.heartvoice.com.cn/ECG-R1/'><img src='https://img.shields.io/badge/Project-Page-green'></a>
|
| 30 |
|
| 31 |
<p align="center">
|
| 32 |
Jiarui Jin, Haoyu Wang, Xingliang Wu, Xiaocheng Fang, Xiang Lan, Zihan Wang<br/>
|
|
|
|
| 39 |
|
| 40 |
## Introduction
|
| 41 |
|
| 42 |
+
Electrocardiography (ECG) serves as an indispensable diagnostic tool in clinical practice, yet existing multimodal large language models (MLLMs) remain unreliable for ECG interpretation, often producing plausible but clinically incorrect analyses. To address this, we propose ECG-R1, the first reasoning MLLM designed for reliable ECG interpretation via three innovations. First, we construct the interpretation corpus using Protocol-Guided Instruction Data Generation, grounding interpretation in measurable ECG features and monograph-defined quantitative thresholds and diagnostic logic. Second, we present a modality-decoupled architecture with Interleaved Modality Dropout to improve robustness and cross-modal consistency when either the ECG signal or ECG image is missing. Third, we present Reinforcement Learning with ECG Diagnostic Evidence Rewards to strengthen evidence-grounded ECG interpretation. Additionally, we systematically evaluate the ECG interpretation capabilities of proprietary, open-source, and medical MLLMs, and provide the first quantitative evidence that severe hallucinations are widespread, suggesting that the public should not directly trust these outputs without independent verification.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 43 |
|
| 44 |
+
## Resources
|
| 45 |
|
| 46 |
+
- **Paper:** π [Arxiv](https://arxiv.org/abs/2602.04279)
|
| 47 |
+
- **Github:** β¨ [Github](https://github.com/PKUDigitalHealth/ECG-R1)
|
| 48 |
+
- **Online Platform:** π [ECG-R1 Platform](http://ai.heartvoice.com.cn/ECG-R1/)
|
| 49 |
+
- **Model:** π€ [ECG-R1-8B](https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL)
|
| 50 |
+
- **Data:** π€ [ECG-Protocol-Guided-Grounding-CoT](https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT)
|
| 51 |
|
| 52 |
## Citation
|
| 53 |
|
|
|
|
| 66 |
```
|
| 67 |
|
| 68 |
## Acknowledgement
|
| 69 |
+
We thank the authors of [PULSE](https://github.com/AIMedLab/PULSE/tree/dev), [ECG-Chat](https://github.com/YubaoZhao/ECG-Chat), [GEM](https://github.com/lanxiang1017/GEM), and [Swift](https://github.com/modelscope/ms-swift) for their publicly released models, datasets, and training codes.
|