Add Arxiv ID to metadata and improve model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +18 -15
README.md CHANGED
@@ -1,12 +1,18 @@
1
  ---
2
- license: apache-2.0
3
- language:
4
- - en
5
  base_model:
6
  - Qwen/Qwen3-VL-8B-Instruct
 
 
 
7
  pipeline_tag: image-text-to-text
8
  tags:
9
  - medical
 
 
 
 
 
 
10
  ---
11
 
12
  <div align="center">
@@ -20,6 +26,7 @@ tags:
20
  <a href='https://arxiv.org/pdf/2602.04279'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
21
  <a href='https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-blue'>
22
  <a href='https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT'><img src='https://img.shields.io/badge/Dataset-Huggingface-yellow'>
 
23
 
24
  <p align="center">
25
  Jiarui Jin, Haoyu Wang, Xingliang Wu, Xiaocheng Fang, Xiang Lan, Zihan Wang<br/>
@@ -32,19 +39,15 @@ tags:
32
 
33
  ## Introduction
34
 
35
- Electrocardiography (ECG) serves as an indispensable diagnostic tool in clinical practice, yet existing multimodal large language models (MLLMs) remain unreliable for ECG interpretation, often producing plausible but clinically incorrect analyses. To address this, we propose ECG-R1, the first reasoning MLLM designed for reliable ECG interpretation via three innovations. First, we construct the interpretation corpus using Protocol-Guided Instruction Data Generation, grounding interpretation in measurable ECG features and monograph-defined quantitative thresholds and diagnostic logic. Second, we present a modality-decoupled architecture with Interleaved Modality Dropout to improve robustness and cross-modal consistency when either the ECG signal or ECG image is missing. Third, we present Reinforcement Learning with ECG Diagnostic Evidence Rewards to strengthen evidence-grounded ECG interpretation. Additionally, we systematically evaluate the ECG interpretation capabilities of proprietary, open-source, and medical MLLMs, and provide the first quantitative evidence that severe hallucinations are widespread, suggesting that the public should not directly trust these outputs without independent verification. Code and data are publicly available at [Github](https://github.com/PKUDigitalHealth/ECG-R1) and [HuggingFace](https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT), and an online platform can be accessed at [ECG-R1-Online-Platform](http://ai.heartvoice.com.cn/ECG-R1/).
36
-
37
- ## Resource
38
-
39
-
40
- #### Paper: πŸ“„ [Arxiv](https://arxiv.org/pdf/2602.04279)
41
-
42
- ### Github: ⌨ [Github](https://github.com/PKUDigitalHealth/ECG-R1)
43
-
44
- #### Model: πŸ€— [ECG-R1-8B](https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL)
45
 
46
- #### Data: πŸ€— [ECG-Protocol-Guided-Grounding-CoT](https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT)
47
 
 
 
 
 
 
48
 
49
  ## Citation
50
 
@@ -63,4 +66,4 @@ If you find ECG-R1 helpful for your research and applications, please cite our p
63
  ```
64
 
65
  ## Acknowledgement
66
- We thank the authors of [PULSE](https://github.com/AIMedLab/PULSE/tree/dev), [ECG-Chat](https://github.com/YubaoZhao/ECG-Chat), [GEM](https://github.com/lanxiang1017/GEM), and [Swift](https://github.com/modelscope/ms-swift) for their publicly released models, datasets, and training codes.
 
1
  ---
 
 
 
2
  base_model:
3
  - Qwen/Qwen3-VL-8B-Instruct
4
+ language:
5
+ - en
6
+ license: apache-2.0
7
  pipeline_tag: image-text-to-text
8
  tags:
9
  - medical
10
+ - ecg
11
+ - multimodal
12
+ - reasoning
13
+ arxiv: 2602.04279
14
+ datasets:
15
+ - PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT
16
  ---
17
 
18
  <div align="center">
 
26
  <a href='https://arxiv.org/pdf/2602.04279'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a>
27
  <a href='https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-blue'>
28
  <a href='https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT'><img src='https://img.shields.io/badge/Dataset-Huggingface-yellow'>
29
+ <a href='http://ai.heartvoice.com.cn/ECG-R1/'><img src='https://img.shields.io/badge/Project-Page-green'></a>
30
 
31
  <p align="center">
32
  Jiarui Jin, Haoyu Wang, Xingliang Wu, Xiaocheng Fang, Xiang Lan, Zihan Wang<br/>
 
39
 
40
  ## Introduction
41
 
42
+ Electrocardiography (ECG) serves as an indispensable diagnostic tool in clinical practice, yet existing multimodal large language models (MLLMs) remain unreliable for ECG interpretation, often producing plausible but clinically incorrect analyses. To address this, we propose ECG-R1, the first reasoning MLLM designed for reliable ECG interpretation via three innovations. First, we construct the interpretation corpus using Protocol-Guided Instruction Data Generation, grounding interpretation in measurable ECG features and monograph-defined quantitative thresholds and diagnostic logic. Second, we present a modality-decoupled architecture with Interleaved Modality Dropout to improve robustness and cross-modal consistency when either the ECG signal or ECG image is missing. Third, we present Reinforcement Learning with ECG Diagnostic Evidence Rewards to strengthen evidence-grounded ECG interpretation. Additionally, we systematically evaluate the ECG interpretation capabilities of proprietary, open-source, and medical MLLMs, and provide the first quantitative evidence that severe hallucinations are widespread, suggesting that the public should not directly trust these outputs without independent verification.
 
 
 
 
 
 
 
 
 
43
 
44
+ ## Resources
45
 
46
+ - **Paper:** πŸ“„ [Arxiv](https://arxiv.org/abs/2602.04279)
47
+ - **Github:** ⌨ [Github](https://github.com/PKUDigitalHealth/ECG-R1)
48
+ - **Online Platform:** 🌐 [ECG-R1 Platform](http://ai.heartvoice.com.cn/ECG-R1/)
49
+ - **Model:** πŸ€— [ECG-R1-8B](https://huggingface.co/PKUDigitalHealth/ECG-R1-8B-RL)
50
+ - **Data:** πŸ€— [ECG-Protocol-Guided-Grounding-CoT](https://huggingface.co/datasets/PKUDigitalHealth/ECG-Protocol-Guided-Grounding-CoT)
51
 
52
  ## Citation
53
 
 
66
  ```
67
 
68
  ## Acknowledgement
69
+ We thank the authors of [PULSE](https://github.com/AIMedLab/PULSE/tree/dev), [ECG-Chat](https://github.com/YubaoZhao/ECG-Chat), [GEM](https://github.com/lanxiang1017/GEM), and [Swift](https://github.com/modelscope/ms-swift) for their publicly released models, datasets, and training codes.