File size: 3,408 Bytes
45949f2 cff8227 0dc55c0 cff8227 339faaf a0ff7c1 45949f2 a0ff7c1 45949f2 cff8227 45949f2 cff8227 45949f2 0dc55c0 cff8227 45949f2 cff8227 45949f2 0dc55c0 45949f2 0dc55c0 45949f2 cff8227 45949f2 cff8227 45949f2 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 | 
**FCMBench** is a multimodal benchmark for credit-risk–oriented workflows. It aims to provide a standard playground to promote collaborative development between academia and industry and provides standardized datasets, prompts, and evaluation scripts across multiple tracks (image, video, speech, agents, etc.)
[简体中文](./README_cn.md)
## 🔥 News
- 【**2026. 03. 16**】✨ We released **FCMBench-V1.1**. This version adds English document images and corresponding QA pairs, expands the covered document types to 26, and increases the dataset to 5,198 images and 13,806 QA samples.
- 【**2026. 01. 01**】We are proud to launch **FCMBench-V1.0**, which covers 18 core certificate types, including 4,043 privacy-compliant images and 8,446 QA samples. It involves 3 types of Perception tasks and 4 types of Reasoning tasks, which are cross-referenced with 10 categories of robustness inferences. All the tasks and inferences are derived from real-world critical scenarios.
> **Status:** Public release (v1.1).
> **Maintainers:** [奇富科技 / Qfin Holdings](https://github.com/QFIN-tech)
> **Contact:** [yangyehuisw@126.com]
---
## Tracks Overview
### 1) Vision-Language Track (✅ Available)
Image-based financial document understanding:
- **Entry:** [Vision-Language Track](vision_language)
- **Inputs:** document images + text prompts (JSONL, one sample per line)
- **Outputs:** text responses (JSONL, one sample per line)
- **Evaluation:** [Evaluation Script](vision_language/evaluation.py)
#### Paper & Project Links
- [**Paper (arXiv)**](https://arxiv.org/abs/2601.00150)
- [**Paper (PDF)**](https://github.com/QFIN-tech/FCMBench/tree/main/TechnicalReport)
- [**Project Page**](https://github.com/QFIN-tech/FCMBench/tree/main/vision_language)
- [**Leaderboard**](https://qfin-tech.github.io/FCMBench)
- [**Sample Data**](https://qfin-tech.github.io/FCMBench/Examples.html)
- [**Dataset (ModelScope)**](https://modelscope.cn/datasets/QFIN/FCMBench-Data)
- [**Dataset (Hugging Face)**](https://huggingface.co/datasets/QFIN/FCMBench-Data)
#### Reference Model Demo
We also provide access to an interactive demo of our Qfin-VL-Instruct model, which achieves strong performance on FCMBench.
If you are interested in trying the Gradio demo, please contact [yangyehui-jk@qifu.com] with the following information:
- Name
- Affiliation / Organization
- Intended use (e.g., research exploration, benchmarking reference)
- Contact email
Access will be granted on a case-by-case basis.
### 2) Video Understanding Track (🕒 Coming Soon)
### 3) Speech Understanding & Generation Track (🕒 Coming Soon)
### 4) Multi-step / Agentic Track (🕒 Coming Soon)
## Citation
```
@misc{yang2026fcmbenchcomprehensivefinancialcredit,
title={FCMBench: A Comprehensive Financial Credit Multimodal Benchmark for Real-world Applications},
author={Yehui Yang and Dalu Yang and Wenshuo Zhou and Fangxin Shang and Yifan Liu and Jie Ren and Haojun Fei and Qing Yang and Yanwu Xu and Tao Chen},
year={2026},
eprint={2601.00150},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2601.00150},
}
```
## Star History
[](https://www.star-history.com/#QFIN-tech/FCMBench&type=date&legend=top-left)
|