FCMBench-V1.0 / README.md
parapluie-s's picture
Update README.md
a0ff7c1 verified
metadata
configs:
  - config_name: benchmark
    data_files:
      - split: test
        path: vision_language/FCMBench_v1.0_testset_full.jsonl
license: apache-2.0
task_categories:
  - table-question-answering
  - text-classification
pretty_name: FCMBench
size_categories:
  - 10K<n<100K

Main Image

FCMBench is a multimodal benchmark for credit-risk–oriented workflows. It aims to provide a standard playground to promote collaborative development between academia and industry and provides standardized datasets, prompts, and evaluation scripts across multiple tracks (image, video, speech, agents, etc.)

简体中文

🔥 News

  • 2026. 01. 01】✨ We are proud to launch FCMBench-V1.0, which covers 18 core certificate types, including 4,043 privacy-compliant images and 8,446 QA samples. It involves 3 types of Perception tasks and 4 types of Reasoning tasks, which are cross-referenced with 10 categories of robustness inferences. All the tasks and inferences are derived from real-world critical scenarios.

Status: Public release (v1.0).
Maintainers: 奇富科技 / Qfin Holdings
Contact: [yangyehui-jk@qifu.com]


Tracks Overview

1) Vision-Language Track (✅ Available, FCMBench-V1.0)

Image-based financial document understanding:

Paper & Project Links

Reference Model Demo

We also provide access to an interactive demo of our Qfin-VL-Instruct model, which achieves strong performance on FCMBench-V1.0. If you are interested in trying the Gradio demo, please contact [yangyehui-jk@qifu.com] with the following information:

  • Name
  • Affiliation / Organization
  • Intended use (e.g., research exploration, benchmarking reference)
  • Contact email

Access will be granted on a case-by-case basis.

2) Video Understanding Track (🕒 Coming Soon)

3) Speech Understanding & Generation Track (🕒 Coming Soon)

4) Multi-step / Agentic Track (🕒 Coming Soon)

Citation

@misc{yang2026fcmbenchcomprehensivefinancialcredit,
      title={FCMBench: A Comprehensive Financial Credit Multimodal Benchmark for Real-world Applications}, 
      author={Yehui Yang and Dalu Yang and Wenshuo Zhou and Fangxin Shang and Yifan Liu and Jie Ren and Haojun Fei and Qing Yang and Yanwu Xu and Tao Chen},
      year={2026},
      eprint={2601.00150},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2601.00150}, 
}