Datasets:
File size: 2,107 Bytes
a92aafe 62d94b7 a92aafe | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 | {
"dataset_name": "sfe-scientists-first-exam",
"pretty_name": "SFE (Scientists First Exam) Multimodal Science Benchmark",
"name_zh": "SFE 科学家第一考试多模态评测数据集",
"organization": "obaydata",
"description": "SFE (Scientists First Exam) is a multimodal scientific LLM evaluation benchmark. It uses original research data combined with expert annotations to construct high-difficulty scientific reasoning datasets, evaluating models' full-chain research capabilities from signal perception and attribute understanding to complex reasoning.\n\n**Domains:** Astronomy, Earth Science, Life Science, Materials Science, and more.\n\n**Data Production:** Domain experts combine the latest research results to annotate high-difficulty multimodal image-text questions with answer explanations.\n\n**Difficulty Standard:** Data is produced to the pass@4 standard where the model answers correctly at most twice out of four attempts.",
"description_zh": "SFE(Scientists First Exam)是多模态科学大模型评测基准,核心是用原始科研数据 + 专家标注构建高难度科学推理数据集,评估模型从信号感知、属性理解到复杂推理的全链条科研能力。\n\n由专家结合最新研究成果在天文学、地球科学、生命科学、材料科学等领域进行标注,生产出具备高难度的多模态图文题目并出具答案解析。生产数据按照pass@4标准模型至多答对两次。",
"scale": "Weekly production capacity: 300 entries",
"format": "JSON with images",
"use_case": "Evaluating multimodal LLMs on expert-level scientific reasoning across multiple disciplines",
"license": "cc-by-nc-4.0",
"copyright": "Full copyright provided",
"tags": [
"science",
"multimodal",
"benchmark",
"evaluation",
"astronomy",
"earth-science",
"biology",
"materials-science"
],
"task_categories": [
"visual-question-answering",
"question-answering"
],
"sample_link": "",
"contact": "simon.su@obaydata.com",
"homepage": "https://obaydata.com"
} |