sfe-scientists-first-exam / metadata.json
simonsu2000's picture
Fix false-positive edit in metadata.json - restore original text
62d94b7 verified
{
"dataset_name": "sfe-scientists-first-exam",
"pretty_name": "SFE (Scientists First Exam) Multimodal Science Benchmark",
"name_zh": "SFE 科学家第一考试多模态评测数据集",
"organization": "obaydata",
"description": "SFE (Scientists First Exam) is a multimodal scientific LLM evaluation benchmark. It uses original research data combined with expert annotations to construct high-difficulty scientific reasoning datasets, evaluating models' full-chain research capabilities from signal perception and attribute understanding to complex reasoning.\n\n**Domains:** Astronomy, Earth Science, Life Science, Materials Science, and more.\n\n**Data Production:** Domain experts combine the latest research results to annotate high-difficulty multimodal image-text questions with answer explanations.\n\n**Difficulty Standard:** Data is produced to the pass@4 standard where the model answers correctly at most twice out of four attempts.",
"description_zh": "SFE(Scientists First Exam)是多模态科学大模型评测基准,核心是用原始科研数据 + 专家标注构建高难度科学推理数据集,评估模型从信号感知、属性理解到复杂推理的全链条科研能力。\n\n由专家结合最新研究成果在天文学、地球科学、生命科学、材料科学等领域进行标注,生产出具备高难度的多模态图文题目并出具答案解析。生产数据按照pass@4标准模型至多答对两次。",
"scale": "Weekly production capacity: 300 entries",
"format": "JSON with images",
"use_case": "Evaluating multimodal LLMs on expert-level scientific reasoning across multiple disciplines",
"license": "cc-by-nc-4.0",
"copyright": "Full copyright provided",
"tags": [
"science",
"multimodal",
"benchmark",
"evaluation",
"astronomy",
"earth-science",
"biology",
"materials-science"
],
"task_categories": [
"visual-question-answering",
"question-answering"
],
"sample_link": "",
"contact": "simon.su@obaydata.com",
"homepage": "https://obaydata.com"
}