|
|
--- |
|
|
license: mit |
|
|
task_categories: |
|
|
- question-answering |
|
|
language: |
|
|
- en |
|
|
- zh |
|
|
--- |
|
|
|
|
|
# ORION Evaluation Results |
|
|
|
|
|
ORION is a multilingual benchmark designed to evaluate open-domain reasoning across diverse web-related domains. Each example in the dataset requires multi-step logical composition grounded in verifiable sources, challenging advanced AI assistants and retrieval-augmented models. The dataset consists of 310 questions (170 in Chinese and 140 in English), each accompanied by verified answers, diverse acceptable variants (e.g., aliases or synonymous expressions, separated by `|`), and evidence URLs to ensure fair and flexible evaluation.The table below reports the accuracy of three AI systems on ORION. |
|
|
|
|
|
| **AI System** | **Chinese (%)** | **English (%)** | **Overall (%)** | |
|
|
|--------------------------|------------------|------------------|------------------| |
|
|
| Kimi Exploration Edition | 14.7 | 20.0 | 17.1 | |
|
|
| Doubao Search | 23.5 | 30.7 | 26.8 | |
|
|
| Qwen2.5-Max Search | 20.0 | 20.7 | 20.3 | |
|
|
|
|
|
|
|
|
# Citation |