| | --- |
| | license: cc-by-nc-4.0 |
| | task_categories: |
| | - question-answering |
| | - table-question-answering |
| | language: |
| | - en |
| | tags: |
| | - table-qa |
| | - text-table-qa |
| | - multi-hop-reasoning |
| | - sql |
| | - benchmark |
| | pretty_name: SPARTA |
| | size_categories: |
| | - 1K<n<10K |
| | configs: |
| | - config_name: workload_movie |
| | default: true |
| | data_files: |
| | - split: test |
| | path: workload_movie/test-* |
| | - split: validation |
| | path: workload_movie/validation-* |
| | - config_name: workload_nba |
| | data_files: |
| | - split: test |
| | path: workload_nba/test-* |
| | - split: validation |
| | path: workload_nba/validation-* |
| | - config_name: workload_medical |
| | data_files: |
| | - split: test |
| | path: workload_medical/test-* |
| | - split: validation |
| | path: workload_medical/validation-* |
| | --- |
| | |
| | # SPARTA Benchmark |
| |
|
| | [Paper](https://openreview.net/pdf?id=8KE9qvKhM4) | [Code](https://github.com/pshlego/SPARTA) | [Project Page](https://sparta-projectpage.github.io/) | [Leaderboard](https://sparta.postech.ac.kr/) |
| |
|
| | SPARTA introduces a groundbreaking benchmark for tree-structured multi-hop question answering (QA) across text and tables, addressing the critical shortcomings of existing datasets like HybridQA and OTT-QA, which suffer from shallow reasoning, annotation errors, and limited scale. By constructing a unified reference fact database that merges source tables with grounding tables derived from unstructured passages, our end-to-end framework automates the generation of thousands of high-fidelity QA pairs—requiring only a quarter of the annotation effort—while incorporating advanced operations like aggregation, grouping, and deep nested predicates. Innovative techniques such as provenance-based refinement and realistic-structure enforcement ensure executable, semantically sound queries that mimic real-world complexity, spanning domains like NBA, movies, and medicine. On SPARTA, state-of-the-art models plummet by over 30 F1 points, exposing gaps in cross-modal reasoning and paving the way for more robust QA systems. |
| |
|
| | ## Dataset Structure |
| |
|
| | The dataset contains **3 configs** across 3 domains (Movie, NBA, Medical), each with test and validation splits. |
| |
|
| | ### Domains |
| | - **Movie**: Movie database with films, genres, ratings, and biographical text |
| | - **NBA**: Basketball statistics with player info, teams, awards, and game narratives |
| | - **Medical**: Healthcare data with appointments, billing, treatments, and patient records |
| |
|
| | ### Configs |
| |
|
| | | Config | Splits | Description | |
| | |--------|--------|-------------| |
| | | `workload_movie` | test, validation | 565 + 565 Movie domain queries | |
| | | `workload_nba` | test, validation | 565 + 565 NBA domain queries | |
| | | `workload_medical` | test, validation | 565 + 565 Medical domain queries | |
| |
|
| | ### Splits |
| |
|
| | - **test**: `question_id`, `question`, `table` (3 fields only, answers withheld) |
| | - **validation**: includes SQL query, answer, query metadata (13 fields) |
| |
|
| | ### Validation Split Fields |
| | | Field | Type | Description | |
| | |-------|------|-------------| |
| | | `question_id` | string | Unique query ID (`{domain}:{idx}`) | |
| | | `question` | string | Natural language question | |
| | | `table` | list[string] | Relevant table names | |
| | | `sql_query` | string | SQL query | |
| | | `answer` | list[string] | Ground truth answers | |
| | | `is_nested` | bool | Whether query is nested | |
| | | `is_aggregated` | bool | Whether query uses aggregation | |
| | | `height` | int | Query nesting depth | |
| | | `breadth` | dict | Breadth per level | |
| | | `max_breadth` | int | Maximum breadth | |
| | | `type` | string | Query type description | |
| | | `clause_usage` | dict | SQL clause usage (WHERE, GROUP BY, etc.) | |
| | | `aggregate_usage` | dict | Aggregate function usage (SUM, COUNT, etc.) | |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | |
| | # Load test split (questions only) |
| | test_ds = load_dataset("pshlego/SPARTA", "workload_nba", split="test") |
| | |
| | # Load validation split (with answers and metadata) |
| | val_ds = load_dataset("pshlego/SPARTA", "workload_nba", split="validation") |
| | ``` |
| |
|
| | ## License |
| |
|
| | CC BY-NC 4.0 |
| |
|