--- language: - en task_categories: - question-answering - table-question-answering tags: - scientific-reasoning - tabular-data - complex-reasoning - algorithmic-reasoning - math pretty_name: SciTaRC size_categories: - n<1K configs: - config_name: default data_files: - split: test path: data/test-* dataset_info: features: - name: paper dtype: string - name: relevant_tables list: list: string - name: tables list: list: string - name: fulltext dtype: string - name: question dtype: string - name: answer dtype: string - name: plan dtype: string splits: - name: test num_bytes: 48991529 num_examples: 371 download_size: 13748575 dataset_size: 48991529 --- # Dataset Card for SciTaRC ## Dataset Description - **Paper:** SciTaRC: Benchmarking QA on Scientific Tabular Data that Requires Language Reasoning and Complex Computation ### Dataset Summary **SciTaRC** (Scientific Table Reasoning and Computation) is an expert-authored benchmark designed to evaluate Large Language Models (LLMs) on complex question-answering tasks over real-world scientific tables. Unlike existing benchmarks that focus on simple table-text integration or single-step operations, SciTaRC focuses on **composite reasoning**—requiring models to execute interdependent operations such as descriptive analysis, complex arithmetic, and ranking across detailed scientific tables. To facilitate granular diagnosis of model failures, every instance includes an expert-annotated **pseudo-code plan** that explicitly outlines the algorithmic reasoning steps required to reach the correct answer. ## Dataset Structure The dataset is provided as a single `test` split containing 370 expert-annotated instances. ### Data Instances A typical instance contains the question, the ground truth answer, the expert-authored pseudo-code plan, the LaTeX representations of the relevant tables, and the full text of the source paper. ### Data Fields Each JSON object in the dataset contains the following fields: - `paper` *(string)*: The arXiv ID of the source scientific paper (e.g., `"2401.06769"`). - `question` *(string)*: The complex, multi-step question asked about the tabular data. - `answer` *(string)*: The ground-truth answer. - `plan` *(string)*: The expert-authored pseudo-code blueprint. It explicitly structures the logical and mathematical operations required to solve the question (e.g., `SELECT`, `LOOP`, `COMPUTE`). - `relevant_tables` *(list of lists of strings)*: The exact LaTeX source code for the specific table(s) required to answer the question. - `tables` *(list of lists of strings)*: The LaTeX source code for all tables and figures extracted from the paper. - `fulltext` *(string)*: The complete LaTeX source text of the original scientific paper, providing full context. ## Citation If you use this dataset, please cite the original paper: ```bibtex @misc{scitarc2026, title={SciTaRC: Benchmarking QA on Scientific Tabular Data that Requires Language Reasoning and Complex Computation}, author={Wang, Hexuan and Ren, Yaxuan and Bommireddypalli, Srikar and Chen, Shuxian and Prabhudesai, Adarsh and Baral, Elina and Zhou, Rongkun and Koehn, Philipp}, year={2026}, url={[Insert ArXiv URL here]} }