xinshuo commited on
Commit
fd119ca
·
verified ·
1 Parent(s): a48795f

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +122 -142
README.md CHANGED
@@ -1,198 +1,178 @@
1
- # Scientific Coding SWE Dataset (JSONL Format)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
 
3
  ## Dataset Description
4
 
5
- This dataset contains 244 samples extracted from the [xinshuo/Scientific_Coding_SWE_dataset](https://huggingface.co/datasets/xinshuo/Scientific_Coding_SWE_dataset), specifically the `json_dump` field from each sample. The data is provided in JSONL (JSON Lines) format for easy streaming and processing.
 
 
 
 
6
 
7
  ### Dataset Summary
8
 
9
- The Scientific Coding SWE (Software Engineering) dataset focuses on real-world code issues and fixes from prominent scientific computing projects. Each sample represents a complete software engineering task including:
10
 
11
- - Issue descriptions
12
- - Code patches (fixes)
13
  - Test cases
14
- - Metadata about the changes
15
- - Execution results
16
 
17
  ### Supported Tasks
18
 
19
- - **Software Engineering Task Resolution**: Understanding and resolving code issues in scientific computing contexts
20
- - **Code Generation**: Generating fixes for missing or incorrect code
21
- - **Test-Driven Development**: Working with test cases and validation
22
- - **Documentation Generation**: Understanding code documentation needs
23
 
24
- ### Languages
25
 
26
- The dataset covers two primary programming languages:
27
- - **C++**: 159 samples (65.2%)
28
- - **Python**: 85 samples (34.8%)
29
 
30
- ## Dataset Structure
31
 
32
- ### Data Format
 
 
 
 
33
 
34
- The dataset is provided as a single JSONL file (`dataset_json_dumps.jsonl`), where each line is a complete JSON object representing one sample.
35
 
36
  ### Data Fields
37
 
38
- Each sample contains the following fields:
39
-
40
- - `org` (string): Organization name (e.g., "einsteintoolkit", "Qiskit", "openmm")
41
- - `repo` (string): Repository name
42
- - `number` (int): Issue/PR number
43
- - `state` (string): State of the issue (e.g., "open", "closed")
44
- - `title` (string): Title of the issue/PR
45
- - `body` (string): Description of the issue
46
- - `base` (dict): Base branch information with keys:
47
- - `label`: Branch label
48
- - `ref`: Branch reference
49
- - `sha`: Commit SHA
50
- - `resolved_issues` (list): List of related issues with:
51
- - `number`: Issue number
52
- - `title`: Issue title
53
- - `body`: Issue description
54
- - `fix_patch` (string): Git diff/patch for the fix
55
- - `test_patch` (string): Git diff/patch for tests
56
- - `doc_patch` (string): Git diff/patch for documentation
57
- - `tag` (string): Additional tags
58
- - `number_interval` (string): Number interval information
59
- - `lang` (string): Programming language (cpp, python)
60
- - `language` (string): Same as lang
61
- - `instance_id` (string): Unique instance identifier
62
- - `fixed_tests` (dict): Test results for fixed code
63
- - `p2p_tests` (dict): Patch-to-pass tests
64
- - `f2p_tests` (dict): Fail-to-pass tests
65
- - `s2p_tests` (dict): Skip-to-pass tests
66
- - `n2p_tests` (dict): None-to-pass tests
67
- - `run_result` (dict): Execution results with counts of passed/failed/skipped tests
68
- - `test_patch_result` (dict): Test patch execution results
69
- - `fix_patch_result` (dict): Fix patch execution results
70
- - `hints` (string): Hints for solving the issue
71
- - `count_new_files` (int): Number of new files added
72
- - `count_new_entities` (int): Number of new code entities
73
- - `workdir` (string): Working directory path
74
- - `fix_meta_info` (dict): Metadata about the fix
75
- - `test_meta_info` (dict): Metadata about tests
76
- - `doc_meta_info` (dict): Metadata about documentation
77
- - `meta_info` (dict): General metadata
78
- - `pre_commands` (list): Commands to run before testing
79
-
80
- ### Data Instances
81
-
82
- An example instance (truncated for brevity):
83
-
84
- ```json
85
- {
86
- "org": "einsteintoolkit",
87
- "repo": "Cactus",
88
- "number": 1,
89
- "state": "open",
90
- "title": "Add missing LinearExtrapBnd.c in CactusExamples/SampleBoundary",
91
- "body": "This PR adds the missing source file LinearExtrapBnd.c to complete the CactusExamples/SampleBoundary thorn implementation.",
92
- "language": "cpp",
93
- "instance_id": "einsteintoolkit__cactus_1",
94
- "fix_patch": "diff --git a/arrangements/CactusExamples/SampleBoundary/src/LinearExtrapBnd.c ...",
95
- "run_result": {
96
- "passed_count": 1,
97
- "failed_count": 0,
98
- "skipped_count": 0
99
- }
100
- }
101
- ```
102
 
103
- ### Data Splits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
104
 
105
- This dataset contains a single split with all 244 samples.
 
 
 
106
 
107
- ## Source Projects
 
 
108
 
109
- The dataset includes samples from the following scientific computing projects:
110
 
111
- | Organization | Count | Description |
112
- |-------------|-------|-------------|
113
- | openmm | 54 | High-performance molecular dynamics simulation |
114
- | pyscf | 47 | Python-based Simulations of Chemistry Framework |
115
- | rdkit | 46 | Cheminformatics and machine learning software |
116
- | einsteintoolkit | 40 | Computational infrastructure for numerical relativity and astrophysics |
117
- | Qiskit | 38 | Open-source quantum computing framework |
118
- | AMReX-Codes | 19 | Block-structured adaptive mesh refinement framework |
119
 
120
- ## Dataset Creation
 
 
 
121
 
122
- ### Source Data
123
 
124
- This dataset is derived from the [xinshuo/Scientific_Coding_SWE_dataset](https://huggingface.co/datasets/xinshuo/Scientific_Coding_SWE_dataset), which curates software engineering tasks from real GitHub issues and pull requests in scientific computing projects.
 
 
 
 
 
125
 
126
- ### Data Collection Process
127
 
128
- The data was extracted by:
129
- 1. Loading the original dataset using the HuggingFace `datasets` library
130
- 2. Extracting only the `json_dump` field from each sample
131
- 3. Converting to JSONL format for efficient streaming and processing
132
 
133
  ## Considerations for Using the Data
134
 
135
- ### Social Impact of Dataset
136
 
137
- This dataset can be used to:
138
- - Train and evaluate AI models for scientific software engineering tasks
139
- - Improve automated code generation and bug fixing in scientific computing
140
- - Understand common patterns in scientific software development
141
 
142
- ### Discussion of Biases
143
 
144
- The dataset may have the following biases:
145
- - Overrepresentation of certain programming languages (C++ and Python)
146
  - Focus on specific scientific domains (physics, chemistry, quantum computing)
147
- - Representation limited to open-source projects with public GitHub repositories
 
148
 
149
  ## Additional Information
150
 
151
  ### Dataset Curators
152
 
153
- This JSONL version was created from the original xinshuo/Scientific_Coding_SWE_dataset.
 
 
154
 
155
  ### Licensing Information
156
 
157
- Please refer to the original dataset for licensing information.
158
 
159
  ### Citation Information
160
 
161
- If you use this dataset, please cite the original source:
162
-
163
  ```bibtex
164
- @dataset{scientific_coding_swe_dataset,
165
- title={Scientific Coding SWE Dataset},
166
  author={xinshuo},
167
  year={2024},
168
- url={https://huggingface.co/datasets/xinshuo/Scientific_Coding_SWE_dataset}
169
  }
170
  ```
171
 
172
  ### Contributions
173
 
174
- For questions or issues with this JSONL format, please refer to the dataset repository.
175
-
176
- ## Usage Example
177
-
178
- ```python
179
- import json
180
-
181
- # Read and process the dataset
182
- with open('dataset_json_dumps.jsonl', 'r') as f:
183
- for line in f:
184
- sample = json.loads(line)
185
- print(f"Issue: {sample['title']}")
186
- print(f"Language: {sample['language']}")
187
- print(f"Organization: {sample['org']}")
188
- print("---")
189
- ```
190
-
191
- Or use with HuggingFace datasets:
192
-
193
- ```python
194
- from datasets import load_dataset
195
-
196
- dataset = load_dataset("xinshuo/test_jsonl", data_files="dataset_json_dumps.jsonl")
197
- ```
198
 
 
1
+ ---
2
+ dataset_info:
3
+ features:
4
+ - name: question_id
5
+ dtype: string
6
+ - name: question_type
7
+ dtype: string
8
+ - name: description
9
+ dtype: string
10
+ - name: content
11
+ dtype: string
12
+ - name: environment
13
+ dtype: string
14
+ - name: answer
15
+ dtype: string
16
+ - name: test
17
+ dtype: string
18
+ - name: scoring_config
19
+ dtype: string
20
+ splits:
21
+ - name: msb_type
22
+ num_bytes: 15728640
23
+ num_examples: 244
24
+ - name: et_type
25
+ num_bytes: 709558272
26
+ num_examples: 1085
27
+ - name: scicobench_all
28
+ num_bytes: 666894336
29
+ num_examples: 1329
30
+ download_size: 1392181248
31
+ dataset_size: 1392181248
32
+ configs:
33
+ - config_name: default
34
+ data_files:
35
+ - split: msb_type
36
+ path: data/msb_type.jsonl
37
+ - split: et_type
38
+ path: data/et_type.jsonl
39
+ - split: scicobench_all
40
+ path: data/scicobench_all.jsonl
41
+ default: true
42
+ ---
43
+
44
+ # Scientific Coding Benchmark Dataset
45
 
46
  ## Dataset Description
47
 
48
+ This dataset contains scientific coding questions and benchmarks from multiple sources, organized into three splits:
49
+
50
+ - **msb_type** (244 samples): Multi-source Scientific Benchmark questions focusing on software engineering tasks
51
+ - **et_type** (1,085 samples): Einstein Toolkit related coding questions with extensive documentation
52
+ - **scicobench_all** (1,329 samples): Combined and unified format of all scientific coding questions
53
 
54
  ### Dataset Summary
55
 
56
+ The Scientific Coding Benchmark dataset is designed for evaluating AI models on real-world scientific software development tasks. Each sample includes:
57
 
58
+ - Question descriptions and context
59
+ - Code environment specifications
60
  - Test cases
61
+ - Expected answers/solutions
62
+ - Scoring configurations
63
 
64
  ### Supported Tasks
65
 
66
+ - **Scientific Code Generation**: Generate code solutions for scientific computing problems
67
+ - **Software Engineering**: Fix bugs and implement features in scientific software
68
+ - **Documentation Understanding**: Work with complex scientific software documentation
69
+ - **Test-Driven Development**: Write code that passes specific test cases
70
 
71
+ ## Dataset Structure
72
 
73
+ ### Data Splits
 
 
74
 
75
+ The dataset has three splits:
76
 
77
+ | Split | Samples | Size | Description |
78
+ |-------|---------|------|-------------|
79
+ | msb_type | 244 | ~15 MB | Software engineering tasks from scientific projects |
80
+ | et_type | 1,085 | ~677 MB | Einstein Toolkit code generation tasks |
81
+ | scicobench_all | 1,329 | ~636 MB | Unified format combining all question types |
82
 
83
+ **Default Split**: `msb_type`
84
 
85
  ### Data Fields
86
 
87
+ Each sample in the unified format contains:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
88
 
89
+ - `question_id` (string): Unique identifier for the question
90
+ - `question_type` (string): Type/category of the question
91
+ - `description` (string): High-level description of the task
92
+ - `content` (string): Detailed question content and requirements
93
+ - `environment` (string): Environment setup and configuration details
94
+ - `answer` (string): Expected answer or solution
95
+ - `test` (string): Test cases and validation logic
96
+ - `scoring_config` (string): Configuration for scoring the solution
97
+
98
+ ### Data Loading
99
+
100
+ ```python
101
+ from datasets import load_dataset
102
+
103
+ # Load default split (msb_type)
104
+ dataset = load_dataset("xinshuo/test_jsonl")
105
 
106
+ # Load specific split
107
+ msb_data = load_dataset("xinshuo/test_jsonl", split="msb_type")
108
+ et_data = load_dataset("xinshuo/test_jsonl", split="et_type")
109
+ all_data = load_dataset("xinshuo/test_jsonl", split="scicobench_all")
110
 
111
+ # Access a sample
112
+ print(msb_data[0])
113
+ ```
114
 
115
+ ## Source Data
116
 
117
+ ### MSB Type Questions (244 samples)
 
 
 
 
 
 
 
118
 
119
+ Derived from the Scientific Coding SWE dataset, focusing on:
120
+ - **Organizations**: einsteintoolkit, openmm, pyscf, rdkit, Qiskit, AMReX-Codes
121
+ - **Languages**: C++ (65%), Python (35%)
122
+ - **Tasks**: Bug fixes, feature implementation, code completion
123
 
124
+ ### ET Type Questions (1,085 samples)
125
 
126
+ Einstein Toolkit code generation tasks including:
127
+ - Thorn implementations
128
+ - Interface definitions
129
+ - Parameter configurations
130
+ - Schedule definitions
131
+ - Extensive documentation context
132
 
133
+ ### SciCoBench All (1,329 samples)
134
 
135
+ Unified format combining both question types with standardized structure for comprehensive evaluation.
 
 
 
136
 
137
  ## Considerations for Using the Data
138
 
139
+ ### Use Cases
140
 
141
+ - Training code generation models for scientific computing
142
+ - Benchmarking AI coding assistants on scientific software tasks
143
+ - Research in automated scientific software development
144
+ - Education in scientific programming practices
145
 
146
+ ### Limitations
147
 
 
 
148
  - Focus on specific scientific domains (physics, chemistry, quantum computing)
149
+ - Primarily C++ and Python languages
150
+ - May require domain-specific knowledge to evaluate correctly
151
 
152
  ## Additional Information
153
 
154
  ### Dataset Curators
155
 
156
+ This dataset combines and unifies questions from:
157
+ - xinshuo/Scientific_Coding_SWE_dataset
158
+ - xinshuo/ET_1k
159
 
160
  ### Licensing Information
161
 
162
+ Please refer to the original datasets for licensing information.
163
 
164
  ### Citation Information
165
 
 
 
166
  ```bibtex
167
+ @dataset{scicobench,
168
+ title={Scientific Coding Benchmark Dataset},
169
  author={xinshuo},
170
  year={2024},
171
+ url={https://huggingface.co/datasets/xinshuo/test_jsonl}
172
  }
173
  ```
174
 
175
  ### Contributions
176
 
177
+ For questions or issues, please refer to the dataset repository on HuggingFace.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
178