xinshuo commited on
Commit
3797f78
·
verified ·
1 Parent(s): fd119ca

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +115 -66
README.md CHANGED
@@ -1,42 +1,16 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: question_id
5
- dtype: string
6
- - name: question_type
7
- dtype: string
8
- - name: description
9
- dtype: string
10
- - name: content
11
- dtype: string
12
- - name: environment
13
- dtype: string
14
- - name: answer
15
- dtype: string
16
- - name: test
17
- dtype: string
18
- - name: scoring_config
19
- dtype: string
20
- splits:
21
- - name: msb_type
22
- num_bytes: 15728640
23
- num_examples: 244
24
- - name: et_type
25
- num_bytes: 709558272
26
- num_examples: 1085
27
- - name: scicobench_all
28
- num_bytes: 666894336
29
- num_examples: 1329
30
- download_size: 1392181248
31
- dataset_size: 1392181248
32
  configs:
33
- - config_name: default
34
  data_files:
35
- - split: msb_type
36
  path: data/msb_type.jsonl
37
- - split: et_type
 
 
38
  path: data/et_type.jsonl
39
- - split: scicobench_all
 
 
40
  path: data/scicobench_all.jsonl
41
  default: true
42
  ---
@@ -45,21 +19,15 @@ configs:
45
 
46
  ## Dataset Description
47
 
48
- This dataset contains scientific coding questions and benchmarks from multiple sources, organized into three splits:
49
 
50
  - **msb_type** (244 samples): Multi-source Scientific Benchmark questions focusing on software engineering tasks
51
  - **et_type** (1,085 samples): Einstein Toolkit related coding questions with extensive documentation
52
- - **scicobench_all** (1,329 samples): Combined and unified format of all scientific coding questions
53
 
54
  ### Dataset Summary
55
 
56
- The Scientific Coding Benchmark dataset is designed for evaluating AI models on real-world scientific software development tasks. Each sample includes:
57
-
58
- - Question descriptions and context
59
- - Code environment specifications
60
- - Test cases
61
- - Expected answers/solutions
62
- - Scoring configurations
63
 
64
  ### Supported Tasks
65
 
@@ -70,24 +38,44 @@ The Scientific Coding Benchmark dataset is designed for evaluating AI models on
70
 
71
  ## Dataset Structure
72
 
73
- ### Data Splits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
74
 
75
- The dataset has three splits:
 
 
76
 
77
- | Split | Samples | Size | Description |
78
- |-------|---------|------|-------------|
79
- | msb_type | 244 | ~15 MB | Software engineering tasks from scientific projects |
80
- | et_type | 1,085 | ~677 MB | Einstein Toolkit code generation tasks |
81
- | scicobench_all | 1,329 | ~636 MB | Unified format combining all question types |
82
 
83
- **Default Split**: `msb_type`
 
 
84
 
85
  ### Data Fields
86
 
87
- Each sample in the unified format contains:
 
 
88
 
89
  - `question_id` (string): Unique identifier for the question
90
- - `question_type` (string): Type/category of the question
91
  - `description` (string): High-level description of the task
92
  - `content` (string): Detailed question content and requirements
93
  - `environment` (string): Environment setup and configuration details
@@ -95,22 +83,24 @@ Each sample in the unified format contains:
95
  - `test` (string): Test cases and validation logic
96
  - `scoring_config` (string): Configuration for scoring the solution
97
 
98
- ### Data Loading
99
 
100
- ```python
101
- from datasets import load_dataset
102
-
103
- # Load default split (msb_type)
104
- dataset = load_dataset("xinshuo/test_jsonl")
 
105
 
106
- # Load specific split
107
- msb_data = load_dataset("xinshuo/test_jsonl", split="msb_type")
108
- et_data = load_dataset("xinshuo/test_jsonl", split="et_type")
109
- all_data = load_dataset("xinshuo/test_jsonl", split="scicobench_all")
110
 
111
- # Access a sample
112
- print(msb_data[0])
113
- ```
 
 
 
 
114
 
115
  ## Source Data
116
 
@@ -134,6 +124,52 @@ Einstein Toolkit code generation tasks including:
134
 
135
  Unified format combining both question types with standardized structure for comprehensive evaluation.
136
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
137
  ## Considerations for Using the Data
138
 
139
  ### Use Cases
@@ -143,11 +179,24 @@ Unified format combining both question types with standardized structure for com
143
  - Research in automated scientific software development
144
  - Education in scientific programming practices
145
 
 
 
 
 
 
 
 
 
 
 
 
 
146
  ### Limitations
147
 
148
  - Focus on specific scientific domains (physics, chemistry, quantum computing)
149
  - Primarily C++ and Python languages
150
  - May require domain-specific knowledge to evaluate correctly
 
151
 
152
  ## Additional Information
153
 
 
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  configs:
3
+ - config_name: msb_type
4
  data_files:
5
+ - split: train
6
  path: data/msb_type.jsonl
7
+ - config_name: et_type
8
+ data_files:
9
+ - split: train
10
  path: data/et_type.jsonl
11
+ - config_name: scicobench_all
12
+ data_files:
13
+ - split: train
14
  path: data/scicobench_all.jsonl
15
  default: true
16
  ---
 
19
 
20
  ## Dataset Description
21
 
22
+ This dataset contains scientific coding questions and benchmarks from multiple sources, organized into three configurations:
23
 
24
  - **msb_type** (244 samples): Multi-source Scientific Benchmark questions focusing on software engineering tasks
25
  - **et_type** (1,085 samples): Einstein Toolkit related coding questions with extensive documentation
26
+ - **scicobench_all** (1,329 samples, **default**): Combined and unified format of all scientific coding questions
27
 
28
  ### Dataset Summary
29
 
30
+ The Scientific Coding Benchmark dataset is designed for evaluating AI models on real-world scientific software development tasks. The dataset provides both raw data formats and a unified format for comprehensive evaluation.
 
 
 
 
 
 
31
 
32
  ### Supported Tasks
33
 
 
38
 
39
  ## Dataset Structure
40
 
41
+ ### Data Configurations
42
+
43
+ The dataset has three configurations with different schemas:
44
+
45
+ | Configuration | Samples | Size | Description |
46
+ |---------------|---------|------|-------------|
47
+ | msb_type | 244 | ~15 MB | Software engineering tasks (raw format) |
48
+ | et_type | 1,085 | ~677 MB | Einstein Toolkit tasks (raw format) |
49
+ | **scicobench_all** | 1,329 | ~636 MB | **Unified format (default)** |
50
+
51
+ **Default Configuration**: `scicobench_all`
52
+
53
+ ### Data Loading
54
+
55
+ ```python
56
+ from datasets import load_dataset
57
 
58
+ # Load default configuration (scicobench_all - unified format)
59
+ dataset = load_dataset("xinshuo/test_jsonl")
60
+ print(f"Loaded {len(dataset['train'])} samples")
61
 
62
+ # Load specific configurations
63
+ msb_data = load_dataset("xinshuo/test_jsonl", "msb_type")
64
+ et_data = load_dataset("xinshuo/test_jsonl", "et_type")
65
+ all_data = load_dataset("xinshuo/test_jsonl", "scicobench_all")
 
66
 
67
+ # Access a sample
68
+ print(dataset['train'][0])
69
+ ```
70
 
71
  ### Data Fields
72
 
73
+ #### Unified Format (scicobench_all)
74
+
75
+ Each sample contains:
76
 
77
  - `question_id` (string): Unique identifier for the question
78
+ - `question_type` (string): Type/category of the question (MSB or ET)
79
  - `description` (string): High-level description of the task
80
  - `content` (string): Detailed question content and requirements
81
  - `environment` (string): Environment setup and configuration details
 
83
  - `test` (string): Test cases and validation logic
84
  - `scoring_config` (string): Configuration for scoring the solution
85
 
86
+ #### MSB Type Format (msb_type)
87
 
88
+ Raw format from Scientific Coding SWE dataset with 33 fields including:
89
+ - `org`, `repo`, `number`, `state`, `title`, `body`
90
+ - `fix_patch`, `test_patch`, `doc_patch`
91
+ - `language`, `instance_id`
92
+ - `fixed_tests`, `run_result`, `test_patch_result`, `fix_patch_result`
93
+ - And more... (see samples for full schema)
94
 
95
+ #### ET Type Format (et_type)
 
 
 
96
 
97
+ Raw format from Einstein Toolkit with 15 fields including:
98
+ - `thorn_name`, `url`, `configuration`
99
+ - `interface`, `param`, `schedule`
100
+ - `src_filename`, `src_code`
101
+ - `context`, `doc_tex_content`, `readme_content`
102
+ - `combined_doc_context`
103
+ - And more... (see samples for full schema)
104
 
105
  ## Source Data
106
 
 
124
 
125
  Unified format combining both question types with standardized structure for comprehensive evaluation.
126
 
127
+ ## Usage Examples
128
+
129
+ ### Load and Iterate Through Questions
130
+
131
+ ```python
132
+ from datasets import load_dataset
133
+
134
+ # Load the unified format (recommended)
135
+ dataset = load_dataset("xinshuo/test_jsonl")
136
+
137
+ for sample in dataset['train']:
138
+ print(f"Question ID: {sample['question_id']}")
139
+ print(f"Type: {sample['question_type']}")
140
+ print(f"Description: {sample['description'][:100]}...")
141
+ print("-" * 80)
142
+ ```
143
+
144
+ ### Filter by Question Type
145
+
146
+ ```python
147
+ # Filter MSB-type questions from the unified dataset
148
+ msb_questions = dataset['train'].filter(lambda x: x['question_type'] == 'MSB')
149
+ print(f"MSB questions: {len(msb_questions)}")
150
+
151
+ # Filter ET-type questions
152
+ et_questions = dataset['train'].filter(lambda x: x['question_type'] == 'ET')
153
+ print(f"ET questions: {len(et_questions)}")
154
+ ```
155
+
156
+ ### Work with Raw Formats
157
+
158
+ ```python
159
+ # Load raw MSB format for detailed metadata
160
+ msb_raw = load_dataset("xinshuo/test_jsonl", "msb_type")
161
+ sample = msb_raw['train'][0]
162
+ print(f"Organization: {sample['org']}")
163
+ print(f"Repository: {sample['repo']}")
164
+ print(f"Fix patch:\n{sample['fix_patch'][:200]}...")
165
+
166
+ # Load raw ET format for full documentation
167
+ et_raw = load_dataset("xinshuo/test_jsonl", "et_type")
168
+ sample = et_raw['train'][0]
169
+ print(f"Thorn: {sample['thorn_name']}")
170
+ print(f"Documentation: {sample['combined_doc_context'][:200]}...")
171
+ ```
172
+
173
  ## Considerations for Using the Data
174
 
175
  ### Use Cases
 
179
  - Research in automated scientific software development
180
  - Education in scientific programming practices
181
 
182
+ ### Recommended Configuration
183
+
184
+ For most use cases, we recommend using the **default `scicobench_all` configuration**, which provides:
185
+ - Unified schema across all questions
186
+ - Standardized fields for easier processing
187
+ - Complete coverage of both MSB and ET question types
188
+
189
+ Use the raw configurations (`msb_type` or `et_type`) only if you need:
190
+ - Original metadata and detailed execution results
191
+ - Domain-specific fields
192
+ - Full documentation context
193
+
194
  ### Limitations
195
 
196
  - Focus on specific scientific domains (physics, chemistry, quantum computing)
197
  - Primarily C++ and Python languages
198
  - May require domain-specific knowledge to evaluate correctly
199
+ - Different configurations have different schemas
200
 
201
  ## Additional Information
202