|
|
--- |
|
|
license: mit |
|
|
task_categories: |
|
|
- text-generation |
|
|
language: |
|
|
- ja |
|
|
- en |
|
|
tags: |
|
|
- code |
|
|
|
|
|
configs: |
|
|
- config_name: ja |
|
|
data_files: |
|
|
- split: test |
|
|
path: |
|
|
- sakuraeval/ja.parquet |
|
|
- config_name: en |
|
|
data_files: |
|
|
- split: test |
|
|
path: |
|
|
- sakuraeval/en.parquet |
|
|
--- |
|
|
# SakuraEval |
|
|
|
|
|
## Dataset Description |
|
|
SakuraEval is a Japan-specific code generation benchmark dataset. |
|
|
It is designed independently and does not rely on translation from English benchmarks such as HumanEval or JHumanEval. |
|
|
|
|
|
## Dataset Structure |
|
|
```python |
|
|
from datasets import load_dataset |
|
|
load_dataset("kogi-jwu/sakuraeval", "ja") |
|
|
DatasetDict({ |
|
|
test: Dataset({ |
|
|
features: ['task_id', 'category', 'prompt', 'canonical_solution', 'test', 'entry_point'], |
|
|
num_rows: 164 |
|
|
}) |
|
|
}) |
|
|
``` |
|
|
|
|
|
## Data Fields |
|
|
- task_id: Identifier for the data sample. |
|
|
- category: Task category. |
|
|
- prompt: Input for the model, including the function header and docstring that describes the task. |
|
|
- canonical_solution: Solution to the problem presented in the prompt. |
|
|
- test: Function(s) to test the generated code for correctness. |
|
|
- entry_point: Entry point function to begin testing. |
|
|
|
|
|
### Category Breakdown |
|
|
| Category | Number of Tasks | |
|
|
|----------------------------------|-----------------| |
|
|
| 文化(Culture) | 34 | |
|
|
| 風習(Customs) | 27 | |
|
|
| 日本地理(Japanese Geography) | 10 | |
|
|
| 公民・法律(Law and Civics) | 11 | |
|
|
| 数学・科学(Math and Science) | 21 | |
|
|
| 単位変換(Unit Conversion) | 11 | |
|
|
| 日本語処理(Japanese Language) | 43 | |
|
|
| その他(Other) | 7 | |
|
|
|
|
|
|
|
|
## Languages |
|
|
The dataset contains coding problems in 2 natural languages: English and Japanese. |
|
|
|