File size: 5,676 Bytes
eaf0baf 8b2cafe eaf0baf 8b2cafe eaf0baf 0d7a29b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 |
---
dataset_info:
features:
- name: instruction
dtype: string
- name: inputs
struct:
- name: function
dtype: string
- name: tests
dtype: string
- name: outputs
sequence: string
- name: meta
struct:
- name: id
dtype: int32
- name: canonical_solution
dtype: string
- name: entry_point
dtype: string
splits:
- name: test
num_bytes: 294098
num_examples: 164
download_size: 101670
dataset_size: 294098
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
---
# **ruCodeEval**
## Task Description
Russian Code Evaluation (ruCodeEval) is the Russian analog of the original HumanEval dataset, created to evaluate the ability of language models to generate code in the Python programming language to solve simple problems.
The dataset contains `164` tasks and aims to measure the functional correctness of code generation based on information from the function's documentation lines—a text description of the function's operation and several examples of results for different input data.
This set is a private analogue of ruHumanEval.
Evaluated skills: Instruction Following, Code Perception, Completion, Algorithms & Data Structures
Contributors: Artem Chervyakov, Albina Akhmetgareeva, Katerina Kolomeytseva, Anastasia Kozlova, Alena Fenogenova
### Motivation
This task tests the ability of models to generate simple Python programs based on a description (condition) in natural language. Since large models have in their training corpus a proportion of texts (programs) written in various programming languages, they are assumed to have the ability to understand and write code for simple tasks.
## Dataset Description
### Data fields
- `instruction` is a string containing instructions for the task;
- `inputs` is a dictionary that contains the following information:
- `function` is a line containing the function signature, as well as its docstring in the form of an unwritten function;
- `tests` is a list of dictionaries that contain input data of test cases for a given task (variants of input data on which the final function code is tested);
- `outputs` is a two-dimensional array of size (n_samples, n_tests), where n_samples is the number of samples required to calculate the pass@k metric, n_tests is the number of test cases in tests; each list in the outputs is the same and contains correct answers to all test cases as strings;
- `meta` is a dictionary containing meta information:
- `id` is an integer indicating the index of the example;
- `canonical_solution` is the canonical solution;
- `entry_point` is the function name.
### Data formatting example
Below is an example from the dataset:
```json
{
"instruction": "Необходимо реализовать логику на языке Python для следующей программы\n{function}",
"inputs": {
"function": "\n\ndef greatest_common_divisor(a: int, b: int) -> int:\n \"\"\"Верните наибольший общий делитель двух целых чисел a и b.\n Примеры: \n greatest_common_divisor(3, 5) \n 1 \n greatest_common_divisor(25, 15) \n 5\n \"\"\"",
"tests": "[{'a': 100, 'b': 50}, {'a': 98, 'b': 56}, {'a': 540, 'b': 288}, {'a': 81, 'b': 27}, {'a': 33, 'b': 55}, {'a': 7, 'b': 13}, {'a': 14, 'b': 28}, {'a': 10, 'b': 25}, {'a': 12, 'b': 54}, {'a': 21, 'b': 35}]"
},
"outputs": [
"50",
"14",
"36",
"27",
"11",
"1",
"14",
"5",
"6",
"7"
],
"meta": {
"id": 13,
"canonical_solution": "\n\n def query_gcd(a: int, b: int) -> int:\n return a if b == 0 else query_gcd(b, a % b)\n return query_gcd(a, b) \n\n",
"entry_point": "greatest_common_divisor"
}
}
```
### Prompts
For this task 10 prompts of varying difficulty were created. Example:
```json
"Допишите код на языке Python в соответствии с условием, приведенным в описании\n{function}"
```
### Dataset Creation
The test set was manually collected from open sources according to the format of the original open set [openai_humaneval](https://huggingface.co/datasets/openai_humaneval), adjusted the dataset to avoid data leakage in training and took into account the corrections described in [2].
## Evaluation
### Metrics
The model is evaluated using the `pass@k` metric, which is computed as follows:
$$ pass@k:=\mathbb{E}_{problems}\left[1-\frac{\binom{n-c}{k}}{\binom{n}{k}}\right] $$
Notation: *n* is the total number of generated solution options, *c* is the number of solutions that are correct, *k* is the selected indicator, how many options are taken into account.
To calculate `pass@k`, `n ≥ k` solutions are generated for each problem and are run through test cases (we use n = 10 and k ≤ 10 and an average of 10 test cases per problem). Then, the number of the correct solutions is calculated (`c ≤ n`). The solution is considered to be correct if it passes all test cases. That means the result of running solutions on test cases should be equal to the correct answers (outputs) for one problem. Such an evaluation process yields an unbiased score.
## References
[1] Chen, Mark, et al. "Evaluating large language models trained on code." arXiv preprint arXiv:2107.03374 (2021).
[2] Jiawei Liu, Chunqiu Steven Xia, Yuyao Wang, Lingming Zhang Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation. |