Commit
·
bdf1661
1
Parent(s):
fc95ebd
Add
Browse files- data/python/data/humanevalbugs.jsonl +0 -0
- humaneval-x-bugs.py +12 -6
data/python/data/humanevalbugs.jsonl
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
humaneval-x-bugs.py
CHANGED
|
@@ -57,42 +57,42 @@ class HumanEvalXBugs(datasets.GeneratorBasedBuilder):
|
|
| 57 |
name="python",
|
| 58 |
description="Python HumanEvalBugs",
|
| 59 |
features=[
|
| 60 |
-
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point"
|
| 61 |
]
|
| 62 |
),
|
| 63 |
HumanEvalXBugsConfig(
|
| 64 |
name="cpp",
|
| 65 |
description="C++ HumanEvalBugs",
|
| 66 |
features=[
|
| 67 |
-
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point"
|
| 68 |
]
|
| 69 |
),
|
| 70 |
HumanEvalXBugsConfig(
|
| 71 |
name="go",
|
| 72 |
description="Go HumanEvalBugs",
|
| 73 |
features=[
|
| 74 |
-
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point"
|
| 75 |
]
|
| 76 |
),
|
| 77 |
HumanEvalXBugsConfig(
|
| 78 |
name="java",
|
| 79 |
description="Java HumanEvalBugs",
|
| 80 |
features=[
|
| 81 |
-
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point"
|
| 82 |
]
|
| 83 |
),
|
| 84 |
HumanEvalXBugsConfig(
|
| 85 |
name="js",
|
| 86 |
description="JavaScript HumanEvalBugs",
|
| 87 |
features=[
|
| 88 |
-
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point"
|
| 89 |
]
|
| 90 |
),
|
| 91 |
HumanEvalXBugsConfig(
|
| 92 |
name="rust",
|
| 93 |
description="JavaScript HumanEvalBugs",
|
| 94 |
features=[
|
| 95 |
-
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point"
|
| 96 |
]
|
| 97 |
),
|
| 98 |
]
|
|
@@ -115,6 +115,9 @@ class HumanEvalXBugs(datasets.GeneratorBasedBuilder):
|
|
| 115 |
"test": datasets.Value("string"),
|
| 116 |
"test_setup": datasets.Value("string"),
|
| 117 |
"example_test": datasets.Value("string"),
|
|
|
|
|
|
|
|
|
|
| 118 |
}
|
| 119 |
),
|
| 120 |
homepage=_HOMEPAGE,
|
|
@@ -158,5 +161,8 @@ class HumanEvalXBugs(datasets.GeneratorBasedBuilder):
|
|
| 158 |
"test": row["test"],
|
| 159 |
"test_setup": row.get("test_setup", ""), # Only for Go
|
| 160 |
"example_test": row["example_test"],
|
|
|
|
|
|
|
|
|
|
| 161 |
}
|
| 162 |
key += 1
|
|
|
|
| 57 |
name="python",
|
| 58 |
description="Python HumanEvalBugs",
|
| 59 |
features=[
|
| 60 |
+
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point", "signature", "docstring", "instruction"
|
| 61 |
]
|
| 62 |
),
|
| 63 |
HumanEvalXBugsConfig(
|
| 64 |
name="cpp",
|
| 65 |
description="C++ HumanEvalBugs",
|
| 66 |
features=[
|
| 67 |
+
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point", "signature", "docstring", "instruction"
|
| 68 |
]
|
| 69 |
),
|
| 70 |
HumanEvalXBugsConfig(
|
| 71 |
name="go",
|
| 72 |
description="Go HumanEvalBugs",
|
| 73 |
features=[
|
| 74 |
+
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point", "signature", "docstring", "instruction"
|
| 75 |
]
|
| 76 |
),
|
| 77 |
HumanEvalXBugsConfig(
|
| 78 |
name="java",
|
| 79 |
description="Java HumanEvalBugs",
|
| 80 |
features=[
|
| 81 |
+
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point", "signature", "docstring", "instruction"
|
| 82 |
]
|
| 83 |
),
|
| 84 |
HumanEvalXBugsConfig(
|
| 85 |
name="js",
|
| 86 |
description="JavaScript HumanEvalBugs",
|
| 87 |
features=[
|
| 88 |
+
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point", "signature", "docstring", "instruction"
|
| 89 |
]
|
| 90 |
),
|
| 91 |
HumanEvalXBugsConfig(
|
| 92 |
name="rust",
|
| 93 |
description="JavaScript HumanEvalBugs",
|
| 94 |
features=[
|
| 95 |
+
"task_id", "prompt", "import", "declaration", "buggy_solution", "canonical_solution", "test", "test_setup", "example_test", "bug_type", "failure_symptoms", "entry_point", "signature", "docstring", "instruction"
|
| 96 |
]
|
| 97 |
),
|
| 98 |
]
|
|
|
|
| 115 |
"test": datasets.Value("string"),
|
| 116 |
"test_setup": datasets.Value("string"),
|
| 117 |
"example_test": datasets.Value("string"),
|
| 118 |
+
"signature": datasets.Value("string"),
|
| 119 |
+
"docstring": datasets.Value("string"),
|
| 120 |
+
"instruction": datasets.Value("string"),
|
| 121 |
}
|
| 122 |
),
|
| 123 |
homepage=_HOMEPAGE,
|
|
|
|
| 161 |
"test": row["test"],
|
| 162 |
"test_setup": row.get("test_setup", ""), # Only for Go
|
| 163 |
"example_test": row["example_test"],
|
| 164 |
+
"signature": row["signature"],
|
| 165 |
+
"docstring": row["docstring"],
|
| 166 |
+
"instruction": row["instruction"],
|
| 167 |
}
|
| 168 |
key += 1
|