Normalize split_name values in nano_bm25_subset_config.json to short names
Browse files
nano_bm25_subset_config.json
CHANGED
|
@@ -23,7 +23,7 @@
|
|
| 23 |
},
|
| 24 |
"splits": [
|
| 25 |
{
|
| 26 |
-
"split_name": "
|
| 27 |
"tokenization_plan": {
|
| 28 |
"mode": "transformer",
|
| 29 |
"language": "unknown",
|
|
@@ -96,7 +96,7 @@
|
|
| 96 |
]
|
| 97 |
},
|
| 98 |
{
|
| 99 |
-
"split_name": "
|
| 100 |
"tokenization_plan": {
|
| 101 |
"mode": "transformer",
|
| 102 |
"language": "unknown",
|
|
@@ -169,7 +169,7 @@
|
|
| 169 |
]
|
| 170 |
},
|
| 171 |
{
|
| 172 |
-
"split_name": "
|
| 173 |
"tokenization_plan": {
|
| 174 |
"mode": "transformer",
|
| 175 |
"language": "unknown",
|
|
@@ -242,7 +242,7 @@
|
|
| 242 |
]
|
| 243 |
},
|
| 244 |
{
|
| 245 |
-
"split_name": "
|
| 246 |
"tokenization_plan": {
|
| 247 |
"mode": "transformer",
|
| 248 |
"language": "unknown",
|
|
@@ -315,7 +315,7 @@
|
|
| 315 |
]
|
| 316 |
},
|
| 317 |
{
|
| 318 |
-
"split_name": "
|
| 319 |
"tokenization_plan": {
|
| 320 |
"mode": "transformer",
|
| 321 |
"language": "unknown",
|
|
@@ -388,7 +388,7 @@
|
|
| 388 |
]
|
| 389 |
},
|
| 390 |
{
|
| 391 |
-
"split_name": "
|
| 392 |
"tokenization_plan": {
|
| 393 |
"mode": "transformer",
|
| 394 |
"language": "unknown",
|
|
@@ -461,4 +461,4 @@
|
|
| 461 |
]
|
| 462 |
}
|
| 463 |
]
|
| 464 |
-
}
|
|
|
|
| 23 |
},
|
| 24 |
"splits": [
|
| 25 |
{
|
| 26 |
+
"split_name": "go",
|
| 27 |
"tokenization_plan": {
|
| 28 |
"mode": "transformer",
|
| 29 |
"language": "unknown",
|
|
|
|
| 96 |
]
|
| 97 |
},
|
| 98 |
{
|
| 99 |
+
"split_name": "java",
|
| 100 |
"tokenization_plan": {
|
| 101 |
"mode": "transformer",
|
| 102 |
"language": "unknown",
|
|
|
|
| 169 |
]
|
| 170 |
},
|
| 171 |
{
|
| 172 |
+
"split_name": "javascript",
|
| 173 |
"tokenization_plan": {
|
| 174 |
"mode": "transformer",
|
| 175 |
"language": "unknown",
|
|
|
|
| 242 |
]
|
| 243 |
},
|
| 244 |
{
|
| 245 |
+
"split_name": "php",
|
| 246 |
"tokenization_plan": {
|
| 247 |
"mode": "transformer",
|
| 248 |
"language": "unknown",
|
|
|
|
| 315 |
]
|
| 316 |
},
|
| 317 |
{
|
| 318 |
+
"split_name": "python",
|
| 319 |
"tokenization_plan": {
|
| 320 |
"mode": "transformer",
|
| 321 |
"language": "unknown",
|
|
|
|
| 388 |
]
|
| 389 |
},
|
| 390 |
{
|
| 391 |
+
"split_name": "ruby",
|
| 392 |
"tokenization_plan": {
|
| 393 |
"mode": "transformer",
|
| 394 |
"language": "unknown",
|
|
|
|
| 461 |
]
|
| 462 |
}
|
| 463 |
]
|
| 464 |
+
}
|