Update overall_tokenization_stats.json
Browse files
overall_tokenization_stats.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
-
"total_tokens":
|
| 3 |
-
"total_samples":
|
| 4 |
"subset_stats": {
|
| 5 |
"000_00000": {
|
| 6 |
"num_tokens": 2624538498,
|
|
@@ -227,12 +227,12 @@
|
|
| 227 |
"output_size": 1731576827
|
| 228 |
},
|
| 229 |
"004_00004": {
|
| 230 |
-
"num_tokens":
|
| 231 |
-
"num_samples":
|
| 232 |
"skipped": false,
|
| 233 |
"resumed_from": 0,
|
| 234 |
-
"output_samples":
|
| 235 |
-
"output_size":
|
| 236 |
},
|
| 237 |
"004_00005": {
|
| 238 |
"num_tokens": 1126129557,
|
|
|
|
| 1 |
{
|
| 2 |
+
"total_tokens": 71041037918,
|
| 3 |
+
"total_samples": 61064248,
|
| 4 |
"subset_stats": {
|
| 5 |
"000_00000": {
|
| 6 |
"num_tokens": 2624538498,
|
|
|
|
| 227 |
"output_size": 1731576827
|
| 228 |
},
|
| 229 |
"004_00004": {
|
| 230 |
+
"num_tokens": 1179979945,
|
| 231 |
+
"num_samples": 946911,
|
| 232 |
"skipped": false,
|
| 233 |
"resumed_from": 0,
|
| 234 |
+
"output_samples": 946911,
|
| 235 |
+
"output_size": 1614882994
|
| 236 |
},
|
| 237 |
"004_00005": {
|
| 238 |
"num_tokens": 1126129557,
|