Upload folder using huggingface_hub
Browse files- .amlignore +6 -0
- .amlignore.amltmp +6 -0
- chunk_0/data-00000-of-00001.arrow +3 -0
- chunk_0/dataset_info.json +29 -0
- chunk_0/state.json +13 -0
- chunk_1/data-00000-of-00001.arrow +3 -0
- chunk_1/dataset_info.json +29 -0
- chunk_1/state.json +13 -0
- tokenization_log.txt +10 -0
.amlignore
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
## This file was auto generated by the Azure Machine Learning Studio. Please do not remove.
|
| 2 |
+
## Read more about the .amlignore file here: https://docs.microsoft.com/azure/machine-learning/how-to-save-write-experiment-files#storage-limits-of-experiment-snapshots
|
| 3 |
+
|
| 4 |
+
.ipynb_aml_checkpoints/
|
| 5 |
+
*.amltmp
|
| 6 |
+
*.amltemp
|
.amlignore.amltmp
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
## This file was auto generated by the Azure Machine Learning Studio. Please do not remove.
|
| 2 |
+
## Read more about the .amlignore file here: https://docs.microsoft.com/azure/machine-learning/how-to-save-write-experiment-files#storage-limits-of-experiment-snapshots
|
| 3 |
+
|
| 4 |
+
.ipynb_aml_checkpoints/
|
| 5 |
+
*.amltmp
|
| 6 |
+
*.amltemp
|
chunk_0/data-00000-of-00001.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d18c64fc4b64811b1ed487e0e3499c61504a5982f54a8b6aabd5c84520da61a
|
| 3 |
+
size 14071720
|
chunk_0/dataset_info.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"citation": "",
|
| 3 |
+
"description": "",
|
| 4 |
+
"features": {
|
| 5 |
+
"input_ids": {
|
| 6 |
+
"feature": {
|
| 7 |
+
"dtype": "int32",
|
| 8 |
+
"_type": "Value"
|
| 9 |
+
},
|
| 10 |
+
"_type": "Sequence"
|
| 11 |
+
},
|
| 12 |
+
"token_type_ids": {
|
| 13 |
+
"feature": {
|
| 14 |
+
"dtype": "int8",
|
| 15 |
+
"_type": "Value"
|
| 16 |
+
},
|
| 17 |
+
"_type": "Sequence"
|
| 18 |
+
},
|
| 19 |
+
"attention_mask": {
|
| 20 |
+
"feature": {
|
| 21 |
+
"dtype": "int8",
|
| 22 |
+
"_type": "Value"
|
| 23 |
+
},
|
| 24 |
+
"_type": "Sequence"
|
| 25 |
+
}
|
| 26 |
+
},
|
| 27 |
+
"homepage": "",
|
| 28 |
+
"license": ""
|
| 29 |
+
}
|
chunk_0/state.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_data_files": [
|
| 3 |
+
{
|
| 4 |
+
"filename": "data-00000-of-00001.arrow"
|
| 5 |
+
}
|
| 6 |
+
],
|
| 7 |
+
"_fingerprint": "b7eddb6c9520c638",
|
| 8 |
+
"_format_columns": null,
|
| 9 |
+
"_format_kwargs": {},
|
| 10 |
+
"_format_type": null,
|
| 11 |
+
"_output_all_columns": false,
|
| 12 |
+
"_split": null
|
| 13 |
+
}
|
chunk_1/data-00000-of-00001.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f2f46949baefb0d3ba7db7c6bf60861cc24ba9b16f9fe4d52eaeef9ad1e2d1bc
|
| 3 |
+
size 7395040
|
chunk_1/dataset_info.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"citation": "",
|
| 3 |
+
"description": "",
|
| 4 |
+
"features": {
|
| 5 |
+
"input_ids": {
|
| 6 |
+
"feature": {
|
| 7 |
+
"dtype": "int32",
|
| 8 |
+
"_type": "Value"
|
| 9 |
+
},
|
| 10 |
+
"_type": "Sequence"
|
| 11 |
+
},
|
| 12 |
+
"token_type_ids": {
|
| 13 |
+
"feature": {
|
| 14 |
+
"dtype": "int8",
|
| 15 |
+
"_type": "Value"
|
| 16 |
+
},
|
| 17 |
+
"_type": "Sequence"
|
| 18 |
+
},
|
| 19 |
+
"attention_mask": {
|
| 20 |
+
"feature": {
|
| 21 |
+
"dtype": "int8",
|
| 22 |
+
"_type": "Value"
|
| 23 |
+
},
|
| 24 |
+
"_type": "Sequence"
|
| 25 |
+
}
|
| 26 |
+
},
|
| 27 |
+
"homepage": "",
|
| 28 |
+
"license": ""
|
| 29 |
+
}
|
chunk_1/state.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_data_files": [
|
| 3 |
+
{
|
| 4 |
+
"filename": "data-00000-of-00001.arrow"
|
| 5 |
+
}
|
| 6 |
+
],
|
| 7 |
+
"_fingerprint": "72f22be3c6ea39e6",
|
| 8 |
+
"_format_columns": null,
|
| 9 |
+
"_format_kwargs": {},
|
| 10 |
+
"_format_type": null,
|
| 11 |
+
"_output_all_columns": false,
|
| 12 |
+
"_split": null
|
| 13 |
+
}
|
tokenization_log.txt
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Tokenization Log
|
| 2 |
+
Start time: 2025-09-08 14:32:54
|
| 3 |
+
End time: 2025-09-08 14:33:05
|
| 4 |
+
Duration: 00:00:10
|
| 5 |
+
Total rows tokenized: 15015
|
| 6 |
+
Total chunks saved: 2
|
| 7 |
+
Chunk size: 10000
|
| 8 |
+
Tokenizer file: tokenizer_output/tokenizer.json
|
| 9 |
+
Shortest sequence length: 5
|
| 10 |
+
Longest sequence length: 8102
|