Upload dataset
Browse files
README.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
---
|
| 2 |
dataset_info:
|
| 3 |
-
|
| 4 |
features:
|
| 5 |
- name: _id
|
| 6 |
dtype: string
|
|
@@ -38,6 +38,62 @@ dataset_info:
|
|
| 38 |
num_examples: 100
|
| 39 |
download_size: 65425
|
| 40 |
dataset_size: 65450
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
configs:
|
| 42 |
- config_name: metadata
|
| 43 |
data_files:
|
|
@@ -51,4 +107,8 @@ configs:
|
|
| 51 |
path: metadata/batch_3-*
|
| 52 |
- split: batch_4
|
| 53 |
path: metadata/batch_4-*
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
---
|
|
|
|
| 1 |
---
|
| 2 |
dataset_info:
|
| 3 |
+
- config_name: metadata
|
| 4 |
features:
|
| 5 |
- name: _id
|
| 6 |
dtype: string
|
|
|
|
| 38 |
num_examples: 100
|
| 39 |
download_size: 65425
|
| 40 |
dataset_size: 65450
|
| 41 |
+
- config_name: tokenizers
|
| 42 |
+
features:
|
| 43 |
+
- name: tokenizer_id
|
| 44 |
+
dtype: int64
|
| 45 |
+
- name: hash
|
| 46 |
+
dtype: string
|
| 47 |
+
- name: vocab_hash
|
| 48 |
+
dtype: string
|
| 49 |
+
- name: vocab_size
|
| 50 |
+
dtype: int64
|
| 51 |
+
- name: model_type
|
| 52 |
+
dtype: string
|
| 53 |
+
- name: num_merges
|
| 54 |
+
dtype: int64
|
| 55 |
+
- name: has_normalizer
|
| 56 |
+
dtype: bool
|
| 57 |
+
- name: has_pre_tokenizer
|
| 58 |
+
dtype: bool
|
| 59 |
+
- name: has_post_processor
|
| 60 |
+
dtype: bool
|
| 61 |
+
- name: has_decoder
|
| 62 |
+
dtype: bool
|
| 63 |
+
- name: num_added_tokens
|
| 64 |
+
dtype: int64
|
| 65 |
+
- name: normalizer_type
|
| 66 |
+
dtype: string
|
| 67 |
+
- name: pre_tokenizer_type
|
| 68 |
+
dtype: string
|
| 69 |
+
- name: decoder_type
|
| 70 |
+
dtype: string
|
| 71 |
+
- name: normalizer_types
|
| 72 |
+
list: string
|
| 73 |
+
- name: pre_tokenizer_types
|
| 74 |
+
list: string
|
| 75 |
+
- name: decoder_types
|
| 76 |
+
list: string
|
| 77 |
+
- name: version
|
| 78 |
+
dtype: string
|
| 79 |
+
- name: added_tokens
|
| 80 |
+
dtype: string
|
| 81 |
+
- name: normalizer
|
| 82 |
+
dtype: string
|
| 83 |
+
- name: pre_tokenizer
|
| 84 |
+
dtype: string
|
| 85 |
+
- name: post_processor
|
| 86 |
+
dtype: string
|
| 87 |
+
- name: decoder
|
| 88 |
+
dtype: string
|
| 89 |
+
- name: model
|
| 90 |
+
dtype: string
|
| 91 |
+
splits:
|
| 92 |
+
- name: train
|
| 93 |
+
num_bytes: 640975277
|
| 94 |
+
num_examples: 195
|
| 95 |
+
download_size: 359764431
|
| 96 |
+
dataset_size: 640975277
|
| 97 |
configs:
|
| 98 |
- config_name: metadata
|
| 99 |
data_files:
|
|
|
|
| 107 |
path: metadata/batch_3-*
|
| 108 |
- split: batch_4
|
| 109 |
path: metadata/batch_4-*
|
| 110 |
+
- config_name: tokenizers
|
| 111 |
+
data_files:
|
| 112 |
+
- split: train
|
| 113 |
+
path: tokenizers/train-*
|
| 114 |
---
|
tokenizers/train-00000-of-00002.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d00d15956f9739e9e4ba44249d16b132991e0a5afc5ce39749b297d8fde16afd
|
| 3 |
+
size 198187265
|
tokenizers/train-00001-of-00002.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e5b1661358b8d4bc5b049cfb239318ca52ebb803c1ceb1eb6b5dad44121cce8c
|
| 3 |
+
size 161577166
|