christopher commited on
Commit
8872ef7
·
verified ·
1 Parent(s): 9d6ae2e

Upload dataset

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  dataset_info:
3
- config_name: metadata
4
  features:
5
  - name: _id
6
  dtype: string
@@ -38,6 +38,62 @@ dataset_info:
38
  num_examples: 100
39
  download_size: 65425
40
  dataset_size: 65450
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  configs:
42
  - config_name: metadata
43
  data_files:
@@ -51,4 +107,8 @@ configs:
51
  path: metadata/batch_3-*
52
  - split: batch_4
53
  path: metadata/batch_4-*
 
 
 
 
54
  ---
 
1
  ---
2
  dataset_info:
3
+ - config_name: metadata
4
  features:
5
  - name: _id
6
  dtype: string
 
38
  num_examples: 100
39
  download_size: 65425
40
  dataset_size: 65450
41
+ - config_name: tokenizers
42
+ features:
43
+ - name: tokenizer_id
44
+ dtype: int64
45
+ - name: hash
46
+ dtype: string
47
+ - name: vocab_hash
48
+ dtype: string
49
+ - name: vocab_size
50
+ dtype: int64
51
+ - name: model_type
52
+ dtype: string
53
+ - name: num_merges
54
+ dtype: int64
55
+ - name: has_normalizer
56
+ dtype: bool
57
+ - name: has_pre_tokenizer
58
+ dtype: bool
59
+ - name: has_post_processor
60
+ dtype: bool
61
+ - name: has_decoder
62
+ dtype: bool
63
+ - name: num_added_tokens
64
+ dtype: int64
65
+ - name: normalizer_type
66
+ dtype: string
67
+ - name: pre_tokenizer_type
68
+ dtype: string
69
+ - name: decoder_type
70
+ dtype: string
71
+ - name: normalizer_types
72
+ list: string
73
+ - name: pre_tokenizer_types
74
+ list: string
75
+ - name: decoder_types
76
+ list: string
77
+ - name: version
78
+ dtype: string
79
+ - name: added_tokens
80
+ dtype: string
81
+ - name: normalizer
82
+ dtype: string
83
+ - name: pre_tokenizer
84
+ dtype: string
85
+ - name: post_processor
86
+ dtype: string
87
+ - name: decoder
88
+ dtype: string
89
+ - name: model
90
+ dtype: string
91
+ splits:
92
+ - name: train
93
+ num_bytes: 640975277
94
+ num_examples: 195
95
+ download_size: 359764431
96
+ dataset_size: 640975277
97
  configs:
98
  - config_name: metadata
99
  data_files:
 
107
  path: metadata/batch_3-*
108
  - split: batch_4
109
  path: metadata/batch_4-*
110
+ - config_name: tokenizers
111
+ data_files:
112
+ - split: train
113
+ path: tokenizers/train-*
114
  ---
tokenizers/train-00000-of-00002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d00d15956f9739e9e4ba44249d16b132991e0a5afc5ce39749b297d8fde16afd
3
+ size 198187265
tokenizers/train-00001-of-00002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5b1661358b8d4bc5b049cfb239318ca52ebb803c1ceb1eb6b5dad44121cce8c
3
+ size 161577166