AnonymousArt commited on
Commit
0f4eb78
·
1 Parent(s): 0c45ac1

Upload tokenizer

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ay": 250127,
3
+ "bn": 250121,
4
+ "bs": 250120,
5
+ "cs": 250124,
6
+ "eo": 250119,
7
+ "eu": 250123,
8
+ "gd": 250103,
9
+ "gn": 250109,
10
+ "jv": 250104,
11
+ "kn": 250105,
12
+ "lg": 250114,
13
+ "mk": 250100,
14
+ "mn": 250116,
15
+ "mr": 250112,
16
+ "ms": 250111,
17
+ "nl": 250107,
18
+ "ps": 250113,
19
+ "sd": 250110,
20
+ "sl": 250129,
21
+ "so": 250102,
22
+ "sq": 250128,
23
+ "sr": 250122,
24
+ "sv": 250125,
25
+ "sw": 250117,
26
+ "tk": 250108,
27
+ "tl": 250115,
28
+ "ts": 250126,
29
+ "yi": 250101,
30
+ "zh-CN": 250106,
31
+ "zh-TW": 250118
32
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "eos_token": "</s>",
3
+ "pad_token": "<pad>",
4
+ "unk_token": "<unk>"
5
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db723bb2cbad20506465e5c75220514c9280a8718792fc7d2f39aa9af04e4caf
3
+ size 16332610
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": null,
3
+ "eos_token": "</s>",
4
+ "extra_ids": 0,
5
+ "model_max_length": 1000000000000000019884624838656,
6
+ "pad_token": "<pad>",
7
+ "return_tensors": "pt",
8
+ "sp_model_kwargs": {},
9
+ "special_tokens_map_file": "C:\\Users\\alek/.cache\\huggingface\\hub\\models--google--mt5-small\\snapshots\\38f23af8ec210eb6c376d40e9c56bd25a80f195d\\special_tokens_map.json",
10
+ "tokenizer_class": "T5Tokenizer",
11
+ "unk_token": "<unk>",
12
+ "use_fast": true
13
+ }