CWKSC commited on
Commit
ec9489f
·
verified ·
1 Parent(s): 3048d72

Upload tokenizer

Browse files
Files changed (4) hide show
  1. README.md +3 -3
  2. added_tokens.json +2 -2
  3. tokenizer_config.json +4 -5
  4. vocab.json +14 -16
README.md CHANGED
@@ -1,11 +1,11 @@
1
  ---
 
 
 
2
  library_name: transformers
3
  license: mit
4
- base_model: facebook/w2v-bert-2.0
5
  tags:
6
  - generated_from_trainer
7
- datasets:
8
- - common_voice_16_0
9
  model-index:
10
  - name: w2v-bert-2.0-mongolian-colab-CV16.0
11
  results: []
 
1
  ---
2
+ base_model: facebook/w2v-bert-2.0
3
+ datasets:
4
+ - common_voice_16_0
5
  library_name: transformers
6
  license: mit
 
7
  tags:
8
  - generated_from_trainer
 
 
9
  model-index:
10
  - name: w2v-bert-2.0-mongolian-colab-CV16.0
11
  results: []
added_tokens.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "</s>": 38,
3
- "<s>": 37
4
  }
 
1
  {
2
+ "</s>": 36,
3
+ "<s>": 35
4
  }
tokenizer_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "added_tokens_decoder": {
3
- "35": {
4
  "content": "[UNK]",
5
  "lstrip": true,
6
  "normalized": false,
@@ -8,7 +8,7 @@
8
  "single_word": false,
9
  "special": false
10
  },
11
- "36": {
12
  "content": "[PAD]",
13
  "lstrip": true,
14
  "normalized": false,
@@ -16,7 +16,7 @@
16
  "single_word": false,
17
  "special": false
18
  },
19
- "37": {
20
  "content": "<s>",
21
  "lstrip": false,
22
  "normalized": false,
@@ -24,7 +24,7 @@
24
  "single_word": false,
25
  "special": true
26
  },
27
- "38": {
28
  "content": "</s>",
29
  "lstrip": false,
30
  "normalized": false,
@@ -39,7 +39,6 @@
39
  "eos_token": "</s>",
40
  "model_max_length": 1000000000000000019884624838656,
41
  "pad_token": "[PAD]",
42
- "processor_class": "Wav2Vec2BertProcessor",
43
  "replace_word_delimiter_char": " ",
44
  "target_lang": null,
45
  "tokenizer_class": "Wav2Vec2CTCTokenizer",
 
1
  {
2
  "added_tokens_decoder": {
3
+ "33": {
4
  "content": "[UNK]",
5
  "lstrip": true,
6
  "normalized": false,
 
8
  "single_word": false,
9
  "special": false
10
  },
11
+ "34": {
12
  "content": "[PAD]",
13
  "lstrip": true,
14
  "normalized": false,
 
16
  "single_word": false,
17
  "special": false
18
  },
19
+ "35": {
20
  "content": "<s>",
21
  "lstrip": false,
22
  "normalized": false,
 
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "36": {
28
  "content": "</s>",
29
  "lstrip": false,
30
  "normalized": false,
 
39
  "eos_token": "</s>",
40
  "model_max_length": 1000000000000000019884624838656,
41
  "pad_token": "[PAD]",
 
42
  "replace_word_delimiter_char": " ",
43
  "target_lang": null,
44
  "tokenizer_class": "Wav2Vec2CTCTokenizer",
vocab.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "[PAD]": 36,
3
- "[UNK]": 35,
4
  "|": 0,
5
  "а": 1,
6
  "б": 2,
@@ -22,18 +22,16 @@
22
  "с": 18,
23
  "т": 19,
24
  "у": 20,
25
- "ф": 21,
26
- "х": 22,
27
- "ц": 23,
28
- "ч": 24,
29
- "ш": 25,
30
- "ъ": 26,
31
- "ы": 27,
32
- "ь": 28,
33
- "э": 29,
34
- "ю": 30,
35
- "я": 31,
36
- "ё": 32,
37
- "ү": 33,
38
- "ө": 34
39
  }
 
1
  {
2
+ "[PAD]": 34,
3
+ "[UNK]": 33,
4
  "|": 0,
5
  "а": 1,
6
  "б": 2,
 
22
  "с": 18,
23
  "т": 19,
24
  "у": 20,
25
+ "х": 21,
26
+ "ц": 22,
27
+ "ч": 23,
28
+ "ш": 24,
29
+ "ы": 25,
30
+ "ь": 26,
31
+ "э": 27,
32
+ "ю": 28,
33
+ "я": 29,
34
+ "ё": 30,
35
+ "ү": 31,
36
+ "ө": 32
 
 
37
  }