jouDance commited on
Commit
ec0e62c
·
verified ·
1 Parent(s): dfb5ffd

Upload 6 files

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3789491e75c831e86012df942b938536720c4d19920f3cfbcacec6d7d6b45290
3
  size 142920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:926d327bcb02cde99d75d29fb7e3d2656623e4ba0ed69fe910f759dbf1692e8a
3
  size 142920
special_tokens_map.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "bos_token": "[CLS]",
3
- "cls_token": "[CLS]",
4
- "eos_token": "[SEP]",
5
- "mask_token": "[MASK]",
6
- "pad_token": "[PAD]",
7
- "sep_token": "[SEP]",
8
- "unk_token": "[UNK]"
9
  }
 
1
  {
2
+ "bos_token": "1",
3
+ "cls_token": "1",
4
+ "eos_token": "2",
5
+ "mask_token": "4",
6
+ "pad_token": "3",
7
+ "sep_token": "2",
8
+ "unk_token": "0"
9
  }
tokenizer.json CHANGED
@@ -47,6 +47,51 @@
47
  "rstrip": false,
48
  "normalized": false,
49
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  }
51
  ],
52
  "normalizer": {
@@ -159,7 +204,17 @@
159
  "[SEP]": 2,
160
  "[PAD]": 3,
161
  "[MASK]": 4,
162
- "token": 5
 
 
 
 
 
 
 
 
 
 
163
  },
164
  "unk_token": "[UNK]"
165
  }
 
47
  "rstrip": false,
48
  "normalized": false,
49
  "special": true
50
+ },
51
+ {
52
+ "id": 5,
53
+ "content": "1",
54
+ "single_word": false,
55
+ "lstrip": false,
56
+ "rstrip": false,
57
+ "normalized": false,
58
+ "special": true
59
+ },
60
+ {
61
+ "id": 6,
62
+ "content": "2",
63
+ "single_word": false,
64
+ "lstrip": false,
65
+ "rstrip": false,
66
+ "normalized": false,
67
+ "special": true
68
+ },
69
+ {
70
+ "id": 7,
71
+ "content": "3",
72
+ "single_word": false,
73
+ "lstrip": false,
74
+ "rstrip": false,
75
+ "normalized": false,
76
+ "special": true
77
+ },
78
+ {
79
+ "id": 8,
80
+ "content": "4",
81
+ "single_word": false,
82
+ "lstrip": false,
83
+ "rstrip": false,
84
+ "normalized": false,
85
+ "special": true
86
+ },
87
+ {
88
+ "id": 14,
89
+ "content": "0",
90
+ "single_word": false,
91
+ "lstrip": false,
92
+ "rstrip": false,
93
+ "normalized": false,
94
+ "special": true
95
  }
96
  ],
97
  "normalizer": {
 
204
  "[SEP]": 2,
205
  "[PAD]": 3,
206
  "[MASK]": 4,
207
+ "1": 5,
208
+ "2": 6,
209
+ "3": 7,
210
+ "4": 8,
211
+ "5": 9,
212
+ "6": 10,
213
+ "7": 11,
214
+ "8": 12,
215
+ "9": 13,
216
+ "0": 14,
217
+ "token": 15
218
  },
219
  "unk_token": "[UNK]"
220
  }
tokenizer_config.json CHANGED
@@ -39,19 +39,59 @@
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  }
43
  },
44
- "bos_token": "[CLS]",
45
  "chat_template": "{% for message in messages %}{{ message['content'] }}{% endfor %}",
46
  "clean_up_tokenization_spaces": false,
47
- "cls_token": "[CLS]",
48
- "eos_token": "[SEP]",
49
- "mask_token": "[MASK]",
50
  "max_new_tokens": 1048576,
51
  "model_max_length": 1048576,
52
- "pad_token": "[PAD]",
53
  "padding_side": "right",
54
- "sep_token": "[SEP]",
55
  "tokenizer_class": "PreTrainedTokenizerFast",
56
- "unk_token": "[UNK]"
57
  }
 
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
+ },
43
+ "5": {
44
+ "content": "1",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "6": {
52
+ "content": "2",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "7": {
60
+ "content": "3",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "8": {
68
+ "content": "4",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "14": {
76
+ "content": "0",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
  }
83
  },
84
+ "bos_token": "1",
85
  "chat_template": "{% for message in messages %}{{ message['content'] }}{% endfor %}",
86
  "clean_up_tokenization_spaces": false,
87
+ "cls_token": "1",
88
+ "eos_token": "2",
89
+ "mask_token": "4",
90
  "max_new_tokens": 1048576,
91
  "model_max_length": 1048576,
92
+ "pad_token": "3",
93
  "padding_side": "right",
94
+ "sep_token": "2",
95
  "tokenizer_class": "PreTrainedTokenizerFast",
96
+ "unk_token": "0"
97
  }