sequencr-ai-admin commited on
Commit
e05d05b
·
verified ·
1 Parent(s): 4f059ae

Upload model trained with Unsloth

Browse files

Upload model trained with Unsloth 2x faster

special_tokens_map.json CHANGED
@@ -6,7 +6,13 @@
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
- "eos_token": "<unk>",
 
 
 
 
 
 
10
  "pad_token": {
11
  "content": "<pad>",
12
  "lstrip": false,
 
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
  "pad_token": {
17
  "content": "<pad>",
18
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0240ce510f08e6c2041724e9043e33be9d251d1e4a4d94eb68cd47b954b61d2
3
- size 17078292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f0a9c36fb297452029a786652fd2a7ad7eea9fd6ff802ad083f7f747a4a3875
3
+ size 17078391
tokenizer_config.json CHANGED
@@ -8006,7 +8006,7 @@
8006
  },
8007
  "bos_token": "<s>",
8008
  "clean_up_tokenization_spaces": false,
8009
- "eos_token": "<unk>",
8010
  "extra_special_tokens": {},
8011
  "model_max_length": 131072,
8012
  "pad_token": "<pad>",
 
8006
  },
8007
  "bos_token": "<s>",
8008
  "clean_up_tokenization_spaces": false,
8009
+ "eos_token": "</s>",
8010
  "extra_special_tokens": {},
8011
  "model_max_length": 131072,
8012
  "pad_token": "<pad>",