amphora commited on
Commit
56265c4
·
verified ·
1 Parent(s): b10b12c

Training in progress, step 68

Browse files
config.json CHANGED
@@ -3,7 +3,7 @@
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
- "bos_token_id": null,
7
  "dtype": "float32",
8
  "eos_token_id": 151645,
9
  "hidden_act": "silu",
 
3
  "Qwen2ForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 151644,
7
  "dtype": "float32",
8
  "eos_token_id": 151645,
9
  "hidden_act": "silu",
debug.log CHANGED
The diff for this file is too large to render. See raw diff
 
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "do_sample": true,
3
  "eos_token_id": [
4
  151645,
 
1
  {
2
+ "bos_token_id": 151644,
3
  "do_sample": true,
4
  "eos_token_id": [
5
  151645,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f8dd47ef850823907fe23974e2531e0fb7a53b71c89081e3e73a5e06d7d5f87
3
  size 7108390424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c75407fe1cd32acd246ca1200cdb3e8863aa93fbd73965a981d30d23ec35a17
3
  size 7108390424
tokenizer_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "add_prefix_space": false,
3
  "backend": "tokenizers",
4
- "bos_token": null,
5
  "clean_up_tokenization_spaces": false,
6
  "eos_token": "<|im_end|>",
7
  "errors": "replace",
 
1
  {
2
  "add_prefix_space": false,
3
  "backend": "tokenizers",
4
+ "bos_token": "<|im_start|>",
5
  "clean_up_tokenization_spaces": false,
6
  "eos_token": "<|im_end|>",
7
  "errors": "replace",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fd8155093cc38d45f771229a6f9ed3099d06e6cc95ad26dc1990811ad048167
3
- size 7249
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e161ebd141b89789d94a5383142dd61c0eab397e92f1778a8efa1d7d410563f
3
+ size 9745