Looyyd commited on
Commit
a262b27
·
verified ·
1 Parent(s): 87cc9e5

Training in progress, step 500

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd73425670b194466bebdcad72f0b22f12bf7b81a2b2005979825d42ed7cb64c
3
  size 3087467144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e23abec9e76fdbecbb072ba0be39bc5f2dc3de100eadcca255dc2e9b86e0ac88
3
  size 3087467144
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:145191e3b102f64e84dab1926006ee71528c1ea536df048ef9ff6c48d01f95de
3
- size 11421995
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json CHANGED
@@ -201,6 +201,7 @@
201
  "extra_special_tokens": {},
202
  "model_max_length": 131072,
203
  "pad_token": "<|im_end|>",
 
204
  "split_special_tokens": false,
205
  "tokenizer_class": "Qwen2Tokenizer",
206
  "unk_token": null
 
201
  "extra_special_tokens": {},
202
  "model_max_length": 131072,
203
  "pad_token": "<|im_end|>",
204
+ "padding_size": "left",
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35d90de8e3d794ef624097c9184550e12868edaa3f1397a7f0aaf3d38c89ddbf
3
- size 6033
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cf8feea7e1c77601dde3f7ff0d74ff8b9e984ac97ec29716b2c05d77635ff5c
3
+ size 5560