abcorrea commited on
Commit
2edadce
·
verified ·
1 Parent(s): d539f69

abcorrea/p1-v1-rep

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: p1-v1-rep
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
 
4
  model_name: p1-v1-rep
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f52f0c4b319c0eb6b6a6b7f67f321d78db7ebdc0952fbb1df267d38571a71bde
3
  size 4965904640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:462b70bacf202530c397054f64a9e7a5b3e4aa2cc784731352eca1fe258a349e
3
  size 4965904640
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bc2247c166f4122035e1977bcc06ed8f5f689ee679701bf5541e3f191df01f4
3
  size 3077766632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bd20f84812086fd7b281c01d5231b90c98fe69c58665096b709461ea8e40015
3
  size 3077766632
special_tokens_map.json CHANGED
@@ -14,7 +14,13 @@
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
- "eos_token": "<|im_end|>",
 
 
 
 
 
 
18
  "pad_token": {
19
  "content": "<|endoftext|>",
20
  "lstrip": false,
 
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
  "pad_token": {
25
  "content": "<|endoftext|>",
26
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -231,7 +231,7 @@
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
- "model_max_length": 32768,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
 
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
+ "model_max_length": 131072,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab1f328ead273ec6079dea2b422b76c0ff227181b28add36cff7ca9fb69750bb
3
  size 6225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b18bef2d432a0e00ee5850a05c34f1c3a6714a0504388473f38db58ae270aef
3
  size 6225