aymanbakiri commited on
Commit
84489cb
verified
1 Parent(s): 75b47f0

Merged SFT + LoRA model for MCQA tasks

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7df57f57c285b610614784893f7166d309cc482f31032a0660011b4413b9101
3
  size 1192135096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da68bd074494bc171c6d279d9459592dbd10c37cf87cff8e5fe6a7e23fa952e2
3
  size 1192135096
special_tokens_map.json CHANGED
@@ -21,5 +21,11 @@
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
- "pad_token": "<|vision_pad|>"
 
 
 
 
 
 
25
  }
 
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
  }
tokenizer_config.json CHANGED
@@ -232,8 +232,8 @@
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
  "model_max_length": 131072,
235
- "pad_token": "<|vision_pad|>",
236
- "padding_side": "left",
237
  "split_special_tokens": false,
238
  "tokenizer_class": "Qwen2Tokenizer",
239
  "unk_token": null
 
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
  "model_max_length": 131072,
235
+ "pad_token": "<|endoftext|>",
236
+ "padding_side": "right",
237
  "split_special_tokens": false,
238
  "tokenizer_class": "Qwen2Tokenizer",
239
  "unk_token": null