DocTron commited on
Commit
9910f58
·
verified ·
1 Parent(s): 7e12abb

Upload folder using huggingface_hub

Browse files
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8fab348ffa916c73929595cb007325d30138cf1d09d95b2bfa20ba209318f840
3
  size 4990497880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:856a66458e853b1b87190bc8db62eb8707e6d1407f221874d564d3a0b44ea65c
3
  size 4990497880
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf1617573d95d322f41c6b894823757e2363657380db56cab7fd09785667d6d0
3
- size 4663133960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:656db255b90507b2eec3b444906d6caf385a045bb0b9056682062606cc0a639a
3
+ size 3885221448
model.safetensors.index.json CHANGED
@@ -1,10 +1,9 @@
1
  {
2
  "metadata": {
3
  "total_parameters": 4437815808,
4
- "total_size": 9653543936
5
  },
6
  "weight_map": {
7
- "lm_head.weight": "model-00002-of-00002.safetensors",
8
  "model.language_model.embed_tokens.weight": "model-00001-of-00002.safetensors",
9
  "model.language_model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
10
  "model.language_model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
  "total_parameters": 4437815808,
4
+ "total_size": 8875631616
5
  },
6
  "weight_map": {
 
7
  "model.language_model.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
  "model.language_model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
9
  "model.language_model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }