ZaandaTeika commited on
Commit
75fb08b
·
1 Parent(s): c587a7c

Upload model

Browse files
added_tokens.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
  "</tool_call>": 151658,
3
- "<extra_0>": 151665,
4
  "<tool_call>": 151657,
5
  "<|box_end|>": 151649,
6
  "<|box_start|>": 151648,
 
1
  {
2
  "</tool_call>": 151658,
 
3
  "<tool_call>": 151657,
4
  "<|box_end|>": 151649,
5
  "<|box_start|>": 151648,
config.json CHANGED
@@ -1,18 +1,22 @@
1
  {
2
  "architectures": [
3
- "Qwen2ForProcessRewardModel"
4
  ],
5
  "attention_dropout": 0.0,
6
- "auto_map": {
7
- "AutoConfig": "configuration_qwen2_rm.Qwen2RMConfig",
8
- "AutoModel": "modeling_qwen2_rm.Qwen2ForProcessRewardModel"
9
- },
10
  "dtype": "bfloat16",
11
  "eos_token_id": 151645,
12
  "hidden_act": "silu",
13
  "hidden_size": 1536,
 
 
 
 
14
  "initializer_range": 0.02,
15
  "intermediate_size": 8960,
 
 
 
 
16
  "layer_types": [
17
  "full_attention",
18
  "full_attention",
@@ -55,8 +59,8 @@
55
  "rope_theta": 10000.0,
56
  "sliding_window": null,
57
  "tie_word_embeddings": true,
58
- "transformers_version": "4.57.5",
59
- "use_cache": false,
60
  "use_sliding_window": false,
61
  "vocab_size": 151936
62
  }
 
1
  {
2
  "architectures": [
3
+ "Qwen2ForTokenClassification"
4
  ],
5
  "attention_dropout": 0.0,
 
 
 
 
6
  "dtype": "bfloat16",
7
  "eos_token_id": 151645,
8
  "hidden_act": "silu",
9
  "hidden_size": 1536,
10
+ "id2label": {
11
+ "0": "error",
12
+ "1": "correct"
13
+ },
14
  "initializer_range": 0.02,
15
  "intermediate_size": 8960,
16
+ "label2id": {
17
+ "correct": 1,
18
+ "error": 0
19
+ },
20
  "layer_types": [
21
  "full_attention",
22
  "full_attention",
 
59
  "rope_theta": 10000.0,
60
  "sliding_window": null,
61
  "tie_word_embeddings": true,
62
+ "transformers_version": "4.56.2",
63
+ "use_cache": true,
64
  "use_sliding_window": false,
65
  "vocab_size": 151936
66
  }
model.safetensors.index.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "metadata": {
3
- "total_parameters": 1546078210,
4
- "total_size": 3092195108
5
  },
6
  "weight_map": {
7
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
@@ -342,9 +342,7 @@
342
  "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
343
  "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
344
  "model.norm.weight": "model-00002-of-00002.safetensors",
345
- "score.0.bias": "model-00002-of-00002.safetensors",
346
- "score.0.weight": "model-00002-of-00002.safetensors",
347
- "score.2.bias": "model-00002-of-00002.safetensors",
348
- "score.2.weight": "model-00002-of-00002.safetensors"
349
  }
350
  }
 
1
  {
2
  "metadata": {
3
+ "total_parameters": 1543736630,
4
+ "total_size": 3087473260
5
  },
6
  "weight_map": {
7
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
 
342
  "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
343
  "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
344
  "model.norm.weight": "model-00002-of-00002.safetensors",
345
+ "score.bias": "model-00002-of-00002.safetensors",
346
+ "score.weight": "model-00002-of-00002.safetensors"
 
 
347
  }
348
  }
special_tokens_map.json CHANGED
@@ -1,103 +1,18 @@
1
  {
2
  "additional_special_tokens": [
3
- {
4
- "content": "<|im_start|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_end|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- },
17
- {
18
- "content": "<|object_ref_start|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
- {
25
- "content": "<|object_ref_end|>",
26
- "lstrip": false,
27
- "normalized": false,
28
- "rstrip": false,
29
- "single_word": false
30
- },
31
- {
32
- "content": "<|box_start|>",
33
- "lstrip": false,
34
- "normalized": false,
35
- "rstrip": false,
36
- "single_word": false
37
- },
38
- {
39
- "content": "<|box_end|>",
40
- "lstrip": false,
41
- "normalized": false,
42
- "rstrip": false,
43
- "single_word": false
44
- },
45
- {
46
- "content": "<|quad_start|>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false
51
- },
52
- {
53
- "content": "<|quad_end|>",
54
- "lstrip": false,
55
- "normalized": false,
56
- "rstrip": false,
57
- "single_word": false
58
- },
59
- {
60
- "content": "<|vision_start|>",
61
- "lstrip": false,
62
- "normalized": false,
63
- "rstrip": false,
64
- "single_word": false
65
- },
66
- {
67
- "content": "<|vision_end|>",
68
- "lstrip": false,
69
- "normalized": false,
70
- "rstrip": false,
71
- "single_word": false
72
- },
73
- {
74
- "content": "<|vision_pad|>",
75
- "lstrip": false,
76
- "normalized": false,
77
- "rstrip": false,
78
- "single_word": false
79
- },
80
- {
81
- "content": "<|image_pad|>",
82
- "lstrip": false,
83
- "normalized": false,
84
- "rstrip": false,
85
- "single_word": false
86
- },
87
- {
88
- "content": "<|video_pad|>",
89
- "lstrip": false,
90
- "normalized": false,
91
- "rstrip": false,
92
- "single_word": false
93
- },
94
- {
95
- "content": "<extra_0>",
96
- "lstrip": false,
97
- "normalized": false,
98
- "rstrip": false,
99
- "single_word": false
100
- }
101
  ],
102
  "eos_token": {
103
  "content": "<|im_end|>",
 
1
  {
2
  "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ],
17
  "eos_token": {
18
  "content": "<|im_end|>",
tokenizer_config.json CHANGED
@@ -177,14 +177,6 @@
177
  "rstrip": false,
178
  "single_word": false,
179
  "special": false
180
- },
181
- "151665": {
182
- "content": "<extra_0>",
183
- "lstrip": false,
184
- "normalized": false,
185
- "rstrip": false,
186
- "single_word": false,
187
- "special": true
188
  }
189
  },
190
  "additional_special_tokens": [
@@ -200,8 +192,7 @@
200
  "<|vision_end|>",
201
  "<|vision_pad|>",
202
  "<|image_pad|>",
203
- "<|video_pad|>",
204
- "<extra_0>"
205
  ],
206
  "bos_token": null,
207
  "clean_up_tokenization_spaces": false,
 
177
  "rstrip": false,
178
  "single_word": false,
179
  "special": false
 
 
 
 
 
 
 
 
180
  }
181
  },
182
  "additional_special_tokens": [
 
192
  "<|vision_end|>",
193
  "<|vision_pad|>",
194
  "<|image_pad|>",
195
+ "<|video_pad|>"
 
196
  ],
197
  "bos_token": null,
198
  "clean_up_tokenization_spaces": false,