Upload folder using huggingface_hub

#63
Files changed (24) hide show
  1. .gitattributes +6 -0
  2. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/ai2d.json +0 -0
  3. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmbench_en_dev.json +3 -0
  4. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmerealworld_lite.json +3 -0
  5. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmmu_val.json +0 -0
  6. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmstar.json +0 -0
  7. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/pope.json +3 -0
  8. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/rank0_metric_eval_done.txt +1 -0
  9. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/rank1_metric_eval_done.txt +1 -0
  10. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/rank2_metric_eval_done.txt +1 -0
  11. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/results.json +488 -0
  12. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/mmbench_en_dev_results.json +1 -0
  13. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/mmbench_en_dev_results.xlsx +3 -0
  14. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/textvqa_submission_2025-06-13-02-34-59.json +0 -0
  15. sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/textvqa_val.json +3 -0
  16. sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/gqa.json +3 -0
  17. sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/rank0_metric_eval_done.txt +1 -0
  18. sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/results.json +79 -0
  19. sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/ocrbench.json +0 -0
  20. sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/rank0_metric_eval_done.txt +1 -0
  21. sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/rank1_metric_eval_done.txt +1 -0
  22. sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/rank2_metric_eval_done.txt +1 -0
  23. sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/results.json +67 -0
  24. sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/results/ocrbench_results.txt +18 -0
.gitattributes CHANGED
@@ -167,3 +167,9 @@ sft/1M3/Full_xmoe/logs/0626_1221_llava_v1.5_gqa_llava_model_args_2d4392/gqa.json
167
  sft/1M3/Full_xmoe/checkpoint-27572/trainer_state.json filter=lfs diff=lfs merge=lfs -text
168
  sft/1M3/Full_xmoe/checkpoint-34462/trainer_state.json filter=lfs diff=lfs merge=lfs -text
169
  sft/1M3/Full_xmoe/checkpoint-20679/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
167
  sft/1M3/Full_xmoe/checkpoint-27572/trainer_state.json filter=lfs diff=lfs merge=lfs -text
168
  sft/1M3/Full_xmoe/checkpoint-34462/trainer_state.json filter=lfs diff=lfs merge=lfs -text
169
  sft/1M3/Full_xmoe/checkpoint-20679/trainer_state.json filter=lfs diff=lfs merge=lfs -text
170
+ sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
171
+ sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
172
+ sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/pope.json filter=lfs diff=lfs merge=lfs -text
173
+ sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
174
+ sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
175
+ sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/gqa.json filter=lfs diff=lfs merge=lfs -text
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6f25b6a0beeedbbfb0c713dbea283fb787b953459cb8bbc6ac4292d8297c715
3
+ size 14568190
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3734a62fb5df8bf60f2b0a9230e02fad8493220b461ed2b535505ca7a1174a53
3
+ size 1994104223
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/pope.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c1da06ebc9c613d22568a13dd7d22b94426ad7b5055510c490e13c11d5c63bb
3
+ size 17579727
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/results.json ADDED
@@ -0,0 +1,488 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6910621761658031,
5
+ "exact_match_stderr,flexible-extract": 0.00831621652855342,
6
+ "alias": "ai2d"
7
+ },
8
+ "mmbench_en_dev": {
9
+ "gpt_eval_score,none": 70.79037800687286,
10
+ "gpt_eval_score_stderr,none": "N/A",
11
+ "submission,none": null,
12
+ "submission_stderr,none": "N/A",
13
+ "alias": "mmbench_en_dev"
14
+ },
15
+ "mmerealworld_lite": {
16
+ "mme_realworld_score,none": 0.3397602918186555,
17
+ "mme_realworld_score_stderr,none": "N/A",
18
+ "alias": "mmerealworld_lite"
19
+ },
20
+ "mmmu_val": {
21
+ "mmmu_acc,none": 0.43444,
22
+ "mmmu_acc_stderr,none": "N/A",
23
+ "alias": "mmmu_val"
24
+ },
25
+ "mmstar": {
26
+ "coarse perception,none": 0.7033947521250198,
27
+ "coarse perception_stderr,none": "N/A",
28
+ "fine-grained perception,none": 0.35849914026037827,
29
+ "fine-grained perception_stderr,none": "N/A",
30
+ "instance reasoning,none": 0.5014961747473528,
31
+ "instance reasoning_stderr,none": "N/A",
32
+ "logical reasoning,none": 0.3802980298029803,
33
+ "logical reasoning_stderr,none": "N/A",
34
+ "math,none": 0.39419050164840064,
35
+ "math_stderr,none": "N/A",
36
+ "science & technology,none": 0.3278874261499387,
37
+ "science & technology_stderr,none": "N/A",
38
+ "alias": "mmstar"
39
+ },
40
+ "pope": {
41
+ "pope_accuracy,none": 0.8436666666666667,
42
+ "pope_accuracy_stderr,none": "N/A",
43
+ "pope_precision,none": 0.8854268891069677,
44
+ "pope_precision_stderr,none": "N/A",
45
+ "pope_recall,none": 0.8057602143335566,
46
+ "pope_recall_stderr,none": "N/A",
47
+ "pope_f1_score,none": 0.8437171244886031,
48
+ "pope_f1_score_stderr,none": "N/A",
49
+ "pope_yes_ratio,none": 0.5,
50
+ "pope_yes_ratio_stderr,none": "N/A",
51
+ "alias": "pope"
52
+ },
53
+ "textvqa_val": {
54
+ "exact_match,none": 0.42702000027894976,
55
+ "exact_match_stderr,none": 0.006769072896119281,
56
+ "submission,none": null,
57
+ "submission_stderr,none": "N/A",
58
+ "alias": "textvqa_val"
59
+ }
60
+ },
61
+ "configs": {
62
+ "ai2d": {
63
+ "task": "ai2d",
64
+ "dataset_path": "lmms-lab/ai2d",
65
+ "dataset_kwargs": {
66
+ "token": true
67
+ },
68
+ "test_split": "test",
69
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7fe3bc428280>",
70
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7fe3bc428550>",
71
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7fe3bc428af0>",
72
+ "description": "",
73
+ "target_delimiter": " ",
74
+ "fewshot_delimiter": "\n\n",
75
+ "metric_list": [
76
+ {
77
+ "metric": "exact_match",
78
+ "aggregation": "mean",
79
+ "higher_is_better": true,
80
+ "ignore_case": true,
81
+ "ignore_punctuation": true
82
+ }
83
+ ],
84
+ "output_type": "generate_until",
85
+ "generation_kwargs": {
86
+ "max_new_tokens": 512,
87
+ "temperature": 0.0,
88
+ "do_sample": false,
89
+ "until": [
90
+ "\n\n"
91
+ ]
92
+ },
93
+ "repeats": 1,
94
+ "filter_list": [
95
+ {
96
+ "name": "flexible-extract",
97
+ "filter": [
98
+ {
99
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
100
+ "group_select": 0,
101
+ "ignore_case": true,
102
+ "ignore_punctuation": true,
103
+ "regex_pattern": "([A-Z])\\."
104
+ }
105
+ ]
106
+ }
107
+ ],
108
+ "should_decontaminate": false,
109
+ "metadata": [
110
+ {
111
+ "version": 0.0
112
+ }
113
+ ],
114
+ "model_specific_prompt_kwargs": {
115
+ "default": {
116
+ "prompt_format": "mcq",
117
+ "pre_prompt": "",
118
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
119
+ },
120
+ "gpt4v": {
121
+ "prompt_format": "mcq",
122
+ "pre_prompt": "",
123
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
124
+ },
125
+ "qwen_vl": {
126
+ "prompt_format": "qa",
127
+ "pre_prompt": "",
128
+ "post_prompt": " Answer:"
129
+ },
130
+ "xcomposer2_4khd": {
131
+ "prompt_format": "mcq_xcomposer",
132
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
133
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
134
+ }
135
+ },
136
+ "model_specific_target_kwargs": {
137
+ "default": "mcq",
138
+ "qwen_vl": "qa"
139
+ }
140
+ },
141
+ "mmbench_en_dev": {
142
+ "task": "mmbench_en_dev",
143
+ "dataset_path": "lmms-lab/MMBench",
144
+ "dataset_name": "en",
145
+ "dataset_kwargs": {
146
+ "token": true
147
+ },
148
+ "test_split": "dev",
149
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7fe41c0c8af0>",
150
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7fe41c056040>",
151
+ "doc_to_target": "answer",
152
+ "process_results": "<function mmbench_process_results at 0x7fe41c056550>",
153
+ "description": "",
154
+ "target_delimiter": " ",
155
+ "fewshot_delimiter": "\n\n",
156
+ "metric_list": [
157
+ {
158
+ "metric": "gpt_eval_score",
159
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7fe41c0c83a0>",
160
+ "higher_is_better": true
161
+ },
162
+ {
163
+ "metric": "submission",
164
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7fe41c0c88b0>",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "generate_until",
169
+ "generation_kwargs": {
170
+ "until": [
171
+ "ASSISTANT:"
172
+ ],
173
+ "max_new_tokens": 1024,
174
+ "temperature": 0.0,
175
+ "top_p": 1.0,
176
+ "num_beams": 1,
177
+ "do_sample": false,
178
+ "image_aspect_ratio": "original"
179
+ },
180
+ "repeats": 1,
181
+ "should_decontaminate": false,
182
+ "model_specific_prompt_kwargs": {
183
+ "default": {
184
+ "pre_prompt": "",
185
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
186
+ }
187
+ },
188
+ "model_specific_generation_kwargs": {
189
+ "llava": {
190
+ "image_aspect_ratio": "original"
191
+ }
192
+ }
193
+ },
194
+ "mmerealworld_lite": {
195
+ "task": "mmerealworld_lite",
196
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
197
+ "dataset_kwargs": {
198
+ "token": true
199
+ },
200
+ "test_split": "train",
201
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7fe3bcbd0430>",
202
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7fe3bcbd0ee0>",
203
+ "doc_to_target": "answer",
204
+ "process_results": "<function mme_realworld_process_results at 0x7fe3bcbd6b80>",
205
+ "description": "",
206
+ "target_delimiter": " ",
207
+ "fewshot_delimiter": "\n\n",
208
+ "metric_list": [
209
+ {
210
+ "metric": "mme_realworld_score",
211
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7fe3bcbdd700>",
212
+ "higher_is_better": true
213
+ }
214
+ ],
215
+ "output_type": "generate_until",
216
+ "generation_kwargs": {
217
+ "max_new_tokens": 16,
218
+ "temperature": 0.0,
219
+ "top_p": 1.0,
220
+ "num_beams": 1,
221
+ "do_sample": false,
222
+ "until": [
223
+ "\n\n"
224
+ ]
225
+ },
226
+ "repeats": 1,
227
+ "should_decontaminate": false,
228
+ "metadata": [
229
+ {
230
+ "version": 0.0
231
+ }
232
+ ],
233
+ "model_specific_prompt_kwargs": {
234
+ "default": {
235
+ "pre_prompt": "",
236
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
237
+ },
238
+ "gpt4v": {
239
+ "pre_prompt": "",
240
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
241
+ },
242
+ "xcomposer2_4khd": {
243
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
244
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
245
+ }
246
+ }
247
+ },
248
+ "mmmu_val": {
249
+ "task": "mmmu_val",
250
+ "dataset_path": "lmms-lab/MMMU",
251
+ "test_split": "validation",
252
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7fe3ced50e50>",
253
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7fe3ced7cb80>",
254
+ "doc_to_target": "answer",
255
+ "process_results": "<function mmmu_process_results at 0x7fe3cec5fa60>",
256
+ "description": "",
257
+ "target_delimiter": " ",
258
+ "fewshot_delimiter": "\n\n",
259
+ "metric_list": [
260
+ {
261
+ "metric": "mmmu_acc",
262
+ "aggregation": "<function mmmu_aggregate_results at 0x7fe3cebb59d0>",
263
+ "higher_is_better": true
264
+ }
265
+ ],
266
+ "output_type": "generate_until",
267
+ "generation_kwargs": {
268
+ "max_new_tokens": 128,
269
+ "until": [
270
+ "\n\n"
271
+ ],
272
+ "image_aspect_ratio": "original"
273
+ },
274
+ "repeats": 1,
275
+ "should_decontaminate": false,
276
+ "metadata": [
277
+ {
278
+ "version": 0.0
279
+ }
280
+ ],
281
+ "model_specific_generation_kwargs": {
282
+ "llava": {
283
+ "image_aspect_ratio": "original"
284
+ }
285
+ }
286
+ },
287
+ "mmstar": {
288
+ "task": "mmstar",
289
+ "dataset_path": "Lin-Chen/MMStar",
290
+ "dataset_kwargs": {
291
+ "token": true
292
+ },
293
+ "test_split": "val",
294
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7fe3bc3838b0>",
295
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7fe3bc383d30>",
296
+ "doc_to_target": "answer",
297
+ "process_results": "<function mmstar_process_results at 0x7fe3bc389280>",
298
+ "description": "",
299
+ "target_delimiter": " ",
300
+ "fewshot_delimiter": "\n\n",
301
+ "metric_list": [
302
+ {
303
+ "metric": "coarse perception",
304
+ "aggregation": "<function mmstar_aggregate_results at 0x7fe3bc389700>",
305
+ "higher_is_better": true
306
+ },
307
+ {
308
+ "metric": "fine-grained perception",
309
+ "aggregation": "<function mmstar_aggregate_results at 0x7fe3bc389af0>",
310
+ "higher_is_better": true
311
+ },
312
+ {
313
+ "metric": "instance reasoning",
314
+ "aggregation": "<function mmstar_aggregate_results at 0x7fe3bc389ee0>",
315
+ "higher_is_better": true
316
+ },
317
+ {
318
+ "metric": "logical reasoning",
319
+ "aggregation": "<function mmstar_aggregate_results at 0x7fe3bc38f310>",
320
+ "higher_is_better": true
321
+ },
322
+ {
323
+ "metric": "science & technology",
324
+ "aggregation": "<function mmstar_aggregate_results at 0x7fe3bc38f700>",
325
+ "higher_is_better": true
326
+ },
327
+ {
328
+ "metric": "math",
329
+ "aggregation": "<function mmstar_aggregate_results at 0x7fe3bc38faf0>",
330
+ "higher_is_better": true
331
+ }
332
+ ],
333
+ "output_type": "generate_until",
334
+ "generation_kwargs": {
335
+ "until": [
336
+ "\n\n"
337
+ ],
338
+ "do_sample": false
339
+ },
340
+ "repeats": 1,
341
+ "should_decontaminate": false,
342
+ "metadata": [
343
+ {
344
+ "version": 0.0
345
+ }
346
+ ],
347
+ "model_specific_prompt_kwargs": {
348
+ "default": {
349
+ "pre_prompt": "",
350
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
351
+ }
352
+ }
353
+ },
354
+ "pope": {
355
+ "task": "pope",
356
+ "dataset_path": "lmms-lab/POPE",
357
+ "dataset_kwargs": {
358
+ "token": true
359
+ },
360
+ "test_split": "test",
361
+ "doc_to_visual": "<function pope_doc_to_visual at 0x7fe3bcc67ee0>",
362
+ "doc_to_text": "<function pope_doc_to_text at 0x7fe3bcc6b4c0>",
363
+ "doc_to_target": "answer",
364
+ "process_results": "<function pope_process_results at 0x7fe3bcc6ba60>",
365
+ "description": "",
366
+ "target_delimiter": " ",
367
+ "fewshot_delimiter": "\n\n",
368
+ "metric_list": [
369
+ {
370
+ "metric": "pope_accuracy",
371
+ "aggregation": "<function pope_aggregate_accuracy at 0x7fe3bcc6e040>",
372
+ "higher_is_better": true
373
+ },
374
+ {
375
+ "metric": "pope_precision",
376
+ "aggregation": "<function pope_aggregate_precision at 0x7fe3bcc6e5e0>",
377
+ "higher_is_better": true
378
+ },
379
+ {
380
+ "metric": "pope_recall",
381
+ "aggregation": "<function pope_aggregate_recall at 0x7fe3bcc6eb80>",
382
+ "higher_is_better": true
383
+ },
384
+ {
385
+ "metric": "pope_f1_score",
386
+ "aggregation": "<function pope_aggregate_f1_score at 0x7fe3bcc73160>",
387
+ "higher_is_better": true
388
+ },
389
+ {
390
+ "metric": "pope_yes_ratio",
391
+ "aggregation": "<function pope_aggregate_yes_ratio at 0x7fe3bcc73700>",
392
+ "higher_is_better": true
393
+ }
394
+ ],
395
+ "output_type": "generate_until",
396
+ "generation_kwargs": {
397
+ "max_new_tokens": 128,
398
+ "temperature": 0.0,
399
+ "top_p": 1.0,
400
+ "num_beams": 1,
401
+ "do_sample": false,
402
+ "until": [
403
+ "\n\n"
404
+ ]
405
+ },
406
+ "repeats": 1,
407
+ "should_decontaminate": false,
408
+ "metadata": [
409
+ {
410
+ "version": 0.0
411
+ }
412
+ ]
413
+ },
414
+ "textvqa_val": {
415
+ "task": "textvqa_val",
416
+ "dataset_path": "lmms-lab/textvqa",
417
+ "test_split": "validation",
418
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7fe3e8d06ee0>",
419
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7fe3e8c143a0>",
420
+ "doc_to_target": "answer",
421
+ "process_results": "<function textvqa_process_results at 0x7fe3e8c14670>",
422
+ "description": "",
423
+ "target_delimiter": " ",
424
+ "fewshot_delimiter": "\n\n",
425
+ "metric_list": [
426
+ {
427
+ "metric": "exact_match",
428
+ "aggregation": "mean",
429
+ "higher_is_better": true,
430
+ "ignore_case": true,
431
+ "ignore_punctuation": true
432
+ },
433
+ {
434
+ "metric": "submission",
435
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7fe3e8d06ca0>",
436
+ "higher_is_better": true
437
+ }
438
+ ],
439
+ "output_type": "generate_until",
440
+ "generation_kwargs": {
441
+ "until": [
442
+ "ASSISTANT:"
443
+ ]
444
+ },
445
+ "repeats": 1,
446
+ "should_decontaminate": false,
447
+ "model_specific_prompt_kwargs": {
448
+ "default": {
449
+ "pre_prompt": "",
450
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
451
+ "ocr": false
452
+ },
453
+ "qwen_vl": {
454
+ "pre_prompt": "",
455
+ "post_prompt": " Answer:"
456
+ }
457
+ }
458
+ }
459
+ },
460
+ "versions": {
461
+ "ai2d": "Yaml",
462
+ "mmbench_en_dev": "Yaml",
463
+ "mmerealworld_lite": "Yaml",
464
+ "mmmu_val": "Yaml",
465
+ "mmstar": "Yaml",
466
+ "pope": "Yaml",
467
+ "textvqa_val": "Yaml"
468
+ },
469
+ "n-shot": {
470
+ "ai2d": 0,
471
+ "mmbench_en_dev": 0,
472
+ "mmerealworld_lite": 0,
473
+ "mmmu_val": 0,
474
+ "mmstar": 0,
475
+ "pope": 0,
476
+ "textvqa_val": 0
477
+ },
478
+ "model_configs": {
479
+ "model": "llava",
480
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/std_0.002_Full_smoe,conv_template=phi35",
481
+ "batch_size": "1",
482
+ "device": null,
483
+ "limit": null,
484
+ "bootstrap_iters": 100000,
485
+ "gen_kwargs": ""
486
+ },
487
+ "git_hash": "289c7fe5"
488
+ }
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.7079037800687286, "category_acc": {"action_recognition": 0.8703703703703703, "attribute_comparison": 0.6590909090909091, "attribute_recognition": 0.9054054054054054, "celebrity_recognition": 0.7777777777777778, "function_reasoning": 0.8734177215189873, "future_prediction": 0.575, "identity_reasoning": 0.9777777777777777, "image_emotion": 0.82, "image_quality": 0.49056603773584906, "image_scene": 0.9519230769230769, "image_style": 0.8867924528301887, "image_topic": 0.8888888888888888, "nature_relation": 0.5833333333333334, "object_localization": 0.4074074074074074, "ocr": 0.6410256410256411, "physical_property_reasoning": 0.44, "physical_relation": 0.5833333333333334, "social_relation": 0.8372093023255814, "spatial_relationship": 0.2, "structuralized_imagetext_understanding": 0.5769230769230769}, "l2_category_acc": {"attribute_reasoning": 0.7336683417085427, "coarse_perception": 0.8277027027027027, "finegrained_perception (cross-instance)": 0.5944055944055944, "finegrained_perception (instance-level)": 0.689419795221843, "logic_reasoning": 0.576271186440678, "relation_reasoning": 0.6782608695652174}}
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:613d6406462bf5f7cdaa2cc19733a452f30570b8bce56bb4d51835a247d142c1
3
+ size 843251
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/submissions/textvqa_submission_2025-06-13-02-34-59.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/std_0.002_Full_smoe/logs/0613_0312_llava..._pope_llava_model_args_35cab6/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6be333dbcdf51b13d860f8881b1106e78c4fa9d3537a0f72b1a7adf5639b15d7
3
+ size 13146060
sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/gqa.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c05f10a5e4ff3162c9290e5b2ad59725b82393898254c5599012ad1a2fc24946
3
+ size 38379389
sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0314_llava_v1.5_gqa_llava_model_args_35cab6/results.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "gqa": {
4
+ "exact_match,none": 0.4698680235331531,
5
+ "exact_match_stderr,none": 0.004450321794749798,
6
+ "alias": "gqa"
7
+ }
8
+ },
9
+ "configs": {
10
+ "gqa": {
11
+ "task": "gqa",
12
+ "dataset_path": "lmms-lab/GQA",
13
+ "dataset_name": "testdev_balanced_instructions",
14
+ "dataset_kwargs": {
15
+ "token": true
16
+ },
17
+ "test_split": "testdev",
18
+ "doc_to_visual": "<function gqa_doc_to_visual at 0x7fdf7085ef70>",
19
+ "doc_to_text": "<function gqa_doc_to_text at 0x7fdf70737280>",
20
+ "doc_to_target": "answer",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "exact_match",
27
+ "aggregation": "mean",
28
+ "higher_is_better": true,
29
+ "ignore_case": true,
30
+ "ignore_punctuation": true
31
+ }
32
+ ],
33
+ "output_type": "generate_until",
34
+ "generation_kwargs": {
35
+ "max_new_tokens": 16,
36
+ "temperature": 0.0,
37
+ "top_p": 1.0,
38
+ "num_beams": 1,
39
+ "do_sample": false,
40
+ "until": [
41
+ "\n\n"
42
+ ]
43
+ },
44
+ "repeats": 1,
45
+ "should_decontaminate": false,
46
+ "metadata": [
47
+ {
48
+ "version": 0.0
49
+ }
50
+ ],
51
+ "model_specific_prompt_kwargs": {
52
+ "default": {
53
+ "pre_prompt": "",
54
+ "post_prompt": "\nAnswer the question using a single word or phrase."
55
+ },
56
+ "qwen_vl": {
57
+ "pre_prompt": "",
58
+ "post_prompt": " Answer:"
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "versions": {
64
+ "gqa": "Yaml"
65
+ },
66
+ "n-shot": {
67
+ "gqa": 0
68
+ },
69
+ "model_configs": {
70
+ "model": "llava",
71
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/std_0.002_Full_smoe,conv_template=phi35",
72
+ "batch_size": "1",
73
+ "device": null,
74
+ "limit": null,
75
+ "bootstrap_iters": 100000,
76
+ "gen_kwargs": ""
77
+ },
78
+ "git_hash": "289c7fe5"
79
+ }
sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/ocrbench.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/results.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ocrbench": {
4
+ "ocrbench_accuracy,none": 0.402,
5
+ "ocrbench_accuracy_stderr,none": "N/A",
6
+ "alias": "ocrbench"
7
+ }
8
+ },
9
+ "configs": {
10
+ "ocrbench": {
11
+ "task": "ocrbench",
12
+ "dataset_path": "echo840/OCRBench",
13
+ "dataset_kwargs": {
14
+ "token": true
15
+ },
16
+ "test_split": "test",
17
+ "doc_to_visual": "<function ocrbench_doc_to_visual at 0x7faafdac61f0>",
18
+ "doc_to_text": "<function ocrbench_doc_to_text at 0x7faafd4a2790>",
19
+ "doc_to_target": "answer",
20
+ "process_results": "<function ocrbench_process_results at 0x7faafd4a2af0>",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "ocrbench_accuracy",
27
+ "aggregation": "<function ocrbench_aggregate_accuracy at 0x7faafd4a2e50>",
28
+ "higher_is_better": true
29
+ }
30
+ ],
31
+ "output_type": "generate_until",
32
+ "generation_kwargs": {
33
+ "max_new_tokens": 128,
34
+ "temperature": 0.0,
35
+ "top_p": 1.0,
36
+ "num_beams": 1,
37
+ "do_sample": false,
38
+ "until": [
39
+ "\n\n"
40
+ ]
41
+ },
42
+ "repeats": 1,
43
+ "should_decontaminate": false,
44
+ "metadata": [
45
+ {
46
+ "version": 0.0
47
+ }
48
+ ]
49
+ }
50
+ },
51
+ "versions": {
52
+ "ocrbench": "Yaml"
53
+ },
54
+ "n-shot": {
55
+ "ocrbench": 0
56
+ },
57
+ "model_configs": {
58
+ "model": "llava",
59
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/std_0.002_Full_smoe,conv_template=phi35",
60
+ "batch_size": "1",
61
+ "device": null,
62
+ "limit": null,
63
+ "bootstrap_iters": 100000,
64
+ "gen_kwargs": ""
65
+ },
66
+ "git_hash": "289c7fe5"
67
+ }
sft/1M3/std_0.002_Full_smoe/logs/0613_0336_llava...bench_llava_model_args_35cab6/results/ocrbench_results.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ######################### OCRBench #############################
2
+ Text Recognition(Total 300): 215
3
+ ---------------- Details of Recognition Score ------------------
4
+ Regular Text Recognition(Total 50): 47
5
+ Irregular Text Recognition(Total 50): 44
6
+ Artistic Text Recognition(Total 50): 43
7
+ Handwriting Recognition(Total 50): 20
8
+ Digit String Recognition(Total 50): 36
9
+ Non-Semantic Text Recognition(Total 50): 25
10
+ ----------------------------------------------------------------
11
+ Scene Text-centric VQA(Total 200): 112
12
+ ----------------------------------------------------------------
13
+ Doc-oriented VQA(Total 200): 22
14
+ ----------------------------------------------------------------
15
+ Key Information Extraction(Total 200): 18
16
+ Handwritten Mathematical Expression Recognition(Total 100): 35
17
+ --------------------- Final Score ------------------------------
18
+ Final Score(Total 1000): 402