DavidNguyen commited on
Commit
01ca196
·
verified ·
1 Parent(s): d808cc9

Upload folder using huggingface_hub

Browse files
Files changed (30) hide show
  1. .gitattributes +6 -0
  2. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/ai2d.json +0 -0
  3. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmbench_en_dev.json +3 -0
  4. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmmu_val.json +0 -0
  5. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmstar.json +0 -0
  6. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank0_metric_eval_done.txt +1 -0
  7. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank1_metric_eval_done.txt +1 -0
  8. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank2_metric_eval_done.txt +1 -0
  9. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank3_metric_eval_done.txt +1 -0
  10. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/results.json +298 -0
  11. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/submissions/mmbench_en_dev_results.json +1 -0
  12. sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/submissions/mmbench_en_dev_results.xlsx +3 -0
  13. sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/gqa.json +3 -0
  14. sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/rank0_metric_eval_done.txt +1 -0
  15. sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/results.json +79 -0
  16. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/mmerealworld_lite.json +3 -0
  17. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/ocrbench.json +0 -0
  18. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/pope.json +3 -0
  19. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/rank0_metric_eval_done.txt +1 -0
  20. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/rank1_metric_eval_done.txt +1 -0
  21. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/rank2_metric_eval_done.txt +1 -0
  22. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/results.json +257 -0
  23. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/results/ocrbench_results.txt +18 -0
  24. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/submissions/textvqa_submission_2025-04-12-18-52-07.json +0 -0
  25. sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/textvqa_val.json +3 -0
  26. sft/smoe_refine_665k_llava/model-00001-of-00003.safetensors +1 -1
  27. sft/smoe_refine_665k_llava/model-00002-of-00003.safetensors +1 -1
  28. sft/smoe_refine_665k_llava/model-00003-of-00003.safetensors +1 -1
  29. sft/smoe_refine_665k_llava/trainer_state.json +0 -0
  30. sft/smoe_refine_665k_llava/training_args.bin +2 -2
.gitattributes CHANGED
@@ -40,3 +40,9 @@ sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_m
40
  sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
41
  sft/smoe_refine_665k_llava/checkpoint-3328/logs/0409_1820_llava...mstar_llava_model_args_b94fe7/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
42
  sft/smoe_refine_665k_llava/checkpoint-3328/logs/0409_1820_llava...mstar_llava_model_args_b94fe7/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
40
  sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
41
  sft/smoe_refine_665k_llava/checkpoint-3328/logs/0409_1820_llava...mstar_llava_model_args_b94fe7/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
42
  sft/smoe_refine_665k_llava/checkpoint-3328/logs/0409_1820_llava...mstar_llava_model_args_b94fe7/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
43
+ sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
44
+ sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
45
+ sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/gqa.json filter=lfs diff=lfs merge=lfs -text
46
+ sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
47
+ sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/pope.json filter=lfs diff=lfs merge=lfs -text
48
+ sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:269f917d9576de83390b3278fb637ffe30f93762af58fb582620e85c50b24fd2
3
+ size 14568181
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6496113989637305,
5
+ "exact_match_stderr,flexible-extract": 0.008586842325753156,
6
+ "alias": "ai2d"
7
+ },
8
+ "mmbench_en_dev": {
9
+ "gpt_eval_score,none": 71.56357388316151,
10
+ "gpt_eval_score_stderr,none": "N/A",
11
+ "submission,none": null,
12
+ "submission_stderr,none": "N/A",
13
+ "alias": "mmbench_en_dev"
14
+ },
15
+ "mmmu_val": {
16
+ "mmmu_acc,none": 0.41889,
17
+ "mmmu_acc_stderr,none": "N/A",
18
+ "alias": "mmmu_val"
19
+ },
20
+ "mmstar": {
21
+ "coarse perception,none": 0.7266511799799377,
22
+ "coarse perception_stderr,none": "N/A",
23
+ "fine-grained perception,none": 0.3833578973225252,
24
+ "fine-grained perception_stderr,none": "N/A",
25
+ "instance reasoning,none": 0.5216438884434536,
26
+ "instance reasoning_stderr,none": "N/A",
27
+ "logical reasoning,none": 0.3982429012131982,
28
+ "logical reasoning_stderr,none": "N/A",
29
+ "math,none": 0.29996435890581835,
30
+ "math_stderr,none": "N/A",
31
+ "science & technology,none": 0.24473379748481924,
32
+ "science & technology_stderr,none": "N/A",
33
+ "alias": "mmstar"
34
+ }
35
+ },
36
+ "configs": {
37
+ "ai2d": {
38
+ "task": "ai2d",
39
+ "dataset_path": "lmms-lab/ai2d",
40
+ "dataset_kwargs": {
41
+ "token": true
42
+ },
43
+ "test_split": "test",
44
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7f573be26af0>",
45
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7f573be26dc0>",
46
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7f573be2b280>",
47
+ "description": "",
48
+ "target_delimiter": " ",
49
+ "fewshot_delimiter": "\n\n",
50
+ "metric_list": [
51
+ {
52
+ "metric": "exact_match",
53
+ "aggregation": "mean",
54
+ "higher_is_better": true,
55
+ "ignore_case": true,
56
+ "ignore_punctuation": true
57
+ }
58
+ ],
59
+ "output_type": "generate_until",
60
+ "generation_kwargs": {
61
+ "max_new_tokens": 512,
62
+ "temperature": 0.0,
63
+ "do_sample": false,
64
+ "until": [
65
+ "\n\n"
66
+ ]
67
+ },
68
+ "repeats": 1,
69
+ "filter_list": [
70
+ {
71
+ "name": "flexible-extract",
72
+ "filter": [
73
+ {
74
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
75
+ "group_select": 0,
76
+ "ignore_case": true,
77
+ "ignore_punctuation": true,
78
+ "regex_pattern": "([A-Z])\\."
79
+ }
80
+ ]
81
+ }
82
+ ],
83
+ "should_decontaminate": false,
84
+ "metadata": [
85
+ {
86
+ "version": 0.0
87
+ }
88
+ ],
89
+ "model_specific_prompt_kwargs": {
90
+ "default": {
91
+ "prompt_format": "mcq",
92
+ "pre_prompt": "",
93
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
94
+ },
95
+ "gpt4v": {
96
+ "prompt_format": "mcq",
97
+ "pre_prompt": "",
98
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
99
+ },
100
+ "qwen_vl": {
101
+ "prompt_format": "qa",
102
+ "pre_prompt": "",
103
+ "post_prompt": " Answer:"
104
+ },
105
+ "xcomposer2_4khd": {
106
+ "prompt_format": "mcq_xcomposer",
107
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
108
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
109
+ }
110
+ },
111
+ "model_specific_target_kwargs": {
112
+ "default": "mcq",
113
+ "qwen_vl": "qa"
114
+ }
115
+ },
116
+ "mmbench_en_dev": {
117
+ "task": "mmbench_en_dev",
118
+ "dataset_path": "lmms-lab/MMBench",
119
+ "dataset_name": "en",
120
+ "dataset_kwargs": {
121
+ "token": true
122
+ },
123
+ "test_split": "dev",
124
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7f579c2d2af0>",
125
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7f579c25d0d0>",
126
+ "doc_to_target": "answer",
127
+ "process_results": "<function mmbench_process_results at 0x7f579c25d670>",
128
+ "description": "",
129
+ "target_delimiter": " ",
130
+ "fewshot_delimiter": "\n\n",
131
+ "metric_list": [
132
+ {
133
+ "metric": "gpt_eval_score",
134
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7f579c2d2280>",
135
+ "higher_is_better": true
136
+ },
137
+ {
138
+ "metric": "submission",
139
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7f579c2d2820>",
140
+ "higher_is_better": true
141
+ }
142
+ ],
143
+ "output_type": "generate_until",
144
+ "generation_kwargs": {
145
+ "until": [
146
+ "ASSISTANT:"
147
+ ],
148
+ "max_new_tokens": 1024,
149
+ "temperature": 0.0,
150
+ "top_p": 1.0,
151
+ "num_beams": 1,
152
+ "do_sample": false,
153
+ "image_aspect_ratio": "original"
154
+ },
155
+ "repeats": 1,
156
+ "should_decontaminate": false,
157
+ "model_specific_prompt_kwargs": {
158
+ "default": {
159
+ "pre_prompt": "",
160
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
161
+ }
162
+ },
163
+ "model_specific_generation_kwargs": {
164
+ "llava": {
165
+ "image_aspect_ratio": "original"
166
+ }
167
+ }
168
+ },
169
+ "mmmu_val": {
170
+ "task": "mmmu_val",
171
+ "dataset_path": "lmms-lab/MMMU",
172
+ "test_split": "validation",
173
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7f574d84dc10>",
174
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7f5773c2f8b0>",
175
+ "doc_to_target": "answer",
176
+ "process_results": "<function mmmu_process_results at 0x7f5773c45700>",
177
+ "description": "",
178
+ "target_delimiter": " ",
179
+ "fewshot_delimiter": "\n\n",
180
+ "metric_list": [
181
+ {
182
+ "metric": "mmmu_acc",
183
+ "aggregation": "<function mmmu_aggregate_results at 0x7f57739225e0>",
184
+ "higher_is_better": true
185
+ }
186
+ ],
187
+ "output_type": "generate_until",
188
+ "generation_kwargs": {
189
+ "max_new_tokens": 128,
190
+ "until": [
191
+ "\n\n"
192
+ ],
193
+ "image_aspect_ratio": "original"
194
+ },
195
+ "repeats": 1,
196
+ "should_decontaminate": false,
197
+ "metadata": [
198
+ {
199
+ "version": 0.0
200
+ }
201
+ ],
202
+ "model_specific_generation_kwargs": {
203
+ "llava": {
204
+ "image_aspect_ratio": "original"
205
+ }
206
+ }
207
+ },
208
+ "mmstar": {
209
+ "task": "mmstar",
210
+ "dataset_path": "Lin-Chen/MMStar",
211
+ "dataset_kwargs": {
212
+ "token": true
213
+ },
214
+ "test_split": "val",
215
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7f573bd7dca0>",
216
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7f573bd880d0>",
217
+ "doc_to_target": "answer",
218
+ "process_results": "<function mmstar_process_results at 0x7f573bd88550>",
219
+ "description": "",
220
+ "target_delimiter": " ",
221
+ "fewshot_delimiter": "\n\n",
222
+ "metric_list": [
223
+ {
224
+ "metric": "coarse perception",
225
+ "aggregation": "<function mmstar_aggregate_results at 0x7f573bd88940>",
226
+ "higher_is_better": true
227
+ },
228
+ {
229
+ "metric": "fine-grained perception",
230
+ "aggregation": "<function mmstar_aggregate_results at 0x7f573bd88ca0>",
231
+ "higher_is_better": true
232
+ },
233
+ {
234
+ "metric": "instance reasoning",
235
+ "aggregation": "<function mmstar_aggregate_results at 0x7f573bd8e040>",
236
+ "higher_is_better": true
237
+ },
238
+ {
239
+ "metric": "logical reasoning",
240
+ "aggregation": "<function mmstar_aggregate_results at 0x7f573bd8e3a0>",
241
+ "higher_is_better": true
242
+ },
243
+ {
244
+ "metric": "science & technology",
245
+ "aggregation": "<function mmstar_aggregate_results at 0x7f573bd8e700>",
246
+ "higher_is_better": true
247
+ },
248
+ {
249
+ "metric": "math",
250
+ "aggregation": "<function mmstar_aggregate_results at 0x7f573bd8ea60>",
251
+ "higher_is_better": true
252
+ }
253
+ ],
254
+ "output_type": "generate_until",
255
+ "generation_kwargs": {
256
+ "until": [
257
+ "\n\n"
258
+ ],
259
+ "do_sample": false
260
+ },
261
+ "repeats": 1,
262
+ "should_decontaminate": false,
263
+ "metadata": [
264
+ {
265
+ "version": 0.0
266
+ }
267
+ ],
268
+ "model_specific_prompt_kwargs": {
269
+ "default": {
270
+ "pre_prompt": "",
271
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
272
+ }
273
+ }
274
+ }
275
+ },
276
+ "versions": {
277
+ "ai2d": "Yaml",
278
+ "mmbench_en_dev": "Yaml",
279
+ "mmmu_val": "Yaml",
280
+ "mmstar": "Yaml"
281
+ },
282
+ "n-shot": {
283
+ "ai2d": 0,
284
+ "mmbench_en_dev": 0,
285
+ "mmmu_val": 0,
286
+ "mmstar": 0
287
+ },
288
+ "model_configs": {
289
+ "model": "llava",
290
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/sft/SMOE/test_Full_smoe,conv_template=phi35",
291
+ "batch_size": "1",
292
+ "device": null,
293
+ "limit": null,
294
+ "bootstrap_iters": 100000,
295
+ "gen_kwargs": ""
296
+ },
297
+ "git_hash": "b61cb97"
298
+ }
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.7156357388316151, "category_acc": {"action_recognition": 0.8888888888888888, "attribute_comparison": 0.75, "attribute_recognition": 0.8783783783783784, "celebrity_recognition": 0.8080808080808081, "function_reasoning": 0.8734177215189873, "future_prediction": 0.475, "identity_reasoning": 0.9777777777777777, "image_emotion": 0.84, "image_quality": 0.5283018867924528, "image_scene": 0.9519230769230769, "image_style": 0.8113207547169812, "image_topic": 0.8888888888888888, "nature_relation": 0.6875, "object_localization": 0.48148148148148145, "ocr": 0.6410256410256411, "physical_property_reasoning": 0.5866666666666667, "physical_relation": 0.4583333333333333, "social_relation": 0.7674418604651163, "spatial_relationship": 0.3111111111111111, "structuralized_imagetext_understanding": 0.41025641025641024}, "l2_category_acc": {"attribute_reasoning": 0.7889447236180904, "coarse_perception": 0.8243243243243243, "finegrained_perception (cross-instance)": 0.6643356643356644, "finegrained_perception (instance-level)": 0.7133105802047781, "logic_reasoning": 0.4322033898305085, "relation_reasoning": 0.6695652173913044}}
sft/smoe_refine_665k_llava/logs/ai2d,mmbench_en_dev,mmmu_val,mmstar/0412_1911_llava...mstar_llava_model_args_4172a9/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b0a389ca00b157517d41368cff5f7ddefb68e34ec2d717f37f55baf31dab422
3
+ size 865153
sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/gqa.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b3eb93ad63ce6a17250d269c1dea1f66b6bedbd6848e4c57bb8501c97901928
3
+ size 38369383
sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/smoe_refine_665k_llava/logs/gqa/0412_1934_llava_v1.5_gqa_llava_model_args_4172a9/results.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "gqa": {
4
+ "exact_match,none": 0.6148831292733344,
5
+ "exact_match_stderr,none": 0.004339143672998995,
6
+ "alias": "gqa"
7
+ }
8
+ },
9
+ "configs": {
10
+ "gqa": {
11
+ "task": "gqa",
12
+ "dataset_path": "lmms-lab/GQA",
13
+ "dataset_name": "testdev_balanced_instructions",
14
+ "dataset_kwargs": {
15
+ "token": true
16
+ },
17
+ "test_split": "testdev",
18
+ "doc_to_visual": "<function gqa_doc_to_visual at 0x7f337a112820>",
19
+ "doc_to_text": "<function gqa_doc_to_text at 0x7f337a063940>",
20
+ "doc_to_target": "answer",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "exact_match",
27
+ "aggregation": "mean",
28
+ "higher_is_better": true,
29
+ "ignore_case": true,
30
+ "ignore_punctuation": true
31
+ }
32
+ ],
33
+ "output_type": "generate_until",
34
+ "generation_kwargs": {
35
+ "max_new_tokens": 16,
36
+ "temperature": 0.0,
37
+ "top_p": 1.0,
38
+ "num_beams": 1,
39
+ "do_sample": false,
40
+ "until": [
41
+ "\n\n"
42
+ ]
43
+ },
44
+ "repeats": 1,
45
+ "should_decontaminate": false,
46
+ "metadata": [
47
+ {
48
+ "version": 0.0
49
+ }
50
+ ],
51
+ "model_specific_prompt_kwargs": {
52
+ "default": {
53
+ "pre_prompt": "",
54
+ "post_prompt": "\nAnswer the question using a single word or phrase."
55
+ },
56
+ "qwen_vl": {
57
+ "pre_prompt": "",
58
+ "post_prompt": " Answer:"
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "versions": {
64
+ "gqa": "Yaml"
65
+ },
66
+ "n-shot": {
67
+ "gqa": 0
68
+ },
69
+ "model_configs": {
70
+ "model": "llava",
71
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/sft/SMOE/test_Full_smoe,conv_template=phi35",
72
+ "batch_size": "1",
73
+ "device": null,
74
+ "limit": null,
75
+ "bootstrap_iters": 100000,
76
+ "gen_kwargs": ""
77
+ },
78
+ "git_hash": "b61cb97"
79
+ }
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43d09dfabbd10426ed87e786b93da8fb445467154d1d8d354f9058a98445489f
3
+ size 1994104230
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/ocrbench.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/pope.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04c67b7ca3ab3e41cc17c5f60c89d33c096a51fd55963a12f578a671cc905da5
3
+ size 17577382
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/results.json ADDED
@@ -0,0 +1,257 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "mmerealworld_lite": {
4
+ "mme_realworld_score,none": 0.32204273058884836,
5
+ "mme_realworld_score_stderr,none": "N/A",
6
+ "alias": "mmerealworld_lite"
7
+ },
8
+ "ocrbench": {
9
+ "ocrbench_accuracy,none": 0.323,
10
+ "ocrbench_accuracy_stderr,none": "N/A",
11
+ "alias": "ocrbench"
12
+ },
13
+ "pope": {
14
+ "pope_accuracy,none": 0.8651111111111112,
15
+ "pope_accuracy_stderr,none": "N/A",
16
+ "pope_precision,none": 0.9292058516196448,
17
+ "pope_precision_stderr,none": "N/A",
18
+ "pope_recall,none": 0.7904444444444444,
19
+ "pope_recall_stderr,none": "N/A",
20
+ "pope_f1_score,none": 0.8542267050912584,
21
+ "pope_f1_score_stderr,none": "N/A",
22
+ "pope_yes_ratio,none": 0.5,
23
+ "pope_yes_ratio_stderr,none": "N/A",
24
+ "alias": "pope"
25
+ },
26
+ "textvqa_val": {
27
+ "exact_match,none": 0.4165600003123283,
28
+ "exact_match_stderr,none": 0.006747433492938556,
29
+ "submission,none": null,
30
+ "submission_stderr,none": "N/A",
31
+ "alias": "textvqa_val"
32
+ }
33
+ },
34
+ "configs": {
35
+ "mmerealworld_lite": {
36
+ "task": "mmerealworld_lite",
37
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
38
+ "dataset_kwargs": {
39
+ "token": true
40
+ },
41
+ "test_split": "train",
42
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7fa81e9ceaf0>",
43
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7fa81e9d1700>",
44
+ "doc_to_target": "answer",
45
+ "process_results": "<function mme_realworld_process_results at 0x7fa81e9d74c0>",
46
+ "description": "",
47
+ "target_delimiter": " ",
48
+ "fewshot_delimiter": "\n\n",
49
+ "metric_list": [
50
+ {
51
+ "metric": "mme_realworld_score",
52
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7fa81e9da160>",
53
+ "higher_is_better": true
54
+ }
55
+ ],
56
+ "output_type": "generate_until",
57
+ "generation_kwargs": {
58
+ "max_new_tokens": 16,
59
+ "temperature": 0.0,
60
+ "top_p": 1.0,
61
+ "num_beams": 1,
62
+ "do_sample": false,
63
+ "until": [
64
+ "\n\n"
65
+ ]
66
+ },
67
+ "repeats": 1,
68
+ "should_decontaminate": false,
69
+ "metadata": [
70
+ {
71
+ "version": 0.0
72
+ }
73
+ ],
74
+ "model_specific_prompt_kwargs": {
75
+ "default": {
76
+ "pre_prompt": "",
77
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
78
+ },
79
+ "gpt4v": {
80
+ "pre_prompt": "",
81
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
82
+ },
83
+ "xcomposer2_4khd": {
84
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
85
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
86
+ }
87
+ }
88
+ },
89
+ "ocrbench": {
90
+ "task": "ocrbench",
91
+ "dataset_path": "echo840/OCRBench",
92
+ "dataset_kwargs": {
93
+ "token": true
94
+ },
95
+ "test_split": "test",
96
+ "doc_to_visual": "<function ocrbench_doc_to_visual at 0x7fa8337ce0d0>",
97
+ "doc_to_text": "<function ocrbench_doc_to_text at 0x7fa833117700>",
98
+ "doc_to_target": "answer",
99
+ "process_results": "<function ocrbench_process_results at 0x7fa833117a60>",
100
+ "description": "",
101
+ "target_delimiter": " ",
102
+ "fewshot_delimiter": "\n\n",
103
+ "metric_list": [
104
+ {
105
+ "metric": "ocrbench_accuracy",
106
+ "aggregation": "<function ocrbench_aggregate_accuracy at 0x7fa833117dc0>",
107
+ "higher_is_better": true
108
+ }
109
+ ],
110
+ "output_type": "generate_until",
111
+ "generation_kwargs": {
112
+ "max_new_tokens": 128,
113
+ "temperature": 0.0,
114
+ "top_p": 1.0,
115
+ "num_beams": 1,
116
+ "do_sample": false,
117
+ "until": [
118
+ "\n\n"
119
+ ]
120
+ },
121
+ "repeats": 1,
122
+ "should_decontaminate": false,
123
+ "metadata": [
124
+ {
125
+ "version": 0.0
126
+ }
127
+ ]
128
+ },
129
+ "pope": {
130
+ "task": "pope",
131
+ "dataset_path": "lmms-lab/POPE",
132
+ "dataset_kwargs": {
133
+ "token": true
134
+ },
135
+ "test_split": "test",
136
+ "doc_to_visual": "<function pope_doc_to_visual at 0x7fa81ea61e50>",
137
+ "doc_to_text": "<function pope_doc_to_text at 0x7fa81ea683a0>",
138
+ "doc_to_target": "answer",
139
+ "process_results": "<function pope_process_results at 0x7fa81ea688b0>",
140
+ "description": "",
141
+ "target_delimiter": " ",
142
+ "fewshot_delimiter": "\n\n",
143
+ "metric_list": [
144
+ {
145
+ "metric": "pope_accuracy",
146
+ "aggregation": "<function pope_aggregate_accuracy at 0x7fa81ea68dc0>",
147
+ "higher_is_better": true
148
+ },
149
+ {
150
+ "metric": "pope_precision",
151
+ "aggregation": "<function pope_aggregate_precision at 0x7fa81ea69310>",
152
+ "higher_is_better": true
153
+ },
154
+ {
155
+ "metric": "pope_recall",
156
+ "aggregation": "<function pope_aggregate_recall at 0x7fa81ea69820>",
157
+ "higher_is_better": true
158
+ },
159
+ {
160
+ "metric": "pope_f1_score",
161
+ "aggregation": "<function pope_aggregate_f1_score at 0x7fa81ea69d30>",
162
+ "higher_is_better": true
163
+ },
164
+ {
165
+ "metric": "pope_yes_ratio",
166
+ "aggregation": "<function pope_aggregate_yes_ratio at 0x7fa81ea71280>",
167
+ "higher_is_better": true
168
+ }
169
+ ],
170
+ "output_type": "generate_until",
171
+ "generation_kwargs": {
172
+ "max_new_tokens": 128,
173
+ "temperature": 0.0,
174
+ "top_p": 1.0,
175
+ "num_beams": 1,
176
+ "do_sample": false,
177
+ "until": [
178
+ "\n\n"
179
+ ]
180
+ },
181
+ "repeats": 1,
182
+ "should_decontaminate": false,
183
+ "metadata": [
184
+ {
185
+ "version": 0.0
186
+ }
187
+ ]
188
+ },
189
+ "textvqa_val": {
190
+ "task": "textvqa_val",
191
+ "dataset_path": "lmms-lab/textvqa",
192
+ "test_split": "validation",
193
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7fa84a8f41f0>",
194
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7fa84a8f4670>",
195
+ "doc_to_target": "answer",
196
+ "process_results": "<function textvqa_process_results at 0x7fa84a8f4940>",
197
+ "description": "",
198
+ "target_delimiter": " ",
199
+ "fewshot_delimiter": "\n\n",
200
+ "metric_list": [
201
+ {
202
+ "metric": "exact_match",
203
+ "aggregation": "mean",
204
+ "higher_is_better": true,
205
+ "ignore_case": true,
206
+ "ignore_punctuation": true
207
+ },
208
+ {
209
+ "metric": "submission",
210
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7fa84a8f4040>",
211
+ "higher_is_better": true
212
+ }
213
+ ],
214
+ "output_type": "generate_until",
215
+ "generation_kwargs": {
216
+ "until": [
217
+ "ASSISTANT:"
218
+ ]
219
+ },
220
+ "repeats": 1,
221
+ "should_decontaminate": false,
222
+ "model_specific_prompt_kwargs": {
223
+ "default": {
224
+ "pre_prompt": "",
225
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
226
+ "ocr": false
227
+ },
228
+ "qwen_vl": {
229
+ "pre_prompt": "",
230
+ "post_prompt": " Answer:"
231
+ }
232
+ }
233
+ }
234
+ },
235
+ "versions": {
236
+ "mmerealworld_lite": "Yaml",
237
+ "ocrbench": "Yaml",
238
+ "pope": "Yaml",
239
+ "textvqa_val": "Yaml"
240
+ },
241
+ "n-shot": {
242
+ "mmerealworld_lite": 0,
243
+ "ocrbench": 0,
244
+ "pope": 0,
245
+ "textvqa_val": 0
246
+ },
247
+ "model_configs": {
248
+ "model": "llava",
249
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/sft/SMOE/test_Full_smoe,conv_template=phi35",
250
+ "batch_size": "1",
251
+ "device": null,
252
+ "limit": null,
253
+ "bootstrap_iters": 100000,
254
+ "gen_kwargs": ""
255
+ },
256
+ "git_hash": "b61cb97"
257
+ }
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/results/ocrbench_results.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ######################### OCRBench #############################
2
+ Text Recognition(Total 300): 180
3
+ ---------------- Details of Recognition Score ------------------
4
+ Regular Text Recognition(Total 50): 46
5
+ Irregular Text Recognition(Total 50): 40
6
+ Artistic Text Recognition(Total 50): 45
7
+ Handwriting Recognition(Total 50): 28
8
+ Digit String Recognition(Total 50): 9
9
+ Non-Semantic Text Recognition(Total 50): 12
10
+ ----------------------------------------------------------------
11
+ Scene Text-centric VQA(Total 200): 111
12
+ ----------------------------------------------------------------
13
+ Doc-oriented VQA(Total 200): 23
14
+ ----------------------------------------------------------------
15
+ Key Information Extraction(Total 200): 9
16
+ Handwritten Mathematical Expression Recognition(Total 100): 0
17
+ --------------------- Final Score ------------------------------
18
+ Final Score(Total 1000): 323
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/submissions/textvqa_submission_2025-04-12-18-52-07.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/logs/ocrbench,textvqa_val,mmerealworld_lite,pope/0412_1933_llava..._pope_llava_model_args_4172a9/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb018759d20b23f05a8f6287cd9f52dbf3bd8915f6d5bffa4a679c0f4749b3bf
3
+ size 13148492
sft/smoe_refine_665k_llava/model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eed1597e630d33e971332c65f90e0abaa9a72d47769733b53e9bc39e1684d8f
3
  size 4972489328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25d5a7cb279b9c52201e13da03fdcfcf3d4bf862f49543a29297a9c4bf94ca13
3
  size 4972489328
sft/smoe_refine_665k_llava/model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ddb3a759a0dd0e05fd3bc46899f02df24184e8eb1b117018b37fa3d6ef293dad
3
  size 4985529648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab05102ad84689c3a903f896d9e82da4045c9b8911256f625b22f4c5d8d99951
3
  size 4985529648
sft/smoe_refine_665k_llava/model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a43396d05647807918cc144c703116819da443aa175c98024702aa939391a252
3
  size 248943552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3044d9ec03437ea57f9f62abd7b681bbb31776b46339997b58a15ad2594db92a
3
  size 248943552
sft/smoe_refine_665k_llava/trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e89343be69cf9344a0c92b8a9b8d2f2e376d7e19a1cc7fe1cba698b6a65fa0c
3
- size 7928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8f88ca76a6d6eb76c2f7f274ca0c114e1f74df026b4738ee990a2acf8268c9b
3
+ size 7992