DavidNguyen commited on
Commit
c4ed689
·
verified ·
1 Parent(s): 6e53354

dddc916d0af8ce5501874b0e9bcdd410372d15780b8254e327e14723fd9da0d7

Browse files
Files changed (37) hide show
  1. .gitattributes +5 -0
  2. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/mmerealworld_lite.json +3 -0
  3. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/ocrbench.json +0 -0
  4. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/pope.json +3 -0
  5. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank0_metric_eval_done.txt +1 -0
  6. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank1_metric_eval_done.txt +1 -0
  7. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank2_metric_eval_done.txt +1 -0
  8. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank3_metric_eval_done.txt +1 -0
  9. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/results.json +257 -0
  10. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/results/ocrbench_results.txt +18 -0
  11. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/submissions/textvqa_submission_2025-04-09-17-19-18.json +0 -0
  12. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/textvqa_val.json +3 -0
  13. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/ai2d.json +0 -0
  14. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmbench_en_dev.json +3 -0
  15. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmmu_val.json +0 -0
  16. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmstar.json +0 -0
  17. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank0_metric_eval_done.txt +1 -0
  18. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank1_metric_eval_done.txt +1 -0
  19. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank2_metric_eval_done.txt +1 -0
  20. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank3_metric_eval_done.txt +1 -0
  21. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/results.json +298 -0
  22. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.json +1 -0
  23. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.xlsx +3 -0
  24. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1843_llava...image_llava_model_args_7bd459/rank1_metric_eval_done.txt +1 -0
  25. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1843_llava...image_llava_model_args_7bd459/rank2_metric_eval_done.txt +1 -0
  26. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1843_llava...image_llava_model_args_7bd459/rank3_metric_eval_done.txt +1 -0
  27. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/mathvista_testmini.json +0 -0
  28. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank0_metric_eval_done.txt +1 -0
  29. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank1_metric_eval_done.txt +1 -0
  30. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank2_metric_eval_done.txt +1 -0
  31. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank3_metric_eval_done.txt +1 -0
  32. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/results.json +79 -0
  33. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/submissions/mathvista_testmini_scores.json +0 -0
  34. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/gpt_response/hallusion_output_vd_model.json +0 -0
  35. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/rank1_metric_eval_done.txt +1 -0
  36. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/rank2_metric_eval_done.txt +1 -0
  37. sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/rank3_metric_eval_done.txt +1 -0
.gitattributes CHANGED
@@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
37
+ sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/pope.json filter=lfs diff=lfs merge=lfs -text
38
+ sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
39
+ sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
40
+ sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b26e5e78b76b521492611ba5cfdc56ad379e4dfc7c5a5930f496708bdc63556
3
+ size 1994104210
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/ocrbench.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/pope.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4daf799d1b7ec517a5af1073d1049ec7bee5b95cdd694cf06778b0d0efbbfa9f
3
+ size 17577558
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/results.json ADDED
@@ -0,0 +1,257 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "mmerealworld_lite": {
4
+ "mme_realworld_score,none": 0.3324648254299114,
5
+ "mme_realworld_score_stderr,none": "N/A",
6
+ "alias": "mmerealworld_lite"
7
+ },
8
+ "ocrbench": {
9
+ "ocrbench_accuracy,none": 0.319,
10
+ "ocrbench_accuracy_stderr,none": "N/A",
11
+ "alias": "ocrbench"
12
+ },
13
+ "pope": {
14
+ "pope_accuracy,none": 0.8682222222222222,
15
+ "pope_accuracy_stderr,none": "N/A",
16
+ "pope_precision,none": 0.9297199170124482,
17
+ "pope_precision_stderr,none": "N/A",
18
+ "pope_recall,none": 0.7966666666666666,
19
+ "pope_recall_stderr,none": "N/A",
20
+ "pope_f1_score,none": 0.8580660603159406,
21
+ "pope_f1_score_stderr,none": "N/A",
22
+ "pope_yes_ratio,none": 0.5,
23
+ "pope_yes_ratio_stderr,none": "N/A",
24
+ "alias": "pope"
25
+ },
26
+ "textvqa_val": {
27
+ "exact_match,none": 0.41270000022649767,
28
+ "exact_match_stderr,none": 0.006748425752359104,
29
+ "submission,none": null,
30
+ "submission_stderr,none": "N/A",
31
+ "alias": "textvqa_val"
32
+ }
33
+ },
34
+ "configs": {
35
+ "mmerealworld_lite": {
36
+ "task": "mmerealworld_lite",
37
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
38
+ "dataset_kwargs": {
39
+ "token": true
40
+ },
41
+ "test_split": "train",
42
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7fe9371bc9d0>",
43
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7fe9371b65e0>",
44
+ "doc_to_target": "answer",
45
+ "process_results": "<function mme_realworld_process_results at 0x7fe9371453a0>",
46
+ "description": "",
47
+ "target_delimiter": " ",
48
+ "fewshot_delimiter": "\n\n",
49
+ "metric_list": [
50
+ {
51
+ "metric": "mme_realworld_score",
52
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7fe937146040>",
53
+ "higher_is_better": true
54
+ }
55
+ ],
56
+ "output_type": "generate_until",
57
+ "generation_kwargs": {
58
+ "max_new_tokens": 16,
59
+ "temperature": 0.0,
60
+ "top_p": 1.0,
61
+ "num_beams": 1,
62
+ "do_sample": false,
63
+ "until": [
64
+ "\n\n"
65
+ ]
66
+ },
67
+ "repeats": 1,
68
+ "should_decontaminate": false,
69
+ "metadata": [
70
+ {
71
+ "version": 0.0
72
+ }
73
+ ],
74
+ "model_specific_prompt_kwargs": {
75
+ "default": {
76
+ "pre_prompt": "",
77
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
78
+ },
79
+ "gpt4v": {
80
+ "pre_prompt": "",
81
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
82
+ },
83
+ "xcomposer2_4khd": {
84
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
85
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
86
+ }
87
+ }
88
+ },
89
+ "ocrbench": {
90
+ "task": "ocrbench",
91
+ "dataset_path": "echo840/OCRBench",
92
+ "dataset_kwargs": {
93
+ "token": true
94
+ },
95
+ "test_split": "test",
96
+ "doc_to_visual": "<function ocrbench_doc_to_visual at 0x7fe95037f040>",
97
+ "doc_to_text": "<function ocrbench_doc_to_text at 0x7fe95f79f670>",
98
+ "doc_to_target": "answer",
99
+ "process_results": "<function ocrbench_process_results at 0x7fe95f79f9d0>",
100
+ "description": "",
101
+ "target_delimiter": " ",
102
+ "fewshot_delimiter": "\n\n",
103
+ "metric_list": [
104
+ {
105
+ "metric": "ocrbench_accuracy",
106
+ "aggregation": "<function ocrbench_aggregate_accuracy at 0x7fe95f79fd30>",
107
+ "higher_is_better": true
108
+ }
109
+ ],
110
+ "output_type": "generate_until",
111
+ "generation_kwargs": {
112
+ "max_new_tokens": 128,
113
+ "temperature": 0.0,
114
+ "top_p": 1.0,
115
+ "num_beams": 1,
116
+ "do_sample": false,
117
+ "until": [
118
+ "\n\n"
119
+ ]
120
+ },
121
+ "repeats": 1,
122
+ "should_decontaminate": false,
123
+ "metadata": [
124
+ {
125
+ "version": 0.0
126
+ }
127
+ ]
128
+ },
129
+ "pope": {
130
+ "task": "pope",
131
+ "dataset_path": "lmms-lab/POPE",
132
+ "dataset_kwargs": {
133
+ "token": true
134
+ },
135
+ "test_split": "test",
136
+ "doc_to_visual": "<function pope_doc_to_visual at 0x7fe9371cfd30>",
137
+ "doc_to_text": "<function pope_doc_to_text at 0x7fe9371d6280>",
138
+ "doc_to_target": "answer",
139
+ "process_results": "<function pope_process_results at 0x7fe9371d6790>",
140
+ "description": "",
141
+ "target_delimiter": " ",
142
+ "fewshot_delimiter": "\n\n",
143
+ "metric_list": [
144
+ {
145
+ "metric": "pope_accuracy",
146
+ "aggregation": "<function pope_aggregate_accuracy at 0x7fe9371d6ca0>",
147
+ "higher_is_better": true
148
+ },
149
+ {
150
+ "metric": "pope_precision",
151
+ "aggregation": "<function pope_aggregate_precision at 0x7fe9371db1f0>",
152
+ "higher_is_better": true
153
+ },
154
+ {
155
+ "metric": "pope_recall",
156
+ "aggregation": "<function pope_aggregate_recall at 0x7fe9371db700>",
157
+ "higher_is_better": true
158
+ },
159
+ {
160
+ "metric": "pope_f1_score",
161
+ "aggregation": "<function pope_aggregate_f1_score at 0x7fe9371dbc10>",
162
+ "higher_is_better": true
163
+ },
164
+ {
165
+ "metric": "pope_yes_ratio",
166
+ "aggregation": "<function pope_aggregate_yes_ratio at 0x7fe9371e0160>",
167
+ "higher_is_better": true
168
+ }
169
+ ],
170
+ "output_type": "generate_until",
171
+ "generation_kwargs": {
172
+ "max_new_tokens": 128,
173
+ "temperature": 0.0,
174
+ "top_p": 1.0,
175
+ "num_beams": 1,
176
+ "do_sample": false,
177
+ "until": [
178
+ "\n\n"
179
+ ]
180
+ },
181
+ "repeats": 1,
182
+ "should_decontaminate": false,
183
+ "metadata": [
184
+ {
185
+ "version": 0.0
186
+ }
187
+ ]
188
+ },
189
+ "textvqa_val": {
190
+ "task": "textvqa_val",
191
+ "dataset_path": "lmms-lab/textvqa",
192
+ "test_split": "validation",
193
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7fe96b6a60d0>",
194
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7fe96b6a6550>",
195
+ "doc_to_target": "answer",
196
+ "process_results": "<function textvqa_process_results at 0x7fe96b6a6820>",
197
+ "description": "",
198
+ "target_delimiter": " ",
199
+ "fewshot_delimiter": "\n\n",
200
+ "metric_list": [
201
+ {
202
+ "metric": "exact_match",
203
+ "aggregation": "mean",
204
+ "higher_is_better": true,
205
+ "ignore_case": true,
206
+ "ignore_punctuation": true
207
+ },
208
+ {
209
+ "metric": "submission",
210
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7fe96b708ee0>",
211
+ "higher_is_better": true
212
+ }
213
+ ],
214
+ "output_type": "generate_until",
215
+ "generation_kwargs": {
216
+ "until": [
217
+ "ASSISTANT:"
218
+ ]
219
+ },
220
+ "repeats": 1,
221
+ "should_decontaminate": false,
222
+ "model_specific_prompt_kwargs": {
223
+ "default": {
224
+ "pre_prompt": "",
225
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
226
+ "ocr": false
227
+ },
228
+ "qwen_vl": {
229
+ "pre_prompt": "",
230
+ "post_prompt": " Answer:"
231
+ }
232
+ }
233
+ }
234
+ },
235
+ "versions": {
236
+ "mmerealworld_lite": "Yaml",
237
+ "ocrbench": "Yaml",
238
+ "pope": "Yaml",
239
+ "textvqa_val": "Yaml"
240
+ },
241
+ "n-shot": {
242
+ "mmerealworld_lite": 0,
243
+ "ocrbench": 0,
244
+ "pope": 0,
245
+ "textvqa_val": 0
246
+ },
247
+ "model_configs": {
248
+ "model": "llava",
249
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/sft/SMOE/Full_smoe/checkpoint-16632,conv_template=phi35",
250
+ "batch_size": "1",
251
+ "device": null,
252
+ "limit": null,
253
+ "bootstrap_iters": 100000,
254
+ "gen_kwargs": ""
255
+ },
256
+ "git_hash": "bf33b90"
257
+ }
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/results/ocrbench_results.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ######################### OCRBench #############################
2
+ Text Recognition(Total 300): 182
3
+ ---------------- Details of Recognition Score ------------------
4
+ Regular Text Recognition(Total 50): 46
5
+ Irregular Text Recognition(Total 50): 42
6
+ Artistic Text Recognition(Total 50): 44
7
+ Handwriting Recognition(Total 50): 26
8
+ Digit String Recognition(Total 50): 11
9
+ Non-Semantic Text Recognition(Total 50): 13
10
+ ----------------------------------------------------------------
11
+ Scene Text-centric VQA(Total 200): 107
12
+ ----------------------------------------------------------------
13
+ Doc-oriented VQA(Total 200): 22
14
+ ----------------------------------------------------------------
15
+ Key Information Extraction(Total 200): 8
16
+ Handwritten Mathematical Expression Recognition(Total 100): 0
17
+ --------------------- Final Score ------------------------------
18
+ Final Score(Total 1000): 319
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/submissions/textvqa_submission_2025-04-09-17-19-18.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1803_llava..._pope_llava_model_args_7bd459/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02035d62838017ab756a6df07b88060c358076cdf3927fca0f0048c65b88a09d
3
+ size 13150501
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4dcfddcb36e62b42fba763210b96466f64630d467f8192bc4803317fc5f3db0
3
+ size 14568169
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/results.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6492875647668394,
5
+ "exact_match_stderr,flexible-extract": 0.008588667905633653,
6
+ "alias": "ai2d"
7
+ },
8
+ "mmbench_en_dev": {
9
+ "gpt_eval_score,none": 71.99312714776632,
10
+ "gpt_eval_score_stderr,none": "N/A",
11
+ "submission,none": null,
12
+ "submission_stderr,none": "N/A",
13
+ "alias": "mmbench_en_dev"
14
+ },
15
+ "mmmu_val": {
16
+ "mmmu_acc,none": 0.42222,
17
+ "mmmu_acc_stderr,none": "N/A",
18
+ "alias": "mmmu_val"
19
+ },
20
+ "mmstar": {
21
+ "coarse perception,none": 0.7030779789873819,
22
+ "coarse perception_stderr,none": "N/A",
23
+ "fine-grained perception,none": 0.34155612871530333,
24
+ "fine-grained perception_stderr,none": "N/A",
25
+ "instance reasoning,none": 0.53201826163044,
26
+ "instance reasoning_stderr,none": "N/A",
27
+ "logical reasoning,none": 0.36875764499526875,
28
+ "logical reasoning_stderr,none": "N/A",
29
+ "math,none": 0.2920175086875167,
30
+ "math_stderr,none": "N/A",
31
+ "science & technology,none": 0.25190144653700547,
32
+ "science & technology_stderr,none": "N/A",
33
+ "alias": "mmstar"
34
+ }
35
+ },
36
+ "configs": {
37
+ "ai2d": {
38
+ "task": "ai2d",
39
+ "dataset_path": "lmms-lab/ai2d",
40
+ "dataset_kwargs": {
41
+ "token": true
42
+ },
43
+ "test_split": "test",
44
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7ef90acbe8b0>",
45
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7ef90acbeb80>",
46
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7ef90ac48040>",
47
+ "description": "",
48
+ "target_delimiter": " ",
49
+ "fewshot_delimiter": "\n\n",
50
+ "metric_list": [
51
+ {
52
+ "metric": "exact_match",
53
+ "aggregation": "mean",
54
+ "higher_is_better": true,
55
+ "ignore_case": true,
56
+ "ignore_punctuation": true
57
+ }
58
+ ],
59
+ "output_type": "generate_until",
60
+ "generation_kwargs": {
61
+ "max_new_tokens": 512,
62
+ "temperature": 0.0,
63
+ "do_sample": false,
64
+ "until": [
65
+ "\n\n"
66
+ ]
67
+ },
68
+ "repeats": 1,
69
+ "filter_list": [
70
+ {
71
+ "name": "flexible-extract",
72
+ "filter": [
73
+ {
74
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
75
+ "group_select": 0,
76
+ "ignore_case": true,
77
+ "ignore_punctuation": true,
78
+ "regex_pattern": "([A-Z])\\."
79
+ }
80
+ ]
81
+ }
82
+ ],
83
+ "should_decontaminate": false,
84
+ "metadata": [
85
+ {
86
+ "version": 0.0
87
+ }
88
+ ],
89
+ "model_specific_prompt_kwargs": {
90
+ "default": {
91
+ "prompt_format": "mcq",
92
+ "pre_prompt": "",
93
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
94
+ },
95
+ "gpt4v": {
96
+ "prompt_format": "mcq",
97
+ "pre_prompt": "",
98
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
99
+ },
100
+ "qwen_vl": {
101
+ "prompt_format": "qa",
102
+ "pre_prompt": "",
103
+ "post_prompt": " Answer:"
104
+ },
105
+ "xcomposer2_4khd": {
106
+ "prompt_format": "mcq_xcomposer",
107
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
108
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
109
+ }
110
+ },
111
+ "model_specific_target_kwargs": {
112
+ "default": "mcq",
113
+ "qwen_vl": "qa"
114
+ }
115
+ },
116
+ "mmbench_en_dev": {
117
+ "task": "mmbench_en_dev",
118
+ "dataset_path": "lmms-lab/MMBench",
119
+ "dataset_name": "en",
120
+ "dataset_kwargs": {
121
+ "token": true
122
+ },
123
+ "test_split": "dev",
124
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7ef969d64e50>",
125
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7ef969d67430>",
126
+ "doc_to_target": "answer",
127
+ "process_results": "<function mmbench_process_results at 0x7ef969d679d0>",
128
+ "description": "",
129
+ "target_delimiter": " ",
130
+ "fewshot_delimiter": "\n\n",
131
+ "metric_list": [
132
+ {
133
+ "metric": "gpt_eval_score",
134
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7ef969d645e0>",
135
+ "higher_is_better": true
136
+ },
137
+ {
138
+ "metric": "submission",
139
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7ef969d64b80>",
140
+ "higher_is_better": true
141
+ }
142
+ ],
143
+ "output_type": "generate_until",
144
+ "generation_kwargs": {
145
+ "until": [
146
+ "ASSISTANT:"
147
+ ],
148
+ "max_new_tokens": 1024,
149
+ "temperature": 0.0,
150
+ "top_p": 1.0,
151
+ "num_beams": 1,
152
+ "do_sample": false,
153
+ "image_aspect_ratio": "original"
154
+ },
155
+ "repeats": 1,
156
+ "should_decontaminate": false,
157
+ "model_specific_prompt_kwargs": {
158
+ "default": {
159
+ "pre_prompt": "",
160
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
161
+ }
162
+ },
163
+ "model_specific_generation_kwargs": {
164
+ "llava": {
165
+ "image_aspect_ratio": "original"
166
+ }
167
+ }
168
+ },
169
+ "mmmu_val": {
170
+ "task": "mmmu_val",
171
+ "dataset_path": "lmms-lab/MMMU",
172
+ "test_split": "validation",
173
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7ef932f079d0>",
174
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7ef932e1a670>",
175
+ "doc_to_target": "answer",
176
+ "process_results": "<function mmmu_process_results at 0x7ef932cf84c0>",
177
+ "description": "",
178
+ "target_delimiter": " ",
179
+ "fewshot_delimiter": "\n\n",
180
+ "metric_list": [
181
+ {
182
+ "metric": "mmmu_acc",
183
+ "aggregation": "<function mmmu_aggregate_results at 0x7ef932c263a0>",
184
+ "higher_is_better": true
185
+ }
186
+ ],
187
+ "output_type": "generate_until",
188
+ "generation_kwargs": {
189
+ "max_new_tokens": 128,
190
+ "until": [
191
+ "\n\n"
192
+ ],
193
+ "image_aspect_ratio": "original"
194
+ },
195
+ "repeats": 1,
196
+ "should_decontaminate": false,
197
+ "metadata": [
198
+ {
199
+ "version": 0.0
200
+ }
201
+ ],
202
+ "model_specific_generation_kwargs": {
203
+ "llava": {
204
+ "image_aspect_ratio": "original"
205
+ }
206
+ }
207
+ },
208
+ "mmstar": {
209
+ "task": "mmstar",
210
+ "dataset_path": "Lin-Chen/MMStar",
211
+ "dataset_kwargs": {
212
+ "token": true
213
+ },
214
+ "test_split": "val",
215
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7ef90ab97a60>",
216
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7ef90ab97e50>",
217
+ "doc_to_target": "answer",
218
+ "process_results": "<function mmstar_process_results at 0x7ef90aba1310>",
219
+ "description": "",
220
+ "target_delimiter": " ",
221
+ "fewshot_delimiter": "\n\n",
222
+ "metric_list": [
223
+ {
224
+ "metric": "coarse perception",
225
+ "aggregation": "<function mmstar_aggregate_results at 0x7ef90aba1700>",
226
+ "higher_is_better": true
227
+ },
228
+ {
229
+ "metric": "fine-grained perception",
230
+ "aggregation": "<function mmstar_aggregate_results at 0x7ef90aba1a60>",
231
+ "higher_is_better": true
232
+ },
233
+ {
234
+ "metric": "instance reasoning",
235
+ "aggregation": "<function mmstar_aggregate_results at 0x7ef90aba1dc0>",
236
+ "higher_is_better": true
237
+ },
238
+ {
239
+ "metric": "logical reasoning",
240
+ "aggregation": "<function mmstar_aggregate_results at 0x7ef90aba9160>",
241
+ "higher_is_better": true
242
+ },
243
+ {
244
+ "metric": "science & technology",
245
+ "aggregation": "<function mmstar_aggregate_results at 0x7ef90aba94c0>",
246
+ "higher_is_better": true
247
+ },
248
+ {
249
+ "metric": "math",
250
+ "aggregation": "<function mmstar_aggregate_results at 0x7ef90aba9820>",
251
+ "higher_is_better": true
252
+ }
253
+ ],
254
+ "output_type": "generate_until",
255
+ "generation_kwargs": {
256
+ "until": [
257
+ "\n\n"
258
+ ],
259
+ "do_sample": false
260
+ },
261
+ "repeats": 1,
262
+ "should_decontaminate": false,
263
+ "metadata": [
264
+ {
265
+ "version": 0.0
266
+ }
267
+ ],
268
+ "model_specific_prompt_kwargs": {
269
+ "default": {
270
+ "pre_prompt": "",
271
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
272
+ }
273
+ }
274
+ }
275
+ },
276
+ "versions": {
277
+ "ai2d": "Yaml",
278
+ "mmbench_en_dev": "Yaml",
279
+ "mmmu_val": "Yaml",
280
+ "mmstar": "Yaml"
281
+ },
282
+ "n-shot": {
283
+ "ai2d": 0,
284
+ "mmbench_en_dev": 0,
285
+ "mmmu_val": 0,
286
+ "mmstar": 0
287
+ },
288
+ "model_configs": {
289
+ "model": "llava",
290
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/sft/SMOE/Full_smoe/checkpoint-16632,conv_template=phi35",
291
+ "batch_size": "1",
292
+ "device": null,
293
+ "limit": null,
294
+ "bootstrap_iters": 100000,
295
+ "gen_kwargs": ""
296
+ },
297
+ "git_hash": "bf33b90"
298
+ }
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.7199312714776632, "category_acc": {"action_recognition": 0.9259259259259259, "attribute_comparison": 0.6818181818181818, "attribute_recognition": 0.8648648648648649, "celebrity_recognition": 0.8080808080808081, "function_reasoning": 0.8734177215189873, "future_prediction": 0.5, "identity_reasoning": 0.9777777777777777, "image_emotion": 0.84, "image_quality": 0.5849056603773585, "image_scene": 0.9423076923076923, "image_style": 0.7924528301886793, "image_topic": 0.8888888888888888, "nature_relation": 0.5208333333333334, "object_localization": 0.5432098765432098, "ocr": 0.6666666666666666, "physical_property_reasoning": 0.5866666666666667, "physical_relation": 0.4583333333333333, "social_relation": 0.8372093023255814, "spatial_relationship": 0.3111111111111111, "structuralized_imagetext_understanding": 0.46153846153846156}, "l2_category_acc": {"attribute_reasoning": 0.7889447236180904, "coarse_perception": 0.8277027027027027, "finegrained_perception (cross-instance)": 0.6573426573426573, "finegrained_perception (instance-level)": 0.7303754266211604, "logic_reasoning": 0.4745762711864407, "relation_reasoning": 0.6260869565217392}}
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1833_llava...mstar_llava_model_args_7bd459/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8e1bf82d81bdc3d83cc5a496f96c0d8c23b646e46eebc33d1afa187aa952791
3
+ size 865173
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1843_llava...image_llava_model_args_7bd459/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1843_llava...image_llava_model_args_7bd459/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1843_llava...image_llava_model_args_7bd459/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/mathvista_testmini.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/results.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "mathvista_testmini": {
4
+ "gpt_eval_score,none": 31.9,
5
+ "gpt_eval_score_stderr,none": "N/A",
6
+ "alias": "mathvista_testmini"
7
+ }
8
+ },
9
+ "configs": {
10
+ "mathvista_testmini": {
11
+ "task": "mathvista_testmini",
12
+ "dataset_path": "AI4Math/MathVista",
13
+ "dataset_kwargs": {
14
+ "token": true
15
+ },
16
+ "test_split": "testmini",
17
+ "doc_to_visual": "<function mathvista_doc_to_visual at 0x7facdb3ecc10>",
18
+ "doc_to_text": "<function mathvista_doc_to_text at 0x7facdb3f8310>",
19
+ "doc_to_target": "answer",
20
+ "process_results": "<function mathvista_process_results at 0x7facdb4029d0>",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "gpt_eval_score",
27
+ "aggregation": "<function mathvista_aggregate_results at 0x7facdaca90d0>",
28
+ "higher_is_better": true
29
+ }
30
+ ],
31
+ "output_type": "generate_until",
32
+ "generation_kwargs": {
33
+ "until": [
34
+ "ASSISTANT:"
35
+ ],
36
+ "max_new_tokens": 1024,
37
+ "temperature": 0.0,
38
+ "top_p": 1.0,
39
+ "num_beams": 1,
40
+ "do_sample": false,
41
+ "image_aspect_ratio": "original"
42
+ },
43
+ "repeats": 1,
44
+ "should_decontaminate": false,
45
+ "model_specific_prompt_kwargs": {
46
+ "default": {
47
+ "shot_type": "format-prompt",
48
+ "shot": 0,
49
+ "use_caption": false,
50
+ "use_ocr": false
51
+ },
52
+ "phi3v": {
53
+ "shot_type": "solution"
54
+ }
55
+ },
56
+ "model_specific_generation_kwargs": {
57
+ "llava": {
58
+ "image_aspect_ratio": "original"
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "versions": {
64
+ "mathvista_testmini": "Yaml"
65
+ },
66
+ "n-shot": {
67
+ "mathvista_testmini": 0
68
+ },
69
+ "model_configs": {
70
+ "model": "llava",
71
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/sft/SMOE/Full_smoe/checkpoint-16632,conv_template=phi35",
72
+ "batch_size": "1",
73
+ "device": null,
74
+ "limit": null,
75
+ "bootstrap_iters": 100000,
76
+ "gen_kwargs": ""
77
+ },
78
+ "git_hash": "bf33b90"
79
+ }
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1848_llava...tmini_llava_model_args_7bd459/submissions/mathvista_testmini_scores.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/gpt_response/hallusion_output_vd_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/smoe_refine_665k_llava/checkpoint-16632/logs/0409_1908_llava...image_llava_model_args_7bd459/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done