DavidNguyen commited on
Commit
2483382
·
verified ·
1 Parent(s): 4f11a8c

25119750f8c75c6ca0431185319ad552584348976044b67513b1e9f31d3e067e

Browse files
Files changed (37) hide show
  1. .gitattributes +2 -0
  2. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/infovqa_val.json +3 -0
  3. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/mmmu_pro_standard.json +0 -0
  4. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/mmmu_pro_vision.json +0 -0
  5. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank0_metric_eval_done.txt +1 -0
  6. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank1_metric_eval_done.txt +1 -0
  7. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank2_metric_eval_done.txt +1 -0
  8. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank3_metric_eval_done.txt +1 -0
  9. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/realworldqa.json +0 -0
  10. sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/results.json +245 -0
  11. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vd_model.json +0 -0
  12. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vs_model.json +0 -0
  13. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/hallusion_bench_image.json +0 -0
  14. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/mathvista_testmini.json +0 -0
  15. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/rank0_metric_eval_done.txt +1 -0
  16. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/rank1_metric_eval_done.txt +1 -0
  17. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/results.json +146 -0
  18. sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/submissions/mathvista_testmini_scores.json +0 -0
  19. sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/pope.json +3 -0
  20. sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank0_metric_eval_done.txt +1 -0
  21. sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank1_metric_eval_done.txt +1 -0
  22. sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank2_metric_eval_done.txt +1 -0
  23. sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank3_metric_eval_done.txt +1 -0
  24. sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/results.json +95 -0
  25. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vd_model.json +0 -0
  26. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vs_model.json +0 -0
  27. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/hallusion_bench_image.json +0 -0
  28. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/mathvista_testmini.json +0 -0
  29. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/rank0_metric_eval_done.txt +1 -0
  30. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/rank1_metric_eval_done.txt +1 -0
  31. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/results.json +146 -0
  32. sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/submissions/mathvista_testmini_scores.json +0 -0
  33. sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/chartqa.json +0 -0
  34. sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/rank0_metric_eval_done.txt +1 -0
  35. sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/rank1_metric_eval_done.txt +1 -0
  36. sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/results.json +168 -0
  37. sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/seedbench_2_plus.json +0 -0
.gitattributes CHANGED
@@ -131,3 +131,5 @@ sft/1M3/Full_smoe/logs/0613_0224_llava..._pope_llava_model_args_bc59f1/pope.json
131
  sft/1M3/Full_smoe/logs/0613_0224_llava..._pope_llava_model_args_bc59f1/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
132
  sft/1M3/Full_smoe/logs/0613_0224_llava..._pope_llava_model_args_bc59f1/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
133
  sft/1M3/Full_smoe/logs/0613_0343_llava_v1.5_gqa_llava_model_args_bc59f1/gqa.json filter=lfs diff=lfs merge=lfs -text
 
 
 
131
  sft/1M3/Full_smoe/logs/0613_0224_llava..._pope_llava_model_args_bc59f1/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
132
  sft/1M3/Full_smoe/logs/0613_0224_llava..._pope_llava_model_args_bc59f1/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
133
  sft/1M3/Full_smoe/logs/0613_0343_llava_v1.5_gqa_llava_model_args_bc59f1/gqa.json filter=lfs diff=lfs merge=lfs -text
134
+ sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
135
+ sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/pope.json filter=lfs diff=lfs merge=lfs -text
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/infovqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cec738425cb57f14dd3611a9fe34493badd8dde4f45b0bf1ff28bbd362a6fc8
3
+ size 576437772
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/mmmu_pro_standard.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/mmmu_pro_vision.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/realworldqa.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_0402_llava...a_val_llava_model_args_bc59f1/results.json ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "infovqa_val": {
4
+ "anls,none": 0.29632274187790075,
5
+ "anls_stderr,none": 0.008629591418151561,
6
+ "alias": "infovqa_val"
7
+ },
8
+ "mmmu_pro": {
9
+ "mmmu_acc,none": 0.1922,
10
+ "mmmu_acc_stderr,none": 0.03656528438248045,
11
+ "alias": "mmmu_pro"
12
+ },
13
+ "mmmu_pro_standard": {
14
+ "mmmu_acc,none": 0.26532,
15
+ "mmmu_acc_stderr,none": "N/A",
16
+ "alias": " - mmmu_pro_standard"
17
+ },
18
+ "mmmu_pro_vision": {
19
+ "mmmu_acc,none": 0.11908,
20
+ "mmmu_acc_stderr,none": "N/A",
21
+ "alias": " - mmmu_pro_vision"
22
+ },
23
+ "realworldqa": {
24
+ "exact_match,flexible-extract": 0.4993464052287582,
25
+ "exact_match_stderr,flexible-extract": 0.0180893496772524,
26
+ "alias": "realworldqa"
27
+ }
28
+ },
29
+ "groups": {
30
+ "mmmu_pro": {
31
+ "mmmu_acc,none": 0.1922,
32
+ "mmmu_acc_stderr,none": 0.03656528438248045,
33
+ "alias": "mmmu_pro"
34
+ }
35
+ },
36
+ "configs": {
37
+ "infovqa_val": {
38
+ "task": "infovqa_val",
39
+ "dataset_path": "lmms-lab/DocVQA",
40
+ "dataset_name": "InfographicVQA",
41
+ "dataset_kwargs": {
42
+ "token": true
43
+ },
44
+ "test_split": "validation",
45
+ "doc_to_visual": "<function infovqa_doc_to_visual at 0x7fe18c88bf70>",
46
+ "doc_to_text": "<function infovqa_doc_to_text at 0x7fe18c603280>",
47
+ "doc_to_target": "answers",
48
+ "description": "",
49
+ "target_delimiter": " ",
50
+ "fewshot_delimiter": "\n\n",
51
+ "metric_list": [
52
+ {
53
+ "metric": "anls",
54
+ "aggregation": "mean",
55
+ "higher_is_better": true
56
+ }
57
+ ],
58
+ "output_type": "generate_until",
59
+ "generation_kwargs": {
60
+ "max_new_tokens": 32,
61
+ "temperature": 0.0,
62
+ "do_sample": false,
63
+ "until": [
64
+ "\n\n"
65
+ ]
66
+ },
67
+ "repeats": 1,
68
+ "should_decontaminate": false,
69
+ "model_specific_prompt_kwargs": {
70
+ "default": {
71
+ "pre_prompt": "",
72
+ "post_prompt": "\nAnswer the question using a single word or phrase."
73
+ }
74
+ }
75
+ },
76
+ "mmmu_pro_standard": {
77
+ "task": "mmmu_pro_standard",
78
+ "dataset_path": "MMMU/MMMU_Pro",
79
+ "dataset_name": "standard (10 options)",
80
+ "test_split": "test",
81
+ "doc_to_visual": "<function mmmu_pro_doc_to_visual at 0x7fe1b87bcd30>",
82
+ "doc_to_text": "<function mmmu_pro_doc_to_text at 0x7fe1b87c7af0>",
83
+ "doc_to_target": "{{answer}}",
84
+ "process_results": "<function mmmu_pro_process_results at 0x7fe1b87cda60>",
85
+ "description": "",
86
+ "target_delimiter": " ",
87
+ "fewshot_delimiter": "\n\n",
88
+ "metric_list": [
89
+ {
90
+ "metric": "mmmu_acc",
91
+ "aggregation": "<function mmmu_pro_aggregate_results at 0x7fe1b87d49d0>",
92
+ "higher_is_better": true
93
+ }
94
+ ],
95
+ "output_type": "generate_until",
96
+ "generation_kwargs": {
97
+ "max_new_tokens": 256,
98
+ "until": [
99
+ "\n\n"
100
+ ]
101
+ },
102
+ "repeats": 1,
103
+ "should_decontaminate": false,
104
+ "metadata": {
105
+ "version": 0.0,
106
+ "interleaved_format": false
107
+ },
108
+ "model_specific_prompt_kwargs": {
109
+ "default": {
110
+ "pre_prompt": "",
111
+ "post_prompt": "Answer with the option letter from the given choices directly."
112
+ }
113
+ }
114
+ },
115
+ "mmmu_pro_vision": {
116
+ "task": "mmmu_pro_vision",
117
+ "dataset_path": "MMMU/MMMU_Pro",
118
+ "dataset_name": "vision",
119
+ "test_split": "test",
120
+ "doc_to_visual": "<function mmmu_pro_doc_to_visual at 0x7fe1b87b1280>",
121
+ "doc_to_text": "Answer with the option letter from the given choices directly.",
122
+ "doc_to_target": "{{answer}}",
123
+ "process_results": "<function mmmu_pro_process_results at 0x7fe1b87b7160>",
124
+ "description": "",
125
+ "target_delimiter": " ",
126
+ "fewshot_delimiter": "\n\n",
127
+ "metric_list": [
128
+ {
129
+ "metric": "mmmu_acc",
130
+ "aggregation": "<function mmmu_pro_aggregate_results at 0x7fe1b87bc0d0>",
131
+ "higher_is_better": true
132
+ }
133
+ ],
134
+ "output_type": "generate_until",
135
+ "generation_kwargs": {
136
+ "max_new_tokens": 256,
137
+ "until": [
138
+ "\n\n"
139
+ ]
140
+ },
141
+ "repeats": 1,
142
+ "should_decontaminate": false,
143
+ "metadata": {
144
+ "version": 0.0,
145
+ "interleaved_format": false
146
+ }
147
+ },
148
+ "realworldqa": {
149
+ "task": "realworldqa",
150
+ "dataset_path": "lmms-lab/RealWorldQA",
151
+ "dataset_kwargs": {
152
+ "token": true
153
+ },
154
+ "test_split": "test",
155
+ "doc_to_visual": "<function realworldqa_doc_to_visual at 0x7fe183dbe040>",
156
+ "doc_to_text": "<function realworldqa_doc_to_text at 0x7fe183dbe670>",
157
+ "doc_to_target": "answer",
158
+ "description": "",
159
+ "target_delimiter": " ",
160
+ "fewshot_delimiter": "\n\n",
161
+ "metric_list": [
162
+ {
163
+ "metric": "exact_match",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true,
166
+ "ignore_case": true,
167
+ "ignore_punctuation": true
168
+ }
169
+ ],
170
+ "output_type": "generate_until",
171
+ "generation_kwargs": {
172
+ "max_new_tokens": 16,
173
+ "temperature": 0.0,
174
+ "top_p": 1.0,
175
+ "num_beams": 1,
176
+ "do_sample": false,
177
+ "until": [
178
+ "\n\n"
179
+ ]
180
+ },
181
+ "repeats": 1,
182
+ "filter_list": [
183
+ {
184
+ "name": "flexible-extract",
185
+ "filter": [
186
+ {
187
+ "function": "<class 'utils.NumberWordsToDigitsFilter'>"
188
+ },
189
+ {
190
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
191
+ "group_select": 0,
192
+ "ignore_case": true,
193
+ "ignore_punctuation": true,
194
+ "regex_pattern": "(\\([A-Z]\\))"
195
+ }
196
+ ]
197
+ }
198
+ ],
199
+ "should_decontaminate": false,
200
+ "metadata": [
201
+ {
202
+ "version": 0.0
203
+ }
204
+ ],
205
+ "model_specific_prompt_kwargs": {
206
+ "default": {
207
+ "pre_prompt": "",
208
+ "post_prompt": ""
209
+ },
210
+ "gpt4v": {
211
+ "pre_prompt": "",
212
+ "post_prompt": ""
213
+ },
214
+ "xcomposer2_4khd": {
215
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
216
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
217
+ }
218
+ }
219
+ }
220
+ },
221
+ "versions": {
222
+ "infovqa_val": "Yaml",
223
+ "mmmu_pro": "N/A",
224
+ "mmmu_pro_standard": "Yaml",
225
+ "mmmu_pro_vision": "Yaml",
226
+ "realworldqa": "Yaml"
227
+ },
228
+ "n-shot": {
229
+ "infovqa_val": 0,
230
+ "mmmu_pro": 0,
231
+ "mmmu_pro_standard": 0,
232
+ "mmmu_pro_vision": 0,
233
+ "realworldqa": 0
234
+ },
235
+ "model_configs": {
236
+ "model": "llava",
237
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_smoe,conv_template=phi35",
238
+ "batch_size": "1",
239
+ "device": null,
240
+ "limit": null,
241
+ "bootstrap_iters": 100000,
242
+ "gen_kwargs": ""
243
+ },
244
+ "git_hash": "289c7fe5"
245
+ }
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vd_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vs_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/hallusion_bench_image.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/mathvista_testmini.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/results.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hallusion_bench_image": {
4
+ "aAcc,none": 47.8444,
5
+ "aAcc_stderr,none": "N/A",
6
+ "fAcc,none": 21.3873,
7
+ "fAcc_stderr,none": "N/A",
8
+ "qAcc,none": 18.4615,
9
+ "qAcc_stderr,none": "N/A",
10
+ "alias": "hallusion_bench_image"
11
+ },
12
+ "mathvista_testmini": {
13
+ "gpt_eval_score,none": 39.9,
14
+ "gpt_eval_score_stderr,none": "N/A",
15
+ "alias": "mathvista_testmini"
16
+ }
17
+ },
18
+ "configs": {
19
+ "hallusion_bench_image": {
20
+ "task": "hallusion_bench_image",
21
+ "dataset_path": "lmms-lab/HallusionBench",
22
+ "dataset_kwargs": {
23
+ "token": true
24
+ },
25
+ "test_split": "image",
26
+ "doc_to_visual": "<function hb_doc_to_visual at 0x7f31170cb700>",
27
+ "doc_to_text": "<function hb_doc_to_text at 0x7f31170cbe50>",
28
+ "doc_to_target": "gt_answer_details",
29
+ "process_results": "<function hb_process_results at 0x7f3116f73670>",
30
+ "description": "",
31
+ "target_delimiter": " ",
32
+ "fewshot_delimiter": "\n\n",
33
+ "metric_list": [
34
+ {
35
+ "metric": "aAcc",
36
+ "aggregation": "<function hb_aggregation_result_aAcc at 0x7f3116f19040>",
37
+ "higher_is_better": true
38
+ },
39
+ {
40
+ "metric": "qAcc",
41
+ "aggregation": "<function hb_aggregation_result_qAcc at 0x7f3116f19670>",
42
+ "higher_is_better": true
43
+ },
44
+ {
45
+ "metric": "fAcc",
46
+ "aggregation": "<function hb_aggregation_result_fAcc at 0x7f3116f19e50>",
47
+ "higher_is_better": true
48
+ }
49
+ ],
50
+ "output_type": "generate_until",
51
+ "generation_kwargs": {
52
+ "max_new_tokens": 128,
53
+ "temperature": 0.0,
54
+ "top_p": 1.0,
55
+ "num_beams": 1,
56
+ "do_sample": false,
57
+ "until": [
58
+ "\n\n"
59
+ ]
60
+ },
61
+ "repeats": 1,
62
+ "should_decontaminate": false,
63
+ "metadata": [
64
+ {
65
+ "version": 0.0
66
+ }
67
+ ],
68
+ "model_specific_prompt_kwargs": {
69
+ "default": {
70
+ "pre_prompt": "",
71
+ "post_prompt": ""
72
+ }
73
+ }
74
+ },
75
+ "mathvista_testmini": {
76
+ "task": "mathvista_testmini",
77
+ "dataset_path": "AI4Math/MathVista",
78
+ "dataset_kwargs": {
79
+ "token": true
80
+ },
81
+ "test_split": "testmini",
82
+ "doc_to_visual": "<function mathvista_doc_to_visual at 0x7f30ec9790d0>",
83
+ "doc_to_text": "<function mathvista_doc_to_text at 0x7f30ec982790>",
84
+ "doc_to_target": "answer",
85
+ "process_results": "<function mathvista_process_results at 0x7f30ec225e50>",
86
+ "description": "",
87
+ "target_delimiter": " ",
88
+ "fewshot_delimiter": "\n\n",
89
+ "metric_list": [
90
+ {
91
+ "metric": "gpt_eval_score",
92
+ "aggregation": "<function mathvista_aggregate_results at 0x7f30ec231550>",
93
+ "higher_is_better": true
94
+ }
95
+ ],
96
+ "output_type": "generate_until",
97
+ "generation_kwargs": {
98
+ "until": [
99
+ "ASSISTANT:"
100
+ ],
101
+ "max_new_tokens": 1024,
102
+ "temperature": 0.0,
103
+ "top_p": 1.0,
104
+ "num_beams": 1,
105
+ "do_sample": false,
106
+ "image_aspect_ratio": "original"
107
+ },
108
+ "repeats": 1,
109
+ "should_decontaminate": false,
110
+ "model_specific_prompt_kwargs": {
111
+ "default": {
112
+ "shot_type": "format-prompt",
113
+ "shot": 0,
114
+ "use_caption": false,
115
+ "use_ocr": false
116
+ },
117
+ "phi3v": {
118
+ "shot_type": "solution"
119
+ }
120
+ },
121
+ "model_specific_generation_kwargs": {
122
+ "llava": {
123
+ "image_aspect_ratio": "original"
124
+ }
125
+ }
126
+ }
127
+ },
128
+ "versions": {
129
+ "hallusion_bench_image": "Yaml",
130
+ "mathvista_testmini": "Yaml"
131
+ },
132
+ "n-shot": {
133
+ "hallusion_bench_image": 0,
134
+ "mathvista_testmini": 0
135
+ },
136
+ "model_configs": {
137
+ "model": "llava",
138
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_smoe,conv_template=phi35",
139
+ "batch_size": "1",
140
+ "device": null,
141
+ "limit": null,
142
+ "bootstrap_iters": 100000,
143
+ "gen_kwargs": ""
144
+ },
145
+ "git_hash": "289c7fe5"
146
+ }
sft/1M3/Full_smoe/logs/0617_1601_llava...image_llava_model_args_bc59f1/submissions/mathvista_testmini_scores.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/pope.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f715e247ed8d6a7da812704f7dc7d3c8026775b2a38b6e549a47bc02586925
3
+ size 17579267
sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_smoe/logs/0626_0643_llava_v1.5_pope_llava_model_args_bc59f1/results.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "pope": {
4
+ "pope_accuracy,none": 0.8466666666666667,
5
+ "pope_accuracy_stderr,none": "N/A",
6
+ "pope_precision,none": 0.8860294117647058,
7
+ "pope_precision_stderr,none": "N/A",
8
+ "pope_recall,none": 0.8033333333333333,
9
+ "pope_recall_stderr,none": "N/A",
10
+ "pope_f1_score,none": 0.8426573426573426,
11
+ "pope_f1_score_stderr,none": "N/A",
12
+ "pope_yes_ratio,none": 0.5,
13
+ "pope_yes_ratio_stderr,none": "N/A",
14
+ "alias": "pope"
15
+ }
16
+ },
17
+ "configs": {
18
+ "pope": {
19
+ "task": "pope",
20
+ "dataset_path": "lmms-lab/POPE",
21
+ "dataset_kwargs": {
22
+ "token": true
23
+ },
24
+ "test_split": "test",
25
+ "doc_to_visual": "<function pope_doc_to_visual at 0x7f05507f3790>",
26
+ "doc_to_text": "<function pope_doc_to_text at 0x7f05507f3d30>",
27
+ "doc_to_target": "answer",
28
+ "process_results": "<function pope_process_results at 0x7f05507f9310>",
29
+ "description": "",
30
+ "target_delimiter": " ",
31
+ "fewshot_delimiter": "\n\n",
32
+ "metric_list": [
33
+ {
34
+ "metric": "pope_accuracy",
35
+ "aggregation": "<function pope_aggregate_accuracy at 0x7f05507f98b0>",
36
+ "higher_is_better": true
37
+ },
38
+ {
39
+ "metric": "pope_precision",
40
+ "aggregation": "<function pope_aggregate_precision at 0x7f05507f9e50>",
41
+ "higher_is_better": true
42
+ },
43
+ {
44
+ "metric": "pope_recall",
45
+ "aggregation": "<function pope_aggregate_recall at 0x7f05507fc430>",
46
+ "higher_is_better": true
47
+ },
48
+ {
49
+ "metric": "pope_f1_score",
50
+ "aggregation": "<function pope_aggregate_f1_score at 0x7f05507fc9d0>",
51
+ "higher_is_better": true
52
+ },
53
+ {
54
+ "metric": "pope_yes_ratio",
55
+ "aggregation": "<function pope_aggregate_yes_ratio at 0x7f05507fcf70>",
56
+ "higher_is_better": true
57
+ }
58
+ ],
59
+ "output_type": "generate_until",
60
+ "generation_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "temperature": 0.0,
63
+ "top_p": 1.0,
64
+ "num_beams": 1,
65
+ "do_sample": false,
66
+ "until": [
67
+ "\n\n"
68
+ ]
69
+ },
70
+ "repeats": 1,
71
+ "should_decontaminate": false,
72
+ "metadata": [
73
+ {
74
+ "version": 0.0
75
+ }
76
+ ]
77
+ }
78
+ },
79
+ "versions": {
80
+ "pope": "Yaml"
81
+ },
82
+ "n-shot": {
83
+ "pope": 0
84
+ },
85
+ "model_configs": {
86
+ "model": "llava",
87
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_smoe,conv_template=phi35",
88
+ "batch_size": "1",
89
+ "device": null,
90
+ "limit": null,
91
+ "bootstrap_iters": 100000,
92
+ "gen_kwargs": ""
93
+ },
94
+ "git_hash": "289c7fe5"
95
+ }
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vd_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/gpt_response/hallusion_output_vs_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/hallusion_bench_image.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/mathvista_testmini.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/results.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hallusion_bench_image": {
4
+ "aAcc,none": 48.1598,
5
+ "aAcc_stderr,none": "N/A",
6
+ "fAcc,none": 21.6763,
7
+ "fAcc_stderr,none": "N/A",
8
+ "qAcc,none": 18.4615,
9
+ "qAcc_stderr,none": "N/A",
10
+ "alias": "hallusion_bench_image"
11
+ },
12
+ "mathvista_testmini": {
13
+ "gpt_eval_score,none": 39.9,
14
+ "gpt_eval_score_stderr,none": "N/A",
15
+ "alias": "mathvista_testmini"
16
+ }
17
+ },
18
+ "configs": {
19
+ "hallusion_bench_image": {
20
+ "task": "hallusion_bench_image",
21
+ "dataset_path": "lmms-lab/HallusionBench",
22
+ "dataset_kwargs": {
23
+ "token": true
24
+ },
25
+ "test_split": "image",
26
+ "doc_to_visual": "<function hb_doc_to_visual at 0x7fc210e580d0>",
27
+ "doc_to_text": "<function hb_doc_to_text at 0x7fc210e58820>",
28
+ "doc_to_target": "gt_answer_details",
29
+ "process_results": "<function hb_process_results at 0x7fc210e7a040>",
30
+ "description": "",
31
+ "target_delimiter": " ",
32
+ "fewshot_delimiter": "\n\n",
33
+ "metric_list": [
34
+ {
35
+ "metric": "aAcc",
36
+ "aggregation": "<function hb_aggregation_result_aAcc at 0x7fc210e7a9d0>",
37
+ "higher_is_better": true
38
+ },
39
+ {
40
+ "metric": "qAcc",
41
+ "aggregation": "<function hb_aggregation_result_qAcc at 0x7fc210cd3040>",
42
+ "higher_is_better": true
43
+ },
44
+ {
45
+ "metric": "fAcc",
46
+ "aggregation": "<function hb_aggregation_result_fAcc at 0x7fc210cd3820>",
47
+ "higher_is_better": true
48
+ }
49
+ ],
50
+ "output_type": "generate_until",
51
+ "generation_kwargs": {
52
+ "max_new_tokens": 128,
53
+ "temperature": 0.0,
54
+ "top_p": 1.0,
55
+ "num_beams": 1,
56
+ "do_sample": false,
57
+ "until": [
58
+ "\n\n"
59
+ ]
60
+ },
61
+ "repeats": 1,
62
+ "should_decontaminate": false,
63
+ "metadata": [
64
+ {
65
+ "version": 0.0
66
+ }
67
+ ],
68
+ "model_specific_prompt_kwargs": {
69
+ "default": {
70
+ "pre_prompt": "",
71
+ "post_prompt": ""
72
+ }
73
+ }
74
+ },
75
+ "mathvista_testmini": {
76
+ "task": "mathvista_testmini",
77
+ "dataset_path": "AI4Math/MathVista",
78
+ "dataset_kwargs": {
79
+ "token": true
80
+ },
81
+ "test_split": "testmini",
82
+ "doc_to_visual": "<function mathvista_doc_to_visual at 0x7fc1e642faf0>",
83
+ "doc_to_text": "<function mathvista_doc_to_text at 0x7fc1e60281f0>",
84
+ "doc_to_target": "answer",
85
+ "process_results": "<function mathvista_process_results at 0x7fc1e60308b0>",
86
+ "description": "",
87
+ "target_delimiter": " ",
88
+ "fewshot_delimiter": "\n\n",
89
+ "metric_list": [
90
+ {
91
+ "metric": "gpt_eval_score",
92
+ "aggregation": "<function mathvista_aggregate_results at 0x7fc1e603af70>",
93
+ "higher_is_better": true
94
+ }
95
+ ],
96
+ "output_type": "generate_until",
97
+ "generation_kwargs": {
98
+ "until": [
99
+ "ASSISTANT:"
100
+ ],
101
+ "max_new_tokens": 1024,
102
+ "temperature": 0.0,
103
+ "top_p": 1.0,
104
+ "num_beams": 1,
105
+ "do_sample": false,
106
+ "image_aspect_ratio": "original"
107
+ },
108
+ "repeats": 1,
109
+ "should_decontaminate": false,
110
+ "model_specific_prompt_kwargs": {
111
+ "default": {
112
+ "shot_type": "format-prompt",
113
+ "shot": 0,
114
+ "use_caption": false,
115
+ "use_ocr": false
116
+ },
117
+ "phi3v": {
118
+ "shot_type": "solution"
119
+ }
120
+ },
121
+ "model_specific_generation_kwargs": {
122
+ "llava": {
123
+ "image_aspect_ratio": "original"
124
+ }
125
+ }
126
+ }
127
+ },
128
+ "versions": {
129
+ "hallusion_bench_image": "Yaml",
130
+ "mathvista_testmini": "Yaml"
131
+ },
132
+ "n-shot": {
133
+ "hallusion_bench_image": 0,
134
+ "mathvista_testmini": 0
135
+ },
136
+ "model_configs": {
137
+ "model": "llava",
138
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_smoe,conv_template=phi35",
139
+ "batch_size": "1",
140
+ "device": null,
141
+ "limit": null,
142
+ "bootstrap_iters": 100000,
143
+ "gen_kwargs": ""
144
+ },
145
+ "git_hash": "289c7fe5"
146
+ }
sft/1M3/Full_smoe/logs/0626_1044_llava...image_llava_model_args_bc59f1/submissions/mathvista_testmini_scores.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/chartqa.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/results.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "chartqa": {
4
+ "relaxed_overall,none": 0.2184,
5
+ "relaxed_overall_stderr,none": 0.008264859294607817,
6
+ "relaxed_human_split,none": 0.2456,
7
+ "relaxed_human_split_stderr,none": 0.012179621773652779,
8
+ "relaxed_augmented_split,none": 0.1912,
9
+ "relaxed_augmented_split_stderr,none": 0.011127133434932011,
10
+ "alias": "chartqa"
11
+ },
12
+ "seedbench_2_plus": {
13
+ "seedbench_2_plus_Chart,none": 0.5111111111111111,
14
+ "seedbench_2_plus_Chart_stderr,none": "N/A",
15
+ "seedbench_2_plus_all,none": 0.5068072024593764,
16
+ "seedbench_2_plus_all_stderr,none": "N/A",
17
+ "seedbench_2_plus_Web,none": 0.5484848484848485,
18
+ "seedbench_2_plus_Web_stderr,none": "N/A",
19
+ "seedbench_2_plus_Map,none": 0.4684014869888476,
20
+ "seedbench_2_plus_Map_stderr,none": "N/A",
21
+ "alias": "seedbench_2_plus"
22
+ }
23
+ },
24
+ "configs": {
25
+ "chartqa": {
26
+ "task": "chartqa",
27
+ "dataset_path": "lmms-lab/ChartQA",
28
+ "dataset_kwargs": {
29
+ "token": true
30
+ },
31
+ "test_split": "test",
32
+ "doc_to_visual": "<function chartqa_doc_to_visual at 0x7fc70ca3cd30>",
33
+ "doc_to_text": "<function chartqa_doc_to_text at 0x7fc70ca42790>",
34
+ "doc_to_target": "answer",
35
+ "process_results": "<function chartqa_process_results at 0x7fc70ca42a60>",
36
+ "description": "",
37
+ "target_delimiter": " ",
38
+ "fewshot_delimiter": "\n\n",
39
+ "metric_list": [
40
+ {
41
+ "metric": "relaxed_overall",
42
+ "aggregation": "mean",
43
+ "higher_is_better": true
44
+ },
45
+ {
46
+ "metric": "relaxed_human_split",
47
+ "aggregation": "mean",
48
+ "higher_is_better": true
49
+ },
50
+ {
51
+ "metric": "relaxed_augmented_split",
52
+ "aggregation": "mean",
53
+ "higher_is_better": true
54
+ }
55
+ ],
56
+ "output_type": "generate_until",
57
+ "generation_kwargs": {
58
+ "max_new_tokens": 16,
59
+ "temperature": 0.0,
60
+ "do_sample": false,
61
+ "until": [
62
+ "\n\n"
63
+ ]
64
+ },
65
+ "repeats": 1,
66
+ "should_decontaminate": false,
67
+ "metadata": [
68
+ {
69
+ "version": 0.0
70
+ }
71
+ ],
72
+ "model_specific_prompt_kwargs": {
73
+ "default": {
74
+ "pre_prompt": "",
75
+ "post_prompt": "\nAnswer the question with a single word."
76
+ },
77
+ "qwen_vl": {
78
+ "pre_prompt": "",
79
+ "post_prompt": " Answer:"
80
+ }
81
+ }
82
+ },
83
+ "seedbench_2_plus": {
84
+ "task": "seedbench_2_plus",
85
+ "dataset_path": "doolayer/SEED-Bench-2-Plus",
86
+ "dataset_kwargs": {
87
+ "token": true
88
+ },
89
+ "test_split": "test",
90
+ "doc_to_visual": "<function seed_doc_to_visual at 0x7fc7357deca0>",
91
+ "doc_to_text": "<function seed_doc_to_text at 0x7fc735804310>",
92
+ "doc_to_target": "answer",
93
+ "process_results": "<function seed_process_result at 0x7fc735804820>",
94
+ "description": "",
95
+ "target_delimiter": " ",
96
+ "fewshot_delimiter": "\n\n",
97
+ "metric_list": [
98
+ {
99
+ "metric": "seedbench_2_plus_Chart",
100
+ "aggregation": "<function seed_aggregation_result at 0x7fc735804d30>",
101
+ "higher_is_better": true
102
+ },
103
+ {
104
+ "metric": "seedbench_2_plus_Map",
105
+ "aggregation": "<function seed_aggregation_result at 0x7fc7354921f0>",
106
+ "higher_is_better": true
107
+ },
108
+ {
109
+ "metric": "seedbench_2_plus_Web",
110
+ "aggregation": "<function seed_aggregation_result at 0x7fc735492670>",
111
+ "higher_is_better": true
112
+ },
113
+ {
114
+ "metric": "seedbench_2_plus_all",
115
+ "aggregation": "<function seed_aggregation_result at 0x7fc735492af0>",
116
+ "higher_is_better": true
117
+ }
118
+ ],
119
+ "output_type": "generate_until",
120
+ "generation_kwargs": {
121
+ "until": [
122
+ "ASSISTANT:"
123
+ ],
124
+ "max_new_tokens": 16,
125
+ "image_aspect_ratio": "original"
126
+ },
127
+ "repeats": 1,
128
+ "should_decontaminate": false,
129
+ "metadata": [
130
+ {
131
+ "version": 0.0
132
+ }
133
+ ],
134
+ "model_specific_prompt_kwargs": {
135
+ "llava": {
136
+ "img_token": "<image>",
137
+ "post_prompt": "Answer with the option's letter from the given choices directly."
138
+ },
139
+ "gpt4V": {
140
+ "img_token": "<image>",
141
+ "post_prompt": "Answer with the option's letter from the given choices directly."
142
+ },
143
+ "default": {
144
+ "img_token": "<image>",
145
+ "post_prompt": "Answer with the option's letter from the given choices directly."
146
+ }
147
+ }
148
+ }
149
+ },
150
+ "versions": {
151
+ "chartqa": "Yaml",
152
+ "seedbench_2_plus": "Yaml"
153
+ },
154
+ "n-shot": {
155
+ "chartqa": 0,
156
+ "seedbench_2_plus": 0
157
+ },
158
+ "model_configs": {
159
+ "model": "llava",
160
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_smoe,conv_template=phi35",
161
+ "batch_size": "1",
162
+ "device": null,
163
+ "limit": null,
164
+ "bootstrap_iters": 100000,
165
+ "gen_kwargs": ""
166
+ },
167
+ "git_hash": "289c7fe5"
168
+ }
sft/1M3/Full_smoe/logs/0626_1106_llava..._plus_llava_model_args_bc59f1/seedbench_2_plus.json ADDED
The diff for this file is too large to render. See raw diff