Upload folder using huggingface_hub

#69
This view is limited to 50 files because it contains too many changes.  See the raw diff here.
Files changed (50) hide show
  1. .gitattributes +7 -0
  2. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/ai2d.json +0 -0
  3. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmbench_en_dev.json +3 -0
  4. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmerealworld_lite.json +3 -0
  5. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmmu_val.json +0 -0
  6. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmstar.json +0 -0
  7. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/pope.json +3 -0
  8. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank0_metric_eval_done.txt +1 -0
  9. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank1_metric_eval_done.txt +1 -0
  10. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank2_metric_eval_done.txt +1 -0
  11. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank3_metric_eval_done.txt +1 -0
  12. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/results.json +488 -0
  13. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/mmbench_en_dev_results.json +1 -0
  14. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/mmbench_en_dev_results.xlsx +3 -0
  15. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/textvqa_submission_2025-06-20-03-52-19.json +0 -0
  16. sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/textvqa_val.json +3 -0
  17. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/docvqa_val.json +0 -0
  18. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/mme.json +0 -0
  19. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank0_metric_eval_done.txt +1 -0
  20. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank1_metric_eval_done.txt +1 -0
  21. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank2_metric_eval_done.txt +1 -0
  22. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank3_metric_eval_done.txt +1 -0
  23. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/results.json +215 -0
  24. sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/scienceqa_img.json +0 -0
  25. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/gpt_response/hallusion_output_vd_model.json +0 -0
  26. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/gpt_response/hallusion_output_vs_model.json +0 -0
  27. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/hallusion_bench_image.json +0 -0
  28. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/mathvista_testmini.json +0 -0
  29. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank0_metric_eval_done.txt +1 -0
  30. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank1_metric_eval_done.txt +1 -0
  31. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank2_metric_eval_done.txt +1 -0
  32. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank3_metric_eval_done.txt +1 -0
  33. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/results.json +146 -0
  34. sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/submissions/mathvista_testmini_scores.json +0 -0
  35. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/infovqa_val.json +3 -0
  36. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/mmmu_pro_standard.json +0 -0
  37. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/mmmu_pro_vision.json +0 -0
  38. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank0_metric_eval_done.txt +1 -0
  39. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank1_metric_eval_done.txt +1 -0
  40. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank2_metric_eval_done.txt +1 -0
  41. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank3_metric_eval_done.txt +1 -0
  42. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/realworldqa.json +0 -0
  43. sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/results.json +245 -0
  44. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/chartqa.json +0 -0
  45. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank0_metric_eval_done.txt +1 -0
  46. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank1_metric_eval_done.txt +1 -0
  47. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank2_metric_eval_done.txt +1 -0
  48. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank3_metric_eval_done.txt +1 -0
  49. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/results.json +168 -0
  50. sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/seedbench_2_plus.json +0 -0
.gitattributes CHANGED
@@ -177,3 +177,10 @@ sft/1M3/std_0.002_Full_smoe/checkpoint-27572/trainer_state.json filter=lfs diff=
177
  sft/1M3/std_0.002_Full_smoe/checkpoint-34462/trainer_state.json filter=lfs diff=lfs merge=lfs -text
178
  sft/1M3/std_0.002_Full_smoe/checkpoint-20679/trainer_state.json filter=lfs diff=lfs merge=lfs -text
179
  sft/1M3/Full_remoe/checkpoint-13786/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
177
  sft/1M3/std_0.002_Full_smoe/checkpoint-34462/trainer_state.json filter=lfs diff=lfs merge=lfs -text
178
  sft/1M3/std_0.002_Full_smoe/checkpoint-20679/trainer_state.json filter=lfs diff=lfs merge=lfs -text
179
  sft/1M3/Full_remoe/checkpoint-13786/trainer_state.json filter=lfs diff=lfs merge=lfs -text
180
+ sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
181
+ sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
182
+ sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/pope.json filter=lfs diff=lfs merge=lfs -text
183
+ sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
184
+ sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
185
+ sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
186
+ sft/1M3/Full_remoe/logs/0626_1526_llava_v1.5_gqa_llava_model_args_5c6951/gqa.json filter=lfs diff=lfs merge=lfs -text
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7e338d074e23e1872cddf1c3265f03335be0392eb8be0887b57239861bb427f
3
+ size 14568172
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d5c161d4829e2d2a8153ac2476056de425f290323213bd691fd720447d34291
3
+ size 1994104219
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/pope.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a102e9d8f0408982a190bb416b7412f87d6753a07de398c912fe57f79d72a26a
3
+ size 17631894
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/results.json ADDED
@@ -0,0 +1,488 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6386010362694301,
5
+ "exact_match_stderr,flexible-extract": 0.008646492043965488,
6
+ "alias": "ai2d"
7
+ },
8
+ "mmbench_en_dev": {
9
+ "gpt_eval_score,none": 56.87285223367697,
10
+ "gpt_eval_score_stderr,none": "N/A",
11
+ "submission,none": null,
12
+ "submission_stderr,none": "N/A",
13
+ "alias": "mmbench_en_dev"
14
+ },
15
+ "mmerealworld_lite": {
16
+ "mme_realworld_score,none": 0.266284523189161,
17
+ "mme_realworld_score_stderr,none": "N/A",
18
+ "alias": "mmerealworld_lite"
19
+ },
20
+ "mmmu_val": {
21
+ "mmmu_acc,none": 0.40778,
22
+ "mmmu_acc_stderr,none": "N/A",
23
+ "alias": "mmmu_val"
24
+ },
25
+ "mmstar": {
26
+ "coarse perception,none": 0.527471270436267,
27
+ "coarse perception_stderr,none": "N/A",
28
+ "fine-grained perception,none": 0.25409604519774015,
29
+ "fine-grained perception_stderr,none": "N/A",
30
+ "instance reasoning,none": 0.4326050159685535,
31
+ "instance reasoning_stderr,none": "N/A",
32
+ "logical reasoning,none": 0.32938293829382936,
33
+ "logical reasoning_stderr,none": "N/A",
34
+ "math,none": 0.3639902432504678,
35
+ "math_stderr,none": "N/A",
36
+ "science & technology,none": 0.22466506472791004,
37
+ "science & technology_stderr,none": "N/A",
38
+ "alias": "mmstar"
39
+ },
40
+ "pope": {
41
+ "pope_accuracy,none": 0.11666666666666667,
42
+ "pope_accuracy_stderr,none": "N/A",
43
+ "pope_precision,none": 0.7718808193668529,
44
+ "pope_precision_stderr,none": "N/A",
45
+ "pope_recall,none": 0.9200887902330743,
46
+ "pope_recall_stderr,none": "N/A",
47
+ "pope_f1_score,none": 0.839493670886076,
48
+ "pope_f1_score_stderr,none": "N/A",
49
+ "pope_yes_ratio,none": 0.5,
50
+ "pope_yes_ratio_stderr,none": "N/A",
51
+ "alias": "pope"
52
+ },
53
+ "textvqa_val": {
54
+ "exact_match,none": 0.228200000166893,
55
+ "exact_match_stderr,none": 0.005759964181277439,
56
+ "submission,none": null,
57
+ "submission_stderr,none": "N/A",
58
+ "alias": "textvqa_val"
59
+ }
60
+ },
61
+ "configs": {
62
+ "ai2d": {
63
+ "task": "ai2d",
64
+ "dataset_path": "lmms-lab/ai2d",
65
+ "dataset_kwargs": {
66
+ "token": true
67
+ },
68
+ "test_split": "test",
69
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7fd1204d2670>",
70
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7fd1204d2940>",
71
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7fd1204d2ee0>",
72
+ "description": "",
73
+ "target_delimiter": " ",
74
+ "fewshot_delimiter": "\n\n",
75
+ "metric_list": [
76
+ {
77
+ "metric": "exact_match",
78
+ "aggregation": "mean",
79
+ "higher_is_better": true,
80
+ "ignore_case": true,
81
+ "ignore_punctuation": true
82
+ }
83
+ ],
84
+ "output_type": "generate_until",
85
+ "generation_kwargs": {
86
+ "max_new_tokens": 512,
87
+ "temperature": 0.0,
88
+ "do_sample": false,
89
+ "until": [
90
+ "\n\n"
91
+ ]
92
+ },
93
+ "repeats": 1,
94
+ "filter_list": [
95
+ {
96
+ "name": "flexible-extract",
97
+ "filter": [
98
+ {
99
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
100
+ "group_select": 0,
101
+ "ignore_case": true,
102
+ "ignore_punctuation": true,
103
+ "regex_pattern": "([A-Z])\\."
104
+ }
105
+ ]
106
+ }
107
+ ],
108
+ "should_decontaminate": false,
109
+ "metadata": [
110
+ {
111
+ "version": 0.0
112
+ }
113
+ ],
114
+ "model_specific_prompt_kwargs": {
115
+ "default": {
116
+ "prompt_format": "mcq",
117
+ "pre_prompt": "",
118
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
119
+ },
120
+ "gpt4v": {
121
+ "prompt_format": "mcq",
122
+ "pre_prompt": "",
123
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
124
+ },
125
+ "qwen_vl": {
126
+ "prompt_format": "qa",
127
+ "pre_prompt": "",
128
+ "post_prompt": " Answer:"
129
+ },
130
+ "xcomposer2_4khd": {
131
+ "prompt_format": "mcq_xcomposer",
132
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
133
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
134
+ }
135
+ },
136
+ "model_specific_target_kwargs": {
137
+ "default": "mcq",
138
+ "qwen_vl": "qa"
139
+ }
140
+ },
141
+ "mmbench_en_dev": {
142
+ "task": "mmbench_en_dev",
143
+ "dataset_path": "lmms-lab/MMBench",
144
+ "dataset_name": "en",
145
+ "dataset_kwargs": {
146
+ "token": true
147
+ },
148
+ "test_split": "dev",
149
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7fd18005aee0>",
150
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7fd180066430>",
151
+ "doc_to_target": "answer",
152
+ "process_results": "<function mmbench_process_results at 0x7fd180066940>",
153
+ "description": "",
154
+ "target_delimiter": " ",
155
+ "fewshot_delimiter": "\n\n",
156
+ "metric_list": [
157
+ {
158
+ "metric": "gpt_eval_score",
159
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7fd18005a790>",
160
+ "higher_is_better": true
161
+ },
162
+ {
163
+ "metric": "submission",
164
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7fd18005aca0>",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "generate_until",
169
+ "generation_kwargs": {
170
+ "until": [
171
+ "ASSISTANT:"
172
+ ],
173
+ "max_new_tokens": 1024,
174
+ "temperature": 0.0,
175
+ "top_p": 1.0,
176
+ "num_beams": 1,
177
+ "do_sample": false,
178
+ "image_aspect_ratio": "original"
179
+ },
180
+ "repeats": 1,
181
+ "should_decontaminate": false,
182
+ "model_specific_prompt_kwargs": {
183
+ "default": {
184
+ "pre_prompt": "",
185
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
186
+ }
187
+ },
188
+ "model_specific_generation_kwargs": {
189
+ "llava": {
190
+ "image_aspect_ratio": "original"
191
+ }
192
+ }
193
+ },
194
+ "mmerealworld_lite": {
195
+ "task": "mmerealworld_lite",
196
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
197
+ "dataset_kwargs": {
198
+ "token": true
199
+ },
200
+ "test_split": "train",
201
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7fd120c7b820>",
202
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7fd120c80310>",
203
+ "doc_to_target": "answer",
204
+ "process_results": "<function mme_realworld_process_results at 0x7fd120c80f70>",
205
+ "description": "",
206
+ "target_delimiter": " ",
207
+ "fewshot_delimiter": "\n\n",
208
+ "metric_list": [
209
+ {
210
+ "metric": "mme_realworld_score",
211
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7fd120c84af0>",
212
+ "higher_is_better": true
213
+ }
214
+ ],
215
+ "output_type": "generate_until",
216
+ "generation_kwargs": {
217
+ "max_new_tokens": 16,
218
+ "temperature": 0.0,
219
+ "top_p": 1.0,
220
+ "num_beams": 1,
221
+ "do_sample": false,
222
+ "until": [
223
+ "\n\n"
224
+ ]
225
+ },
226
+ "repeats": 1,
227
+ "should_decontaminate": false,
228
+ "metadata": [
229
+ {
230
+ "version": 0.0
231
+ }
232
+ ],
233
+ "model_specific_prompt_kwargs": {
234
+ "default": {
235
+ "pre_prompt": "",
236
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
237
+ },
238
+ "gpt4v": {
239
+ "pre_prompt": "",
240
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
241
+ },
242
+ "xcomposer2_4khd": {
243
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
244
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
245
+ }
246
+ }
247
+ },
248
+ "mmmu_val": {
249
+ "task": "mmmu_val",
250
+ "dataset_path": "lmms-lab/MMMU",
251
+ "test_split": "validation",
252
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7fd133c7d280>",
253
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7fd133c7df70>",
254
+ "doc_to_target": "answer",
255
+ "process_results": "<function mmmu_process_results at 0x7fd133aeae50>",
256
+ "description": "",
257
+ "target_delimiter": " ",
258
+ "fewshot_delimiter": "\n\n",
259
+ "metric_list": [
260
+ {
261
+ "metric": "mmmu_acc",
262
+ "aggregation": "<function mmmu_aggregate_results at 0x7fd133abddc0>",
263
+ "higher_is_better": true
264
+ }
265
+ ],
266
+ "output_type": "generate_until",
267
+ "generation_kwargs": {
268
+ "max_new_tokens": 128,
269
+ "until": [
270
+ "\n\n"
271
+ ],
272
+ "image_aspect_ratio": "original"
273
+ },
274
+ "repeats": 1,
275
+ "should_decontaminate": false,
276
+ "metadata": [
277
+ {
278
+ "version": 0.0
279
+ }
280
+ ],
281
+ "model_specific_generation_kwargs": {
282
+ "llava": {
283
+ "image_aspect_ratio": "original"
284
+ }
285
+ }
286
+ },
287
+ "mmstar": {
288
+ "task": "mmstar",
289
+ "dataset_path": "Lin-Chen/MMStar",
290
+ "dataset_kwargs": {
291
+ "token": true
292
+ },
293
+ "test_split": "val",
294
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7fd12042dca0>",
295
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7fd120432160>",
296
+ "doc_to_target": "answer",
297
+ "process_results": "<function mmstar_process_results at 0x7fd120432670>",
298
+ "description": "",
299
+ "target_delimiter": " ",
300
+ "fewshot_delimiter": "\n\n",
301
+ "metric_list": [
302
+ {
303
+ "metric": "coarse perception",
304
+ "aggregation": "<function mmstar_aggregate_results at 0x7fd120432af0>",
305
+ "higher_is_better": true
306
+ },
307
+ {
308
+ "metric": "fine-grained perception",
309
+ "aggregation": "<function mmstar_aggregate_results at 0x7fd120432ee0>",
310
+ "higher_is_better": true
311
+ },
312
+ {
313
+ "metric": "instance reasoning",
314
+ "aggregation": "<function mmstar_aggregate_results at 0x7fd120437310>",
315
+ "higher_is_better": true
316
+ },
317
+ {
318
+ "metric": "logical reasoning",
319
+ "aggregation": "<function mmstar_aggregate_results at 0x7fd120437700>",
320
+ "higher_is_better": true
321
+ },
322
+ {
323
+ "metric": "science & technology",
324
+ "aggregation": "<function mmstar_aggregate_results at 0x7fd120437af0>",
325
+ "higher_is_better": true
326
+ },
327
+ {
328
+ "metric": "math",
329
+ "aggregation": "<function mmstar_aggregate_results at 0x7fd120437ee0>",
330
+ "higher_is_better": true
331
+ }
332
+ ],
333
+ "output_type": "generate_until",
334
+ "generation_kwargs": {
335
+ "until": [
336
+ "\n\n"
337
+ ],
338
+ "do_sample": false
339
+ },
340
+ "repeats": 1,
341
+ "should_decontaminate": false,
342
+ "metadata": [
343
+ {
344
+ "version": 0.0
345
+ }
346
+ ],
347
+ "model_specific_prompt_kwargs": {
348
+ "default": {
349
+ "pre_prompt": "",
350
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
351
+ }
352
+ }
353
+ },
354
+ "pope": {
355
+ "task": "pope",
356
+ "dataset_path": "lmms-lab/POPE",
357
+ "dataset_kwargs": {
358
+ "token": true
359
+ },
360
+ "test_split": "test",
361
+ "doc_to_visual": "<function pope_doc_to_visual at 0x7fd120d14310>",
362
+ "doc_to_text": "<function pope_doc_to_text at 0x7fd120d148b0>",
363
+ "doc_to_target": "answer",
364
+ "process_results": "<function pope_process_results at 0x7fd120d14e50>",
365
+ "description": "",
366
+ "target_delimiter": " ",
367
+ "fewshot_delimiter": "\n\n",
368
+ "metric_list": [
369
+ {
370
+ "metric": "pope_accuracy",
371
+ "aggregation": "<function pope_aggregate_accuracy at 0x7fd120d17430>",
372
+ "higher_is_better": true
373
+ },
374
+ {
375
+ "metric": "pope_precision",
376
+ "aggregation": "<function pope_aggregate_precision at 0x7fd120d179d0>",
377
+ "higher_is_better": true
378
+ },
379
+ {
380
+ "metric": "pope_recall",
381
+ "aggregation": "<function pope_aggregate_recall at 0x7fd120d17f70>",
382
+ "higher_is_better": true
383
+ },
384
+ {
385
+ "metric": "pope_f1_score",
386
+ "aggregation": "<function pope_aggregate_f1_score at 0x7fd120d1e550>",
387
+ "higher_is_better": true
388
+ },
389
+ {
390
+ "metric": "pope_yes_ratio",
391
+ "aggregation": "<function pope_aggregate_yes_ratio at 0x7fd120d1eaf0>",
392
+ "higher_is_better": true
393
+ }
394
+ ],
395
+ "output_type": "generate_until",
396
+ "generation_kwargs": {
397
+ "max_new_tokens": 128,
398
+ "temperature": 0.0,
399
+ "top_p": 1.0,
400
+ "num_beams": 1,
401
+ "do_sample": false,
402
+ "until": [
403
+ "\n\n"
404
+ ]
405
+ },
406
+ "repeats": 1,
407
+ "should_decontaminate": false,
408
+ "metadata": [
409
+ {
410
+ "version": 0.0
411
+ }
412
+ ]
413
+ },
414
+ "textvqa_val": {
415
+ "task": "textvqa_val",
416
+ "dataset_path": "lmms-lab/textvqa",
417
+ "test_split": "validation",
418
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7fd14dfb2310>",
419
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7fd14dfb2790>",
420
+ "doc_to_target": "answer",
421
+ "process_results": "<function textvqa_process_results at 0x7fd14dfb2a60>",
422
+ "description": "",
423
+ "target_delimiter": " ",
424
+ "fewshot_delimiter": "\n\n",
425
+ "metric_list": [
426
+ {
427
+ "metric": "exact_match",
428
+ "aggregation": "mean",
429
+ "higher_is_better": true,
430
+ "ignore_case": true,
431
+ "ignore_punctuation": true
432
+ },
433
+ {
434
+ "metric": "submission",
435
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7fd14dfb20d0>",
436
+ "higher_is_better": true
437
+ }
438
+ ],
439
+ "output_type": "generate_until",
440
+ "generation_kwargs": {
441
+ "until": [
442
+ "ASSISTANT:"
443
+ ]
444
+ },
445
+ "repeats": 1,
446
+ "should_decontaminate": false,
447
+ "model_specific_prompt_kwargs": {
448
+ "default": {
449
+ "pre_prompt": "",
450
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
451
+ "ocr": false
452
+ },
453
+ "qwen_vl": {
454
+ "pre_prompt": "",
455
+ "post_prompt": " Answer:"
456
+ }
457
+ }
458
+ }
459
+ },
460
+ "versions": {
461
+ "ai2d": "Yaml",
462
+ "mmbench_en_dev": "Yaml",
463
+ "mmerealworld_lite": "Yaml",
464
+ "mmmu_val": "Yaml",
465
+ "mmstar": "Yaml",
466
+ "pope": "Yaml",
467
+ "textvqa_val": "Yaml"
468
+ },
469
+ "n-shot": {
470
+ "ai2d": 0,
471
+ "mmbench_en_dev": 0,
472
+ "mmerealworld_lite": 0,
473
+ "mmmu_val": 0,
474
+ "mmstar": 0,
475
+ "pope": 0,
476
+ "textvqa_val": 0
477
+ },
478
+ "model_configs": {
479
+ "model": "llava",
480
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_remoe,conv_template=phi35",
481
+ "batch_size": "1",
482
+ "device": null,
483
+ "limit": null,
484
+ "bootstrap_iters": 100000,
485
+ "gen_kwargs": ""
486
+ },
487
+ "git_hash": "289c7fe5"
488
+ }
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.5687285223367697, "category_acc": {"action_recognition": 0.8333333333333334, "attribute_comparison": 0.5681818181818182, "attribute_recognition": 0.6621621621621622, "celebrity_recognition": 0.5050505050505051, "function_reasoning": 0.7848101265822784, "future_prediction": 0.45, "identity_reasoning": 0.9111111111111111, "image_emotion": 0.6, "image_quality": 0.11320754716981132, "image_scene": 0.8653846153846154, "image_style": 0.5471698113207547, "image_topic": 0.8333333333333334, "nature_relation": 0.4583333333333333, "object_localization": 0.2962962962962963, "ocr": 0.6666666666666666, "physical_property_reasoning": 0.4666666666666667, "physical_relation": 0.5833333333333334, "social_relation": 0.6046511627906976, "spatial_relationship": 0.2, "structuralized_imagetext_understanding": 0.3974358974358974}, "l2_category_acc": {"attribute_reasoning": 0.6934673366834171, "coarse_perception": 0.625, "finegrained_perception (cross-instance)": 0.5524475524475524, "finegrained_perception (instance-level)": 0.5085324232081911, "logic_reasoning": 0.4152542372881356, "relation_reasoning": 0.5391304347826087}}
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e3987dd40b3c51303a503803c88d21648f4b4e77f5160b192a87710d38a5c7
3
+ size 865158
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/submissions/textvqa_submission_2025-06-20-03-52-19.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0620_0432_llava..._pope_llava_model_args_5c6951/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b194e6d65a8f957801b07c572dd505e0a4edb0a4c6d3d18c49c4f89a3fbfcb7a
3
+ size 13193476
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/docvqa_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/mme.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/results.json ADDED
@@ -0,0 +1,215 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "docvqa_val": {
4
+ "anls,none": 0.15797345298186577,
5
+ "anls_stderr,none": 0.0049872282869667045,
6
+ "alias": "docvqa_val"
7
+ },
8
+ "mme": {
9
+ "mme_cognition_score,none": 320.3571428571429,
10
+ "mme_cognition_score_stderr,none": "N/A",
11
+ "mme_percetion_score,none": 1144.748799519808,
12
+ "mme_percetion_score_stderr,none": "N/A",
13
+ "alias": "mme"
14
+ },
15
+ "scienceqa_img": {
16
+ "exact_match,none": 0.7530986613782846,
17
+ "exact_match_stderr,none": 0.009603784276255882,
18
+ "alias": "scienceqa_img"
19
+ }
20
+ },
21
+ "configs": {
22
+ "docvqa_val": {
23
+ "task": "docvqa_val",
24
+ "dataset_path": "lmms-lab/DocVQA",
25
+ "dataset_name": "DocVQA",
26
+ "dataset_kwargs": {
27
+ "token": true
28
+ },
29
+ "test_split": "validation",
30
+ "doc_to_visual": "<function docvqa_doc_to_visual at 0x7f6dbe75e4c0>",
31
+ "doc_to_text": "<function docvqa_doc_to_text at 0x7f6dbe75e820>",
32
+ "doc_to_target": "answers",
33
+ "description": "",
34
+ "target_delimiter": " ",
35
+ "fewshot_delimiter": "\n\n",
36
+ "metric_list": [
37
+ {
38
+ "metric": "anls",
39
+ "aggregation": "mean",
40
+ "higher_is_better": true
41
+ }
42
+ ],
43
+ "output_type": "generate_until",
44
+ "generation_kwargs": {
45
+ "max_new_tokens": 32,
46
+ "temperature": 0.0,
47
+ "do_sample": false,
48
+ "until": [
49
+ "\n\n"
50
+ ]
51
+ },
52
+ "repeats": 1,
53
+ "should_decontaminate": false,
54
+ "model_specific_prompt_kwargs": {
55
+ "default": {
56
+ "pre_prompt": "",
57
+ "post_prompt": "\nAnswer the question using a single word or phrase."
58
+ },
59
+ "qwen_vl": {
60
+ "pre_prompt": "",
61
+ "post_prompt": " Answer:"
62
+ }
63
+ }
64
+ },
65
+ "mme": {
66
+ "task": "mme",
67
+ "dataset_path": "lmms-lab/MME",
68
+ "dataset_kwargs": {
69
+ "token": false
70
+ },
71
+ "test_split": "test",
72
+ "doc_to_visual": "<function mme_doc_to_visual at 0x7f6deacd48b0>",
73
+ "doc_to_text": "<function mme_doc_to_text at 0x7f6de9276160>",
74
+ "doc_to_target": "answer",
75
+ "process_results": "<function mme_process_results at 0x7f6de9276700>",
76
+ "description": "",
77
+ "target_delimiter": " ",
78
+ "fewshot_delimiter": "\n\n",
79
+ "metric_list": [
80
+ {
81
+ "metric": "mme_percetion_score",
82
+ "aggregation": "<function mme_aggregate_results at 0x7f6de9276c10>",
83
+ "higher_is_better": true
84
+ },
85
+ {
86
+ "metric": "mme_cognition_score",
87
+ "aggregation": "<function mme_aggregate_results at 0x7f6de927f0d0>",
88
+ "higher_is_better": true
89
+ }
90
+ ],
91
+ "output_type": "generate_until",
92
+ "generation_kwargs": {
93
+ "max_new_tokens": 16,
94
+ "temperature": 0.0,
95
+ "top_p": 1.0,
96
+ "num_beams": 1,
97
+ "do_sample": false,
98
+ "until": [
99
+ "\n\n"
100
+ ]
101
+ },
102
+ "repeats": 1,
103
+ "should_decontaminate": false,
104
+ "metadata": [
105
+ {
106
+ "version": 0.0
107
+ }
108
+ ],
109
+ "model_specific_prompt_kwargs": {
110
+ "default": {
111
+ "pre_prompt": "",
112
+ "post_prompt": "\nAnswer the question using a single word or phrase."
113
+ },
114
+ "gpt4v": {
115
+ "pre_prompt": "",
116
+ "post_prompt": "\nAnswer the question with Yes or No."
117
+ },
118
+ "qwen_vl": {
119
+ "pre_prompt": "",
120
+ "post_prompt": " Answer:"
121
+ },
122
+ "otterhd": {
123
+ "pre_prompt": "",
124
+ "post_prompt": " Answer:"
125
+ },
126
+ "xcomposer2_4khd": {
127
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
128
+ "post_prompt": " Answer this question briefly[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
129
+ }
130
+ }
131
+ },
132
+ "scienceqa_img": {
133
+ "task": "scienceqa_img",
134
+ "dataset_path": "lmms-lab/ScienceQA",
135
+ "dataset_name": "ScienceQA-IMG",
136
+ "dataset_kwargs": {
137
+ "token": true
138
+ },
139
+ "test_split": "test",
140
+ "doc_to_visual": "<function sqa_doc_to_visual at 0x7f6db4e6fc10>",
141
+ "doc_to_text": "<function sqa_doc_to_text at 0x7f6db4e6fe50>",
142
+ "doc_to_target": "<function sqa_doc_to_target at 0x7f6db4df8280>",
143
+ "process_results": "<function sqa_process_results at 0x7f6db4df85e0>",
144
+ "description": "",
145
+ "target_delimiter": " ",
146
+ "fewshot_delimiter": "\n\n",
147
+ "metric_list": [
148
+ {
149
+ "metric": "exact_match",
150
+ "aggregation": "mean",
151
+ "higher_is_better": true,
152
+ "ignore_case": true,
153
+ "ignore_punctuation": true
154
+ }
155
+ ],
156
+ "output_type": "generate_until",
157
+ "generation_kwargs": {
158
+ "max_new_tokens": 16,
159
+ "temperature": 0.0,
160
+ "do_sample": false,
161
+ "until": [
162
+ "\n\n"
163
+ ],
164
+ "image_aspect_ratio": "original"
165
+ },
166
+ "repeats": 1,
167
+ "should_decontaminate": false,
168
+ "metadata": [
169
+ {
170
+ "version": 0.0
171
+ }
172
+ ],
173
+ "model_specific_prompt_kwargs": {
174
+ "default": {
175
+ "format": "default",
176
+ "pre_prompt": "",
177
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
178
+ },
179
+ "qwen_vl": {
180
+ "format": "qwen_vl"
181
+ },
182
+ "idefics2": {
183
+ "format": "default",
184
+ "pre_prompt": "",
185
+ "post_prompt": "\nAnswer:"
186
+ }
187
+ },
188
+ "model_specific_generation_kwargs": {
189
+ "llava": {
190
+ "image_aspect_ratio": "original"
191
+ }
192
+ }
193
+ }
194
+ },
195
+ "versions": {
196
+ "docvqa_val": "Yaml",
197
+ "mme": "Yaml",
198
+ "scienceqa_img": "Yaml"
199
+ },
200
+ "n-shot": {
201
+ "docvqa_val": 0,
202
+ "mme": 0,
203
+ "scienceqa_img": 0
204
+ },
205
+ "model_configs": {
206
+ "model": "llava",
207
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_remoe,conv_template=phi35",
208
+ "batch_size": "1",
209
+ "device": null,
210
+ "limit": null,
211
+ "bootstrap_iters": 100000,
212
+ "gen_kwargs": ""
213
+ },
214
+ "git_hash": "289c7fe5"
215
+ }
sft/1M3/Full_remoe/logs/0626_1501_llava...l_mme_llava_model_args_5c6951/scienceqa_img.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/gpt_response/hallusion_output_vd_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/gpt_response/hallusion_output_vs_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/hallusion_bench_image.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/mathvista_testmini.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/results.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hallusion_bench_image": {
4
+ "aAcc,none": 44.4795,
5
+ "aAcc_stderr,none": "N/A",
6
+ "fAcc,none": 16.763,
7
+ "fAcc_stderr,none": "N/A",
8
+ "qAcc,none": 11.4286,
9
+ "qAcc_stderr,none": "N/A",
10
+ "alias": "hallusion_bench_image"
11
+ },
12
+ "mathvista_testmini": {
13
+ "gpt_eval_score,none": 34.2,
14
+ "gpt_eval_score_stderr,none": "N/A",
15
+ "alias": "mathvista_testmini"
16
+ }
17
+ },
18
+ "configs": {
19
+ "hallusion_bench_image": {
20
+ "task": "hallusion_bench_image",
21
+ "dataset_path": "lmms-lab/HallusionBench",
22
+ "dataset_kwargs": {
23
+ "token": true
24
+ },
25
+ "test_split": "image",
26
+ "doc_to_visual": "<function hb_doc_to_visual at 0x7fc7f69f60d0>",
27
+ "doc_to_text": "<function hb_doc_to_text at 0x7fc7f69f6820>",
28
+ "doc_to_target": "gt_answer_details",
29
+ "process_results": "<function hb_process_results at 0x7fc7f694c040>",
30
+ "description": "",
31
+ "target_delimiter": " ",
32
+ "fewshot_delimiter": "\n\n",
33
+ "metric_list": [
34
+ {
35
+ "metric": "aAcc",
36
+ "aggregation": "<function hb_aggregation_result_aAcc at 0x7fc7f694c9d0>",
37
+ "higher_is_better": true
38
+ },
39
+ {
40
+ "metric": "qAcc",
41
+ "aggregation": "<function hb_aggregation_result_qAcc at 0x7fc7f68cd040>",
42
+ "higher_is_better": true
43
+ },
44
+ {
45
+ "metric": "fAcc",
46
+ "aggregation": "<function hb_aggregation_result_fAcc at 0x7fc7f68cd820>",
47
+ "higher_is_better": true
48
+ }
49
+ ],
50
+ "output_type": "generate_until",
51
+ "generation_kwargs": {
52
+ "max_new_tokens": 128,
53
+ "temperature": 0.0,
54
+ "top_p": 1.0,
55
+ "num_beams": 1,
56
+ "do_sample": false,
57
+ "until": [
58
+ "\n\n"
59
+ ]
60
+ },
61
+ "repeats": 1,
62
+ "should_decontaminate": false,
63
+ "metadata": [
64
+ {
65
+ "version": 0.0
66
+ }
67
+ ],
68
+ "model_specific_prompt_kwargs": {
69
+ "default": {
70
+ "pre_prompt": "",
71
+ "post_prompt": ""
72
+ }
73
+ }
74
+ },
75
+ "mathvista_testmini": {
76
+ "task": "mathvista_testmini",
77
+ "dataset_path": "AI4Math/MathVista",
78
+ "dataset_kwargs": {
79
+ "token": true
80
+ },
81
+ "test_split": "testmini",
82
+ "doc_to_visual": "<function mathvista_doc_to_visual at 0x7fc7cc111af0>",
83
+ "doc_to_text": "<function mathvista_doc_to_text at 0x7fc7cbd091f0>",
84
+ "doc_to_target": "answer",
85
+ "process_results": "<function mathvista_process_results at 0x7fc7cbd128b0>",
86
+ "description": "",
87
+ "target_delimiter": " ",
88
+ "fewshot_delimiter": "\n\n",
89
+ "metric_list": [
90
+ {
91
+ "metric": "gpt_eval_score",
92
+ "aggregation": "<function mathvista_aggregate_results at 0x7fc7cbd1cf70>",
93
+ "higher_is_better": true
94
+ }
95
+ ],
96
+ "output_type": "generate_until",
97
+ "generation_kwargs": {
98
+ "until": [
99
+ "ASSISTANT:"
100
+ ],
101
+ "max_new_tokens": 1024,
102
+ "temperature": 0.0,
103
+ "top_p": 1.0,
104
+ "num_beams": 1,
105
+ "do_sample": false,
106
+ "image_aspect_ratio": "original"
107
+ },
108
+ "repeats": 1,
109
+ "should_decontaminate": false,
110
+ "model_specific_prompt_kwargs": {
111
+ "default": {
112
+ "shot_type": "format-prompt",
113
+ "shot": 0,
114
+ "use_caption": false,
115
+ "use_ocr": false
116
+ },
117
+ "phi3v": {
118
+ "shot_type": "solution"
119
+ }
120
+ },
121
+ "model_specific_generation_kwargs": {
122
+ "llava": {
123
+ "image_aspect_ratio": "original"
124
+ }
125
+ }
126
+ }
127
+ },
128
+ "versions": {
129
+ "hallusion_bench_image": "Yaml",
130
+ "mathvista_testmini": "Yaml"
131
+ },
132
+ "n-shot": {
133
+ "hallusion_bench_image": 0,
134
+ "mathvista_testmini": 0
135
+ },
136
+ "model_configs": {
137
+ "model": "llava",
138
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_remoe,conv_template=phi35",
139
+ "batch_size": "1",
140
+ "device": null,
141
+ "limit": null,
142
+ "bootstrap_iters": 100000,
143
+ "gen_kwargs": ""
144
+ },
145
+ "git_hash": "289c7fe5"
146
+ }
sft/1M3/Full_remoe/logs/0626_1503_llava...image_llava_model_args_5c6951/submissions/mathvista_testmini_scores.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/infovqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c524c34a69abfc7e6053cd85992721dd5f12b2a266568ed40349392bcf4ff75e
3
+ size 576442330
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/mmmu_pro_standard.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/mmmu_pro_vision.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/realworldqa.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1513_llava...a_val_llava_model_args_5c6951/results.json ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "infovqa_val": {
4
+ "anls,none": 0.2784719742948947,
5
+ "anls_stderr,none": 0.00847107123069025,
6
+ "alias": "infovqa_val"
7
+ },
8
+ "mmmu_pro": {
9
+ "mmmu_acc,none": 0.18988499999999997,
10
+ "mmmu_acc_stderr,none": 0.033387325106623435,
11
+ "alias": "mmmu_pro"
12
+ },
13
+ "mmmu_pro_standard": {
14
+ "mmmu_acc,none": 0.25665,
15
+ "mmmu_acc_stderr,none": "N/A",
16
+ "alias": " - mmmu_pro_standard"
17
+ },
18
+ "mmmu_pro_vision": {
19
+ "mmmu_acc,none": 0.12312,
20
+ "mmmu_acc_stderr,none": "N/A",
21
+ "alias": " - mmmu_pro_vision"
22
+ },
23
+ "realworldqa": {
24
+ "exact_match,flexible-extract": 0.4392156862745098,
25
+ "exact_match_stderr,flexible-extract": 0.017955196829891946,
26
+ "alias": "realworldqa"
27
+ }
28
+ },
29
+ "groups": {
30
+ "mmmu_pro": {
31
+ "mmmu_acc,none": 0.18988499999999997,
32
+ "mmmu_acc_stderr,none": 0.033387325106623435,
33
+ "alias": "mmmu_pro"
34
+ }
35
+ },
36
+ "configs": {
37
+ "infovqa_val": {
38
+ "task": "infovqa_val",
39
+ "dataset_path": "lmms-lab/DocVQA",
40
+ "dataset_name": "InfographicVQA",
41
+ "dataset_kwargs": {
42
+ "token": true
43
+ },
44
+ "test_split": "validation",
45
+ "doc_to_visual": "<function infovqa_doc_to_visual at 0x7f852e796a60>",
46
+ "doc_to_text": "<function infovqa_doc_to_text at 0x7f852e796d30>",
47
+ "doc_to_target": "answers",
48
+ "description": "",
49
+ "target_delimiter": " ",
50
+ "fewshot_delimiter": "\n\n",
51
+ "metric_list": [
52
+ {
53
+ "metric": "anls",
54
+ "aggregation": "mean",
55
+ "higher_is_better": true
56
+ }
57
+ ],
58
+ "output_type": "generate_until",
59
+ "generation_kwargs": {
60
+ "max_new_tokens": 32,
61
+ "temperature": 0.0,
62
+ "do_sample": false,
63
+ "until": [
64
+ "\n\n"
65
+ ]
66
+ },
67
+ "repeats": 1,
68
+ "should_decontaminate": false,
69
+ "model_specific_prompt_kwargs": {
70
+ "default": {
71
+ "pre_prompt": "",
72
+ "post_prompt": "\nAnswer the question using a single word or phrase."
73
+ }
74
+ }
75
+ },
76
+ "mmmu_pro_standard": {
77
+ "task": "mmmu_pro_standard",
78
+ "dataset_path": "MMMU/MMMU_Pro",
79
+ "dataset_name": "standard (10 options)",
80
+ "test_split": "test",
81
+ "doc_to_visual": "<function mmmu_pro_doc_to_visual at 0x7f8556d58790>",
82
+ "doc_to_text": "<function mmmu_pro_doc_to_text at 0x7f8556d62550>",
83
+ "doc_to_target": "{{answer}}",
84
+ "process_results": "<function mmmu_pro_process_results at 0x7f8556d6c4c0>",
85
+ "description": "",
86
+ "target_delimiter": " ",
87
+ "fewshot_delimiter": "\n\n",
88
+ "metric_list": [
89
+ {
90
+ "metric": "mmmu_acc",
91
+ "aggregation": "<function mmmu_pro_aggregate_results at 0x7f8556d76430>",
92
+ "higher_is_better": true
93
+ }
94
+ ],
95
+ "output_type": "generate_until",
96
+ "generation_kwargs": {
97
+ "max_new_tokens": 256,
98
+ "until": [
99
+ "\n\n"
100
+ ]
101
+ },
102
+ "repeats": 1,
103
+ "should_decontaminate": false,
104
+ "metadata": {
105
+ "version": 0.0,
106
+ "interleaved_format": false
107
+ },
108
+ "model_specific_prompt_kwargs": {
109
+ "default": {
110
+ "pre_prompt": "",
111
+ "post_prompt": "Answer with the option letter from the given choices directly."
112
+ }
113
+ }
114
+ },
115
+ "mmmu_pro_vision": {
116
+ "task": "mmmu_pro_vision",
117
+ "dataset_path": "MMMU/MMMU_Pro",
118
+ "dataset_name": "vision",
119
+ "test_split": "test",
120
+ "doc_to_visual": "<function mmmu_pro_doc_to_visual at 0x7f8556d44ca0>",
121
+ "doc_to_text": "Answer with the option letter from the given choices directly.",
122
+ "doc_to_target": "{{answer}}",
123
+ "process_results": "<function mmmu_pro_process_results at 0x7f8556d4db80>",
124
+ "description": "",
125
+ "target_delimiter": " ",
126
+ "fewshot_delimiter": "\n\n",
127
+ "metric_list": [
128
+ {
129
+ "metric": "mmmu_acc",
130
+ "aggregation": "<function mmmu_pro_aggregate_results at 0x7f8556d54af0>",
131
+ "higher_is_better": true
132
+ }
133
+ ],
134
+ "output_type": "generate_until",
135
+ "generation_kwargs": {
136
+ "max_new_tokens": 256,
137
+ "until": [
138
+ "\n\n"
139
+ ]
140
+ },
141
+ "repeats": 1,
142
+ "should_decontaminate": false,
143
+ "metadata": {
144
+ "version": 0.0,
145
+ "interleaved_format": false
146
+ }
147
+ },
148
+ "realworldqa": {
149
+ "task": "realworldqa",
150
+ "dataset_path": "lmms-lab/RealWorldQA",
151
+ "dataset_kwargs": {
152
+ "token": true
153
+ },
154
+ "test_split": "test",
155
+ "doc_to_visual": "<function realworldqa_doc_to_visual at 0x7f8526d32af0>",
156
+ "doc_to_text": "<function realworldqa_doc_to_text at 0x7f8526c43160>",
157
+ "doc_to_target": "answer",
158
+ "description": "",
159
+ "target_delimiter": " ",
160
+ "fewshot_delimiter": "\n\n",
161
+ "metric_list": [
162
+ {
163
+ "metric": "exact_match",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true,
166
+ "ignore_case": true,
167
+ "ignore_punctuation": true
168
+ }
169
+ ],
170
+ "output_type": "generate_until",
171
+ "generation_kwargs": {
172
+ "max_new_tokens": 16,
173
+ "temperature": 0.0,
174
+ "top_p": 1.0,
175
+ "num_beams": 1,
176
+ "do_sample": false,
177
+ "until": [
178
+ "\n\n"
179
+ ]
180
+ },
181
+ "repeats": 1,
182
+ "filter_list": [
183
+ {
184
+ "name": "flexible-extract",
185
+ "filter": [
186
+ {
187
+ "function": "<class 'utils.NumberWordsToDigitsFilter'>"
188
+ },
189
+ {
190
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
191
+ "group_select": 0,
192
+ "ignore_case": true,
193
+ "ignore_punctuation": true,
194
+ "regex_pattern": "(\\([A-Z]\\))"
195
+ }
196
+ ]
197
+ }
198
+ ],
199
+ "should_decontaminate": false,
200
+ "metadata": [
201
+ {
202
+ "version": 0.0
203
+ }
204
+ ],
205
+ "model_specific_prompt_kwargs": {
206
+ "default": {
207
+ "pre_prompt": "",
208
+ "post_prompt": ""
209
+ },
210
+ "gpt4v": {
211
+ "pre_prompt": "",
212
+ "post_prompt": ""
213
+ },
214
+ "xcomposer2_4khd": {
215
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
216
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
217
+ }
218
+ }
219
+ }
220
+ },
221
+ "versions": {
222
+ "infovqa_val": "Yaml",
223
+ "mmmu_pro": "N/A",
224
+ "mmmu_pro_standard": "Yaml",
225
+ "mmmu_pro_vision": "Yaml",
226
+ "realworldqa": "Yaml"
227
+ },
228
+ "n-shot": {
229
+ "infovqa_val": 0,
230
+ "mmmu_pro": 0,
231
+ "mmmu_pro_standard": 0,
232
+ "mmmu_pro_vision": 0,
233
+ "realworldqa": 0
234
+ },
235
+ "model_configs": {
236
+ "model": "llava",
237
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_remoe,conv_template=phi35",
238
+ "batch_size": "1",
239
+ "device": null,
240
+ "limit": null,
241
+ "bootstrap_iters": 100000,
242
+ "gen_kwargs": ""
243
+ },
244
+ "git_hash": "289c7fe5"
245
+ }
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/chartqa.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/results.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "chartqa": {
4
+ "relaxed_overall,none": 0.124,
5
+ "relaxed_overall_stderr,none": 0.006592949779478021,
6
+ "relaxed_human_split,none": 0.1296,
7
+ "relaxed_human_split_stderr,none": 0.009503437497001176,
8
+ "relaxed_augmented_split,none": 0.1184,
9
+ "relaxed_augmented_split_stderr,none": 0.009141772782545212,
10
+ "alias": "chartqa"
11
+ },
12
+ "seedbench_2_plus": {
13
+ "seedbench_2_plus_Chart,none": 0.4728395061728395,
14
+ "seedbench_2_plus_Chart_stderr,none": "N/A",
15
+ "seedbench_2_plus_all,none": 0.48177426438296006,
16
+ "seedbench_2_plus_all_stderr,none": "N/A",
17
+ "seedbench_2_plus_Web,none": 0.5272727272727272,
18
+ "seedbench_2_plus_Web_stderr,none": "N/A",
19
+ "seedbench_2_plus_Map,none": 0.45353159851301117,
20
+ "seedbench_2_plus_Map_stderr,none": "N/A",
21
+ "alias": "seedbench_2_plus"
22
+ }
23
+ },
24
+ "configs": {
25
+ "chartqa": {
26
+ "task": "chartqa",
27
+ "dataset_path": "lmms-lab/ChartQA",
28
+ "dataset_kwargs": {
29
+ "token": true
30
+ },
31
+ "test_split": "test",
32
+ "doc_to_visual": "<function chartqa_doc_to_visual at 0x7f538db64d30>",
33
+ "doc_to_text": "<function chartqa_doc_to_text at 0x7f538db6d790>",
34
+ "doc_to_target": "answer",
35
+ "process_results": "<function chartqa_process_results at 0x7f538db6da60>",
36
+ "description": "",
37
+ "target_delimiter": " ",
38
+ "fewshot_delimiter": "\n\n",
39
+ "metric_list": [
40
+ {
41
+ "metric": "relaxed_overall",
42
+ "aggregation": "mean",
43
+ "higher_is_better": true
44
+ },
45
+ {
46
+ "metric": "relaxed_human_split",
47
+ "aggregation": "mean",
48
+ "higher_is_better": true
49
+ },
50
+ {
51
+ "metric": "relaxed_augmented_split",
52
+ "aggregation": "mean",
53
+ "higher_is_better": true
54
+ }
55
+ ],
56
+ "output_type": "generate_until",
57
+ "generation_kwargs": {
58
+ "max_new_tokens": 16,
59
+ "temperature": 0.0,
60
+ "do_sample": false,
61
+ "until": [
62
+ "\n\n"
63
+ ]
64
+ },
65
+ "repeats": 1,
66
+ "should_decontaminate": false,
67
+ "metadata": [
68
+ {
69
+ "version": 0.0
70
+ }
71
+ ],
72
+ "model_specific_prompt_kwargs": {
73
+ "default": {
74
+ "pre_prompt": "",
75
+ "post_prompt": "\nAnswer the question with a single word."
76
+ },
77
+ "qwen_vl": {
78
+ "pre_prompt": "",
79
+ "post_prompt": " Answer:"
80
+ }
81
+ }
82
+ },
83
+ "seedbench_2_plus": {
84
+ "task": "seedbench_2_plus",
85
+ "dataset_path": "doolayer/SEED-Bench-2-Plus",
86
+ "dataset_kwargs": {
87
+ "token": true
88
+ },
89
+ "test_split": "test",
90
+ "doc_to_visual": "<function seed_doc_to_visual at 0x7f53b6357ca0>",
91
+ "doc_to_text": "<function seed_doc_to_text at 0x7f53b637e310>",
92
+ "doc_to_target": "answer",
93
+ "process_results": "<function seed_process_result at 0x7f53b637e820>",
94
+ "description": "",
95
+ "target_delimiter": " ",
96
+ "fewshot_delimiter": "\n\n",
97
+ "metric_list": [
98
+ {
99
+ "metric": "seedbench_2_plus_Chart",
100
+ "aggregation": "<function seed_aggregation_result at 0x7f53b637ed30>",
101
+ "higher_is_better": true
102
+ },
103
+ {
104
+ "metric": "seedbench_2_plus_Map",
105
+ "aggregation": "<function seed_aggregation_result at 0x7f53b61231f0>",
106
+ "higher_is_better": true
107
+ },
108
+ {
109
+ "metric": "seedbench_2_plus_Web",
110
+ "aggregation": "<function seed_aggregation_result at 0x7f53b6123670>",
111
+ "higher_is_better": true
112
+ },
113
+ {
114
+ "metric": "seedbench_2_plus_all",
115
+ "aggregation": "<function seed_aggregation_result at 0x7f53b6123af0>",
116
+ "higher_is_better": true
117
+ }
118
+ ],
119
+ "output_type": "generate_until",
120
+ "generation_kwargs": {
121
+ "until": [
122
+ "ASSISTANT:"
123
+ ],
124
+ "max_new_tokens": 16,
125
+ "image_aspect_ratio": "original"
126
+ },
127
+ "repeats": 1,
128
+ "should_decontaminate": false,
129
+ "metadata": [
130
+ {
131
+ "version": 0.0
132
+ }
133
+ ],
134
+ "model_specific_prompt_kwargs": {
135
+ "llava": {
136
+ "img_token": "<image>",
137
+ "post_prompt": "Answer with the option's letter from the given choices directly."
138
+ },
139
+ "gpt4V": {
140
+ "img_token": "<image>",
141
+ "post_prompt": "Answer with the option's letter from the given choices directly."
142
+ },
143
+ "default": {
144
+ "img_token": "<image>",
145
+ "post_prompt": "Answer with the option's letter from the given choices directly."
146
+ }
147
+ }
148
+ }
149
+ },
150
+ "versions": {
151
+ "chartqa": "Yaml",
152
+ "seedbench_2_plus": "Yaml"
153
+ },
154
+ "n-shot": {
155
+ "chartqa": 0,
156
+ "seedbench_2_plus": 0
157
+ },
158
+ "model_configs": {
159
+ "model": "llava",
160
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_remoe,conv_template=phi35",
161
+ "batch_size": "1",
162
+ "device": null,
163
+ "limit": null,
164
+ "bootstrap_iters": 100000,
165
+ "gen_kwargs": ""
166
+ },
167
+ "git_hash": "289c7fe5"
168
+ }
sft/1M3/Full_remoe/logs/0626_1520_llava..._plus_llava_model_args_5c6951/seedbench_2_plus.json ADDED
The diff for this file is too large to render. See raw diff