Upload folder using huggingface_hub

#108
This view is limited to 50 files because it contains too many changes.  See the raw diff here.
Files changed (50) hide show
  1. .gitattributes +10 -0
  2. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/gpt_response/hallusion_output_vd_model.json +0 -0
  3. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/gpt_response/hallusion_output_vs_model.json +0 -0
  4. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/hallusion_bench_image.json +0 -0
  5. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/mathvista_testmini.json +0 -0
  6. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/rank0_metric_eval_done.txt +1 -0
  7. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/results.json +146 -0
  8. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/submissions/mathvista_testmini_scores.json +0 -0
  9. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/ai2d.json +0 -0
  10. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/infovqa_val.json +3 -0
  11. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmbench_en_dev.json +3 -0
  12. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mme.json +0 -0
  13. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmerealworld_lite.json +3 -0
  14. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmmu_val.json +0 -0
  15. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmstar.json +0 -0
  16. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/rank0_metric_eval_done.txt +1 -0
  17. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/rank1_metric_eval_done.txt +1 -0
  18. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/rank2_metric_eval_done.txt +1 -0
  19. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/results.json +683 -0
  20. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/scienceqa_img.json +0 -0
  21. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/seedbench_2_plus.json +0 -0
  22. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.json +1 -0
  23. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.xlsx +3 -0
  24. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/textvqa_submission_2025-06-28-17-54-58.json +0 -0
  25. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/textvqa_val.json +3 -0
  26. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1855_llava...bench_llava_model_args_1b5567/rank0_metric_eval_done.txt +1 -0
  27. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1855_llava...bench_llava_model_args_1b5567/results/ocrbench_results.txt +18 -0
  28. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1856_llava...bench_llava_model_args_1b5567/rank1_metric_eval_done.txt +1 -0
  29. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1856_llava...bench_llava_model_args_1b5567/rank2_metric_eval_done.txt +1 -0
  30. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/ai2d.json +0 -0
  31. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/infovqa_val.json +3 -0
  32. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmbench_en_dev.json +3 -0
  33. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mme.json +0 -0
  34. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmerealworld_lite.json +3 -0
  35. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmmu_val.json +0 -0
  36. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmstar.json +0 -0
  37. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank0_metric_eval_done.txt +1 -0
  38. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank1_metric_eval_done.txt +1 -0
  39. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank2_metric_eval_done.txt +1 -0
  40. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank3_metric_eval_done.txt +1 -0
  41. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/results.json +683 -0
  42. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/scienceqa_img.json +0 -0
  43. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/seedbench_2_plus.json +0 -0
  44. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.json +1 -0
  45. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.xlsx +3 -0
  46. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/textvqa_submission_2025-06-28-20-16-15.json +0 -0
  47. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/textvqa_val.json +3 -0
  48. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2117_llava...bench_llava_model_args_1b5567/ocrbench.json +0 -0
  49. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2117_llava...bench_llava_model_args_1b5567/rank0_metric_eval_done.txt +1 -0
  50. sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2117_llava...bench_llava_model_args_1b5567/rank1_metric_eval_done.txt +1 -0
.gitattributes CHANGED
@@ -277,3 +277,13 @@ sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args
277
  sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args_7e5f99/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
278
  sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args_7e5f99/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
279
  sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args_7e5f99/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
277
  sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args_7e5f99/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
278
  sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args_7e5f99/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
279
  sft/1M3/Full_xmoe/checkpoint-27572/logs/0628_2239_llava...l_mme_llava_model_args_7e5f99/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
280
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
281
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
282
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
283
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
284
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
285
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
286
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
287
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
288
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
289
+ sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/gpt_response/hallusion_output_vd_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/gpt_response/hallusion_output_vs_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/hallusion_bench_image.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/mathvista_testmini.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/results.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hallusion_bench_image": {
4
+ "aAcc,none": 45.2156,
5
+ "aAcc_stderr,none": "N/A",
6
+ "fAcc,none": 19.6532,
7
+ "fAcc_stderr,none": "N/A",
8
+ "qAcc,none": 16.044,
9
+ "qAcc_stderr,none": "N/A",
10
+ "alias": "hallusion_bench_image"
11
+ },
12
+ "mathvista_testmini": {
13
+ "gpt_eval_score,none": 35.9,
14
+ "gpt_eval_score_stderr,none": "N/A",
15
+ "alias": "mathvista_testmini"
16
+ }
17
+ },
18
+ "configs": {
19
+ "hallusion_bench_image": {
20
+ "task": "hallusion_bench_image",
21
+ "dataset_path": "lmms-lab/HallusionBench",
22
+ "dataset_kwargs": {
23
+ "token": true
24
+ },
25
+ "test_split": "image",
26
+ "doc_to_visual": "<function hb_doc_to_visual at 0x7f5dd3c6c160>",
27
+ "doc_to_text": "<function hb_doc_to_text at 0x7f5dd3c6c8b0>",
28
+ "doc_to_target": "gt_answer_details",
29
+ "process_results": "<function hb_process_results at 0x7f5dd3c1e0d0>",
30
+ "description": "",
31
+ "target_delimiter": " ",
32
+ "fewshot_delimiter": "\n\n",
33
+ "metric_list": [
34
+ {
35
+ "metric": "aAcc",
36
+ "aggregation": "<function hb_aggregation_result_aAcc at 0x7f5dd3c1ea60>",
37
+ "higher_is_better": true
38
+ },
39
+ {
40
+ "metric": "qAcc",
41
+ "aggregation": "<function hb_aggregation_result_qAcc at 0x7f5dd3bfc0d0>",
42
+ "higher_is_better": true
43
+ },
44
+ {
45
+ "metric": "fAcc",
46
+ "aggregation": "<function hb_aggregation_result_fAcc at 0x7f5dd3bfc8b0>",
47
+ "higher_is_better": true
48
+ }
49
+ ],
50
+ "output_type": "generate_until",
51
+ "generation_kwargs": {
52
+ "max_new_tokens": 128,
53
+ "temperature": 0.0,
54
+ "top_p": 1.0,
55
+ "num_beams": 1,
56
+ "do_sample": false,
57
+ "until": [
58
+ "\n\n"
59
+ ]
60
+ },
61
+ "repeats": 1,
62
+ "should_decontaminate": false,
63
+ "metadata": [
64
+ {
65
+ "version": 0.0
66
+ }
67
+ ],
68
+ "model_specific_prompt_kwargs": {
69
+ "default": {
70
+ "pre_prompt": "",
71
+ "post_prompt": ""
72
+ }
73
+ }
74
+ },
75
+ "mathvista_testmini": {
76
+ "task": "mathvista_testmini",
77
+ "dataset_path": "AI4Math/MathVista",
78
+ "dataset_kwargs": {
79
+ "token": true
80
+ },
81
+ "test_split": "testmini",
82
+ "doc_to_visual": "<function mathvista_doc_to_visual at 0x7f5da8186c10>",
83
+ "doc_to_text": "<function mathvista_doc_to_text at 0x7f5da8191310>",
84
+ "doc_to_target": "answer",
85
+ "process_results": "<function mathvista_process_results at 0x7f5da7d8a9d0>",
86
+ "description": "",
87
+ "target_delimiter": " ",
88
+ "fewshot_delimiter": "\n\n",
89
+ "metric_list": [
90
+ {
91
+ "metric": "gpt_eval_score",
92
+ "aggregation": "<function mathvista_aggregate_results at 0x7f5da7d950d0>",
93
+ "higher_is_better": true
94
+ }
95
+ ],
96
+ "output_type": "generate_until",
97
+ "generation_kwargs": {
98
+ "until": [
99
+ "ASSISTANT:"
100
+ ],
101
+ "max_new_tokens": 1024,
102
+ "temperature": 0.0,
103
+ "top_p": 1.0,
104
+ "num_beams": 1,
105
+ "do_sample": false,
106
+ "image_aspect_ratio": "original"
107
+ },
108
+ "repeats": 1,
109
+ "should_decontaminate": false,
110
+ "model_specific_prompt_kwargs": {
111
+ "default": {
112
+ "shot_type": "format-prompt",
113
+ "shot": 0,
114
+ "use_caption": false,
115
+ "use_ocr": false
116
+ },
117
+ "phi3v": {
118
+ "shot_type": "solution"
119
+ }
120
+ },
121
+ "model_specific_generation_kwargs": {
122
+ "llava": {
123
+ "image_aspect_ratio": "original"
124
+ }
125
+ }
126
+ }
127
+ },
128
+ "versions": {
129
+ "hallusion_bench_image": "Yaml",
130
+ "mathvista_testmini": "Yaml"
131
+ },
132
+ "n-shot": {
133
+ "hallusion_bench_image": 0,
134
+ "mathvista_testmini": 0
135
+ },
136
+ "model_configs": {
137
+ "model": "llava",
138
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_xmoe/checkpoint-6893,conv_template=phi35",
139
+ "batch_size": "1",
140
+ "device": null,
141
+ "limit": null,
142
+ "bootstrap_iters": 100000,
143
+ "gen_kwargs": ""
144
+ },
145
+ "git_hash": "289c7fe5"
146
+ }
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1733_llava...image_llava_model_args_1b5567/submissions/mathvista_testmini_scores.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/infovqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:175256b5261136789c0148ddd386d63a9897c73496cbf98618d9cf6ddc1fe8d6
3
+ size 576436240
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cde429bd98b1a2e21b2db89ef992e8c3b4197f58afcfe3940b5ac16eb08a0185
3
+ size 14568244
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mme.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec3bf2947f0e3e25966b5f05b47bb6f1e5972429c01bdf34cfead6fa8a827e34
3
+ size 1994104277
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/results.json ADDED
@@ -0,0 +1,683 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6635362694300518,
5
+ "exact_match_stderr,flexible-extract": 0.00850419375917845,
6
+ "alias": "ai2d"
7
+ },
8
+ "infovqa_val": {
9
+ "anls,none": 0.27954302034987505,
10
+ "anls_stderr,none": 0.008481044412679035,
11
+ "alias": "infovqa_val"
12
+ },
13
+ "mmbench_en_dev": {
14
+ "gpt_eval_score,none": 64.51890034364261,
15
+ "gpt_eval_score_stderr,none": "N/A",
16
+ "submission,none": null,
17
+ "submission_stderr,none": "N/A",
18
+ "alias": "mmbench_en_dev"
19
+ },
20
+ "mme": {
21
+ "mme_cognition_score,none": 313.92857142857144,
22
+ "mme_cognition_score_stderr,none": "N/A",
23
+ "mme_percetion_score,none": 1271.4210684273708,
24
+ "mme_percetion_score_stderr,none": "N/A",
25
+ "alias": "mme"
26
+ },
27
+ "mmerealworld_lite": {
28
+ "mme_realworld_score,none": 0.28660760812923397,
29
+ "mme_realworld_score_stderr,none": "N/A",
30
+ "alias": "mmerealworld_lite"
31
+ },
32
+ "mmmu_val": {
33
+ "mmmu_acc,none": 0.40556,
34
+ "mmmu_acc_stderr,none": "N/A",
35
+ "alias": "mmmu_val"
36
+ },
37
+ "mmstar": {
38
+ "coarse perception,none": 0.6550428523661194,
39
+ "coarse perception_stderr,none": "N/A",
40
+ "fine-grained perception,none": 0.267563252272169,
41
+ "fine-grained perception_stderr,none": "N/A",
42
+ "instance reasoning,none": 0.47662379503626334,
43
+ "instance reasoning_stderr,none": "N/A",
44
+ "logical reasoning,none": 0.33408956280243407,
45
+ "logical reasoning_stderr,none": "N/A",
46
+ "math,none": 0.37714403457186135,
47
+ "math_stderr,none": "N/A",
48
+ "science & technology,none": 0.239373463952955,
49
+ "science & technology_stderr,none": "N/A",
50
+ "alias": "mmstar"
51
+ },
52
+ "scienceqa_img": {
53
+ "exact_match,none": 0.7917699553792762,
54
+ "exact_match_stderr,none": 0.009043279237864674,
55
+ "alias": "scienceqa_img"
56
+ },
57
+ "seedbench_2_plus": {
58
+ "seedbench_2_plus_Chart,none": 0.5098765432098765,
59
+ "seedbench_2_plus_Chart_stderr,none": "N/A",
60
+ "seedbench_2_plus_all,none": 0.48924022837066317,
61
+ "seedbench_2_plus_all_stderr,none": "N/A",
62
+ "seedbench_2_plus_Web,none": 0.5106060606060606,
63
+ "seedbench_2_plus_Web_stderr,none": "N/A",
64
+ "seedbench_2_plus_Map,none": 0.4510532837670384,
65
+ "seedbench_2_plus_Map_stderr,none": "N/A",
66
+ "alias": "seedbench_2_plus"
67
+ },
68
+ "textvqa_val": {
69
+ "exact_match,none": 0.38190000027418136,
70
+ "exact_match_stderr,none": 0.00666319424399902,
71
+ "submission,none": null,
72
+ "submission_stderr,none": "N/A",
73
+ "alias": "textvqa_val"
74
+ }
75
+ },
76
+ "configs": {
77
+ "ai2d": {
78
+ "task": "ai2d",
79
+ "dataset_path": "lmms-lab/ai2d",
80
+ "dataset_kwargs": {
81
+ "token": true
82
+ },
83
+ "test_split": "test",
84
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7fb831953d30>",
85
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7fb83195b040>",
86
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7fb83195b5e0>",
87
+ "description": "",
88
+ "target_delimiter": " ",
89
+ "fewshot_delimiter": "\n\n",
90
+ "metric_list": [
91
+ {
92
+ "metric": "exact_match",
93
+ "aggregation": "mean",
94
+ "higher_is_better": true,
95
+ "ignore_case": true,
96
+ "ignore_punctuation": true
97
+ }
98
+ ],
99
+ "output_type": "generate_until",
100
+ "generation_kwargs": {
101
+ "max_new_tokens": 512,
102
+ "temperature": 0.0,
103
+ "do_sample": false,
104
+ "until": [
105
+ "\n\n"
106
+ ]
107
+ },
108
+ "repeats": 1,
109
+ "filter_list": [
110
+ {
111
+ "name": "flexible-extract",
112
+ "filter": [
113
+ {
114
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
115
+ "group_select": 0,
116
+ "ignore_case": true,
117
+ "ignore_punctuation": true,
118
+ "regex_pattern": "([A-Z])\\."
119
+ }
120
+ ]
121
+ }
122
+ ],
123
+ "should_decontaminate": false,
124
+ "metadata": [
125
+ {
126
+ "version": 0.0
127
+ }
128
+ ],
129
+ "model_specific_prompt_kwargs": {
130
+ "default": {
131
+ "prompt_format": "mcq",
132
+ "pre_prompt": "",
133
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
134
+ },
135
+ "gpt4v": {
136
+ "prompt_format": "mcq",
137
+ "pre_prompt": "",
138
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
139
+ },
140
+ "qwen_vl": {
141
+ "prompt_format": "qa",
142
+ "pre_prompt": "",
143
+ "post_prompt": " Answer:"
144
+ },
145
+ "xcomposer2_4khd": {
146
+ "prompt_format": "mcq_xcomposer",
147
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
148
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
149
+ }
150
+ },
151
+ "model_specific_target_kwargs": {
152
+ "default": "mcq",
153
+ "qwen_vl": "qa"
154
+ }
155
+ },
156
+ "infovqa_val": {
157
+ "task": "infovqa_val",
158
+ "dataset_path": "lmms-lab/DocVQA",
159
+ "dataset_name": "InfographicVQA",
160
+ "dataset_kwargs": {
161
+ "token": true
162
+ },
163
+ "test_split": "validation",
164
+ "doc_to_visual": "<function infovqa_doc_to_visual at 0x7fb8658eec10>",
165
+ "doc_to_text": "<function infovqa_doc_to_text at 0x7fb8658eeee0>",
166
+ "doc_to_target": "answers",
167
+ "description": "",
168
+ "target_delimiter": " ",
169
+ "fewshot_delimiter": "\n\n",
170
+ "metric_list": [
171
+ {
172
+ "metric": "anls",
173
+ "aggregation": "mean",
174
+ "higher_is_better": true
175
+ }
176
+ ],
177
+ "output_type": "generate_until",
178
+ "generation_kwargs": {
179
+ "max_new_tokens": 32,
180
+ "temperature": 0.0,
181
+ "do_sample": false,
182
+ "until": [
183
+ "\n\n"
184
+ ]
185
+ },
186
+ "repeats": 1,
187
+ "should_decontaminate": false,
188
+ "model_specific_prompt_kwargs": {
189
+ "default": {
190
+ "pre_prompt": "",
191
+ "post_prompt": "\nAnswer the question using a single word or phrase."
192
+ }
193
+ }
194
+ },
195
+ "mmbench_en_dev": {
196
+ "task": "mmbench_en_dev",
197
+ "dataset_path": "lmms-lab/MMBench",
198
+ "dataset_name": "en",
199
+ "dataset_kwargs": {
200
+ "token": true
201
+ },
202
+ "test_split": "dev",
203
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7fb890b6e5e0>",
204
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7fb890b6eaf0>",
205
+ "doc_to_target": "answer",
206
+ "process_results": "<function mmbench_process_results at 0x7fb890b7a040>",
207
+ "description": "",
208
+ "target_delimiter": " ",
209
+ "fewshot_delimiter": "\n\n",
210
+ "metric_list": [
211
+ {
212
+ "metric": "gpt_eval_score",
213
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7fb890be4e50>",
214
+ "higher_is_better": true
215
+ },
216
+ {
217
+ "metric": "submission",
218
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7fb890b6e3a0>",
219
+ "higher_is_better": true
220
+ }
221
+ ],
222
+ "output_type": "generate_until",
223
+ "generation_kwargs": {
224
+ "until": [
225
+ "ASSISTANT:"
226
+ ],
227
+ "max_new_tokens": 1024,
228
+ "temperature": 0.0,
229
+ "top_p": 1.0,
230
+ "num_beams": 1,
231
+ "do_sample": false,
232
+ "image_aspect_ratio": "original"
233
+ },
234
+ "repeats": 1,
235
+ "should_decontaminate": false,
236
+ "model_specific_prompt_kwargs": {
237
+ "default": {
238
+ "pre_prompt": "",
239
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
240
+ }
241
+ },
242
+ "model_specific_generation_kwargs": {
243
+ "llava": {
244
+ "image_aspect_ratio": "original"
245
+ }
246
+ }
247
+ },
248
+ "mme": {
249
+ "task": "mme",
250
+ "dataset_path": "lmms-lab/MME",
251
+ "dataset_kwargs": {
252
+ "token": false
253
+ },
254
+ "test_split": "test",
255
+ "doc_to_visual": "<function mme_doc_to_visual at 0x7fb89264ca60>",
256
+ "doc_to_text": "<function mme_doc_to_text at 0x7fb890bef310>",
257
+ "doc_to_target": "answer",
258
+ "process_results": "<function mme_process_results at 0x7fb890bef8b0>",
259
+ "description": "",
260
+ "target_delimiter": " ",
261
+ "fewshot_delimiter": "\n\n",
262
+ "metric_list": [
263
+ {
264
+ "metric": "mme_percetion_score",
265
+ "aggregation": "<function mme_aggregate_results at 0x7fb890befdc0>",
266
+ "higher_is_better": true
267
+ },
268
+ {
269
+ "metric": "mme_cognition_score",
270
+ "aggregation": "<function mme_aggregate_results at 0x7fb890bf9280>",
271
+ "higher_is_better": true
272
+ }
273
+ ],
274
+ "output_type": "generate_until",
275
+ "generation_kwargs": {
276
+ "max_new_tokens": 16,
277
+ "temperature": 0.0,
278
+ "top_p": 1.0,
279
+ "num_beams": 1,
280
+ "do_sample": false,
281
+ "until": [
282
+ "\n\n"
283
+ ]
284
+ },
285
+ "repeats": 1,
286
+ "should_decontaminate": false,
287
+ "metadata": [
288
+ {
289
+ "version": 0.0
290
+ }
291
+ ],
292
+ "model_specific_prompt_kwargs": {
293
+ "default": {
294
+ "pre_prompt": "",
295
+ "post_prompt": "\nAnswer the question using a single word or phrase."
296
+ },
297
+ "gpt4v": {
298
+ "pre_prompt": "",
299
+ "post_prompt": "\nAnswer the question with Yes or No."
300
+ },
301
+ "qwen_vl": {
302
+ "pre_prompt": "",
303
+ "post_prompt": " Answer:"
304
+ },
305
+ "otterhd": {
306
+ "pre_prompt": "",
307
+ "post_prompt": " Answer:"
308
+ },
309
+ "xcomposer2_4khd": {
310
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
311
+ "post_prompt": " Answer this question briefly[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
312
+ }
313
+ }
314
+ },
315
+ "mmerealworld_lite": {
316
+ "task": "mmerealworld_lite",
317
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
318
+ "dataset_kwargs": {
319
+ "token": true
320
+ },
321
+ "test_split": "train",
322
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7fb8320fdee0>",
323
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7fb8321009d0>",
324
+ "doc_to_target": "answer",
325
+ "process_results": "<function mme_realworld_process_results at 0x7fb832105670>",
326
+ "description": "",
327
+ "target_delimiter": " ",
328
+ "fewshot_delimiter": "\n\n",
329
+ "metric_list": [
330
+ {
331
+ "metric": "mme_realworld_score",
332
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7fb83210b1f0>",
333
+ "higher_is_better": true
334
+ }
335
+ ],
336
+ "output_type": "generate_until",
337
+ "generation_kwargs": {
338
+ "max_new_tokens": 16,
339
+ "temperature": 0.0,
340
+ "top_p": 1.0,
341
+ "num_beams": 1,
342
+ "do_sample": false,
343
+ "until": [
344
+ "\n\n"
345
+ ]
346
+ },
347
+ "repeats": 1,
348
+ "should_decontaminate": false,
349
+ "metadata": [
350
+ {
351
+ "version": 0.0
352
+ }
353
+ ],
354
+ "model_specific_prompt_kwargs": {
355
+ "default": {
356
+ "pre_prompt": "",
357
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
358
+ },
359
+ "gpt4v": {
360
+ "pre_prompt": "",
361
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
362
+ },
363
+ "xcomposer2_4khd": {
364
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
365
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
366
+ }
367
+ }
368
+ },
369
+ "mmmu_val": {
370
+ "task": "mmmu_val",
371
+ "dataset_path": "lmms-lab/MMMU",
372
+ "test_split": "validation",
373
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7fb843b65940>",
374
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7fb843ae4670>",
375
+ "doc_to_target": "answer",
376
+ "process_results": "<function mmmu_process_results at 0x7fb843ab7550>",
377
+ "description": "",
378
+ "target_delimiter": " ",
379
+ "fewshot_delimiter": "\n\n",
380
+ "metric_list": [
381
+ {
382
+ "metric": "mmmu_acc",
383
+ "aggregation": "<function mmmu_aggregate_results at 0x7fb8438a44c0>",
384
+ "higher_is_better": true
385
+ }
386
+ ],
387
+ "output_type": "generate_until",
388
+ "generation_kwargs": {
389
+ "max_new_tokens": 128,
390
+ "until": [
391
+ "\n\n"
392
+ ],
393
+ "image_aspect_ratio": "original"
394
+ },
395
+ "repeats": 1,
396
+ "should_decontaminate": false,
397
+ "metadata": [
398
+ {
399
+ "version": 0.0
400
+ }
401
+ ],
402
+ "model_specific_generation_kwargs": {
403
+ "llava": {
404
+ "image_aspect_ratio": "original"
405
+ }
406
+ }
407
+ },
408
+ "mmstar": {
409
+ "task": "mmstar",
410
+ "dataset_path": "Lin-Chen/MMStar",
411
+ "dataset_kwargs": {
412
+ "token": true
413
+ },
414
+ "test_split": "val",
415
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7fb8318b33a0>",
416
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7fb8318b3820>",
417
+ "doc_to_target": "answer",
418
+ "process_results": "<function mmstar_process_results at 0x7fb8318b3d30>",
419
+ "description": "",
420
+ "target_delimiter": " ",
421
+ "fewshot_delimiter": "\n\n",
422
+ "metric_list": [
423
+ {
424
+ "metric": "coarse perception",
425
+ "aggregation": "<function mmstar_aggregate_results at 0x7fb8318ba1f0>",
426
+ "higher_is_better": true
427
+ },
428
+ {
429
+ "metric": "fine-grained perception",
430
+ "aggregation": "<function mmstar_aggregate_results at 0x7fb8318ba5e0>",
431
+ "higher_is_better": true
432
+ },
433
+ {
434
+ "metric": "instance reasoning",
435
+ "aggregation": "<function mmstar_aggregate_results at 0x7fb8318ba9d0>",
436
+ "higher_is_better": true
437
+ },
438
+ {
439
+ "metric": "logical reasoning",
440
+ "aggregation": "<function mmstar_aggregate_results at 0x7fb8318badc0>",
441
+ "higher_is_better": true
442
+ },
443
+ {
444
+ "metric": "science & technology",
445
+ "aggregation": "<function mmstar_aggregate_results at 0x7fb8318c21f0>",
446
+ "higher_is_better": true
447
+ },
448
+ {
449
+ "metric": "math",
450
+ "aggregation": "<function mmstar_aggregate_results at 0x7fb8318c25e0>",
451
+ "higher_is_better": true
452
+ }
453
+ ],
454
+ "output_type": "generate_until",
455
+ "generation_kwargs": {
456
+ "until": [
457
+ "\n\n"
458
+ ],
459
+ "do_sample": false
460
+ },
461
+ "repeats": 1,
462
+ "should_decontaminate": false,
463
+ "metadata": [
464
+ {
465
+ "version": 0.0
466
+ }
467
+ ],
468
+ "model_specific_prompt_kwargs": {
469
+ "default": {
470
+ "pre_prompt": "",
471
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
472
+ }
473
+ }
474
+ },
475
+ "scienceqa_img": {
476
+ "task": "scienceqa_img",
477
+ "dataset_path": "lmms-lab/ScienceQA",
478
+ "dataset_name": "ScienceQA-IMG",
479
+ "dataset_kwargs": {
480
+ "token": true
481
+ },
482
+ "test_split": "test",
483
+ "doc_to_visual": "<function sqa_doc_to_visual at 0x7fb85c9cedc0>",
484
+ "doc_to_text": "<function sqa_doc_to_text at 0x7fb85c7db040>",
485
+ "doc_to_target": "<function sqa_doc_to_target at 0x7fb85c7db430>",
486
+ "process_results": "<function sqa_process_results at 0x7fb85c7db790>",
487
+ "description": "",
488
+ "target_delimiter": " ",
489
+ "fewshot_delimiter": "\n\n",
490
+ "metric_list": [
491
+ {
492
+ "metric": "exact_match",
493
+ "aggregation": "mean",
494
+ "higher_is_better": true,
495
+ "ignore_case": true,
496
+ "ignore_punctuation": true
497
+ }
498
+ ],
499
+ "output_type": "generate_until",
500
+ "generation_kwargs": {
501
+ "max_new_tokens": 16,
502
+ "temperature": 0.0,
503
+ "do_sample": false,
504
+ "until": [
505
+ "\n\n"
506
+ ],
507
+ "image_aspect_ratio": "original"
508
+ },
509
+ "repeats": 1,
510
+ "should_decontaminate": false,
511
+ "metadata": [
512
+ {
513
+ "version": 0.0
514
+ }
515
+ ],
516
+ "model_specific_prompt_kwargs": {
517
+ "default": {
518
+ "format": "default",
519
+ "pre_prompt": "",
520
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
521
+ },
522
+ "qwen_vl": {
523
+ "format": "qwen_vl"
524
+ },
525
+ "idefics2": {
526
+ "format": "default",
527
+ "pre_prompt": "",
528
+ "post_prompt": "\nAnswer:"
529
+ }
530
+ },
531
+ "model_specific_generation_kwargs": {
532
+ "llava": {
533
+ "image_aspect_ratio": "original"
534
+ }
535
+ }
536
+ },
537
+ "seedbench_2_plus": {
538
+ "task": "seedbench_2_plus",
539
+ "dataset_path": "doolayer/SEED-Bench-2-Plus",
540
+ "dataset_kwargs": {
541
+ "token": true
542
+ },
543
+ "test_split": "test",
544
+ "doc_to_visual": "<function seed_doc_to_visual at 0x7fb84b511ee0>",
545
+ "doc_to_text": "<function seed_doc_to_text at 0x7fb84b525550>",
546
+ "doc_to_target": "answer",
547
+ "process_results": "<function seed_process_result at 0x7fb84b525a60>",
548
+ "description": "",
549
+ "target_delimiter": " ",
550
+ "fewshot_delimiter": "\n\n",
551
+ "metric_list": [
552
+ {
553
+ "metric": "seedbench_2_plus_Chart",
554
+ "aggregation": "<function seed_aggregation_result at 0x7fb84b525f70>",
555
+ "higher_is_better": true
556
+ },
557
+ {
558
+ "metric": "seedbench_2_plus_Map",
559
+ "aggregation": "<function seed_aggregation_result at 0x7fb84b54c430>",
560
+ "higher_is_better": true
561
+ },
562
+ {
563
+ "metric": "seedbench_2_plus_Web",
564
+ "aggregation": "<function seed_aggregation_result at 0x7fb84b54c8b0>",
565
+ "higher_is_better": true
566
+ },
567
+ {
568
+ "metric": "seedbench_2_plus_all",
569
+ "aggregation": "<function seed_aggregation_result at 0x7fb84b54cd30>",
570
+ "higher_is_better": true
571
+ }
572
+ ],
573
+ "output_type": "generate_until",
574
+ "generation_kwargs": {
575
+ "until": [
576
+ "ASSISTANT:"
577
+ ],
578
+ "max_new_tokens": 16,
579
+ "image_aspect_ratio": "original"
580
+ },
581
+ "repeats": 1,
582
+ "should_decontaminate": false,
583
+ "metadata": [
584
+ {
585
+ "version": 0.0
586
+ }
587
+ ],
588
+ "model_specific_prompt_kwargs": {
589
+ "llava": {
590
+ "img_token": "<image>",
591
+ "post_prompt": "Answer with the option's letter from the given choices directly."
592
+ },
593
+ "gpt4V": {
594
+ "img_token": "<image>",
595
+ "post_prompt": "Answer with the option's letter from the given choices directly."
596
+ },
597
+ "default": {
598
+ "img_token": "<image>",
599
+ "post_prompt": "Answer with the option's letter from the given choices directly."
600
+ }
601
+ }
602
+ },
603
+ "textvqa_val": {
604
+ "task": "textvqa_val",
605
+ "dataset_path": "lmms-lab/textvqa",
606
+ "test_split": "validation",
607
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7fb85dc8a9d0>",
608
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7fb85dc8ae50>",
609
+ "doc_to_target": "answer",
610
+ "process_results": "<function textvqa_process_results at 0x7fb85dc3c160>",
611
+ "description": "",
612
+ "target_delimiter": " ",
613
+ "fewshot_delimiter": "\n\n",
614
+ "metric_list": [
615
+ {
616
+ "metric": "exact_match",
617
+ "aggregation": "mean",
618
+ "higher_is_better": true,
619
+ "ignore_case": true,
620
+ "ignore_punctuation": true
621
+ },
622
+ {
623
+ "metric": "submission",
624
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7fb85dc8a790>",
625
+ "higher_is_better": true
626
+ }
627
+ ],
628
+ "output_type": "generate_until",
629
+ "generation_kwargs": {
630
+ "until": [
631
+ "ASSISTANT:"
632
+ ]
633
+ },
634
+ "repeats": 1,
635
+ "should_decontaminate": false,
636
+ "model_specific_prompt_kwargs": {
637
+ "default": {
638
+ "pre_prompt": "",
639
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
640
+ "ocr": false
641
+ },
642
+ "qwen_vl": {
643
+ "pre_prompt": "",
644
+ "post_prompt": " Answer:"
645
+ }
646
+ }
647
+ }
648
+ },
649
+ "versions": {
650
+ "ai2d": "Yaml",
651
+ "infovqa_val": "Yaml",
652
+ "mmbench_en_dev": "Yaml",
653
+ "mme": "Yaml",
654
+ "mmerealworld_lite": "Yaml",
655
+ "mmmu_val": "Yaml",
656
+ "mmstar": "Yaml",
657
+ "scienceqa_img": "Yaml",
658
+ "seedbench_2_plus": "Yaml",
659
+ "textvqa_val": "Yaml"
660
+ },
661
+ "n-shot": {
662
+ "ai2d": 0,
663
+ "infovqa_val": 0,
664
+ "mmbench_en_dev": 0,
665
+ "mme": 0,
666
+ "mmerealworld_lite": 0,
667
+ "mmmu_val": 0,
668
+ "mmstar": 0,
669
+ "scienceqa_img": 0,
670
+ "seedbench_2_plus": 0,
671
+ "textvqa_val": 0
672
+ },
673
+ "model_configs": {
674
+ "model": "llava",
675
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_xmoe/checkpoint-6893,conv_template=phi35",
676
+ "batch_size": "1",
677
+ "device": null,
678
+ "limit": null,
679
+ "bootstrap_iters": 100000,
680
+ "gen_kwargs": ""
681
+ },
682
+ "git_hash": "289c7fe5"
683
+ }
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/scienceqa_img.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/seedbench_2_plus.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.6451890034364262, "category_acc": {"action_recognition": 0.8518518518518519, "attribute_comparison": 0.45454545454545453, "attribute_recognition": 0.7837837837837838, "celebrity_recognition": 0.7575757575757576, "function_reasoning": 0.8354430379746836, "future_prediction": 0.475, "identity_reasoning": 0.9555555555555556, "image_emotion": 0.78, "image_quality": 0.24528301886792453, "image_scene": 0.9519230769230769, "image_style": 0.8490566037735849, "image_topic": 0.8333333333333334, "nature_relation": 0.4791666666666667, "object_localization": 0.2962962962962963, "ocr": 0.6666666666666666, "physical_property_reasoning": 0.4266666666666667, "physical_relation": 0.5833333333333334, "social_relation": 0.7674418604651163, "spatial_relationship": 0.2, "structuralized_imagetext_understanding": 0.47435897435897434}, "l2_category_acc": {"attribute_reasoning": 0.7085427135678392, "coarse_perception": 0.7635135135135135, "finegrained_perception (cross-instance)": 0.5244755244755245, "finegrained_perception (instance-level)": 0.6245733788395904, "logic_reasoning": 0.4745762711864407, "relation_reasoning": 0.6086956521739131}}
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f21fe862d3c2785a4f0650be825aa355a9424698fe5c576440033616c1bc43ca
3
+ size 843215
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/submissions/textvqa_submission_2025-06-28-17-54-58.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1825_llava...l_mme_llava_model_args_1b5567/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ba80382b4fa3937a6856268042977460ca353de906f330a72a3b5e56d230ebd
3
+ size 13143093
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1855_llava...bench_llava_model_args_1b5567/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1855_llava...bench_llava_model_args_1b5567/results/ocrbench_results.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ######################### OCRBench #############################
2
+ Text Recognition(Total 300): 191
3
+ ---------------- Details of Recognition Score ------------------
4
+ Regular Text Recognition(Total 50): 42
5
+ Irregular Text Recognition(Total 50): 44
6
+ Artistic Text Recognition(Total 50): 45
7
+ Handwriting Recognition(Total 50): 16
8
+ Digit String Recognition(Total 50): 24
9
+ Non-Semantic Text Recognition(Total 50): 20
10
+ ----------------------------------------------------------------
11
+ Scene Text-centric VQA(Total 200): 102
12
+ ----------------------------------------------------------------
13
+ Doc-oriented VQA(Total 200): 23
14
+ ----------------------------------------------------------------
15
+ Key Information Extraction(Total 200): 14
16
+ Handwritten Mathematical Expression Recognition(Total 100): 19
17
+ --------------------- Final Score ------------------------------
18
+ Final Score(Total 1000): 349
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1856_llava...bench_llava_model_args_1b5567/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_1856_llava...bench_llava_model_args_1b5567/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/infovqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c393583142b44257982c1359b489a662f7865152cc140e4390427c9aab1ac029
3
+ size 576436240
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ac93c7475024e536312fcd62e05997ad747885abb366b8f03e76d18d14aea80
3
+ size 14568244
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mme.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98e6970a72ea302fb6694ed35dfdda24bc87855289c5d3f2cb27e29773b539ab
3
+ size 1994104277
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/results.json ADDED
@@ -0,0 +1,683 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6596502590673575,
5
+ "exact_match_stderr,flexible-extract": 0.008528080007639032,
6
+ "alias": "ai2d"
7
+ },
8
+ "infovqa_val": {
9
+ "anls,none": 0.27954302034987505,
10
+ "anls_stderr,none": 0.008481044412679035,
11
+ "alias": "infovqa_val"
12
+ },
13
+ "mmbench_en_dev": {
14
+ "gpt_eval_score,none": 64.51890034364261,
15
+ "gpt_eval_score_stderr,none": "N/A",
16
+ "submission,none": null,
17
+ "submission_stderr,none": "N/A",
18
+ "alias": "mmbench_en_dev"
19
+ },
20
+ "mme": {
21
+ "mme_cognition_score,none": 313.92857142857144,
22
+ "mme_cognition_score_stderr,none": "N/A",
23
+ "mme_percetion_score,none": 1271.4210684273708,
24
+ "mme_percetion_score_stderr,none": "N/A",
25
+ "alias": "mme"
26
+ },
27
+ "mmerealworld_lite": {
28
+ "mme_realworld_score,none": 0.28660760812923397,
29
+ "mme_realworld_score_stderr,none": "N/A",
30
+ "alias": "mmerealworld_lite"
31
+ },
32
+ "mmmu_val": {
33
+ "mmmu_acc,none": 0.40556,
34
+ "mmmu_acc_stderr,none": "N/A",
35
+ "alias": "mmmu_val"
36
+ },
37
+ "mmstar": {
38
+ "coarse perception,none": 0.6550428523661194,
39
+ "coarse perception_stderr,none": "N/A",
40
+ "fine-grained perception,none": 0.267563252272169,
41
+ "fine-grained perception_stderr,none": "N/A",
42
+ "instance reasoning,none": 0.47662379503626334,
43
+ "instance reasoning_stderr,none": "N/A",
44
+ "logical reasoning,none": 0.33408956280243407,
45
+ "logical reasoning_stderr,none": "N/A",
46
+ "math,none": 0.37714403457186135,
47
+ "math_stderr,none": "N/A",
48
+ "science & technology,none": 0.239373463952955,
49
+ "science & technology_stderr,none": "N/A",
50
+ "alias": "mmstar"
51
+ },
52
+ "scienceqa_img": {
53
+ "exact_match,none": 0.7917699553792762,
54
+ "exact_match_stderr,none": 0.009043279237864684,
55
+ "alias": "scienceqa_img"
56
+ },
57
+ "seedbench_2_plus": {
58
+ "seedbench_2_plus_Chart,none": 0.5098765432098765,
59
+ "seedbench_2_plus_Chart_stderr,none": "N/A",
60
+ "seedbench_2_plus_all,none": 0.48924022837066317,
61
+ "seedbench_2_plus_all_stderr,none": "N/A",
62
+ "seedbench_2_plus_Web,none": 0.5106060606060606,
63
+ "seedbench_2_plus_Web_stderr,none": "N/A",
64
+ "seedbench_2_plus_Map,none": 0.4510532837670384,
65
+ "seedbench_2_plus_Map_stderr,none": "N/A",
66
+ "alias": "seedbench_2_plus"
67
+ },
68
+ "textvqa_val": {
69
+ "exact_match,none": 0.38190000027418136,
70
+ "exact_match_stderr,none": 0.006663194243999023,
71
+ "submission,none": null,
72
+ "submission_stderr,none": "N/A",
73
+ "alias": "textvqa_val"
74
+ }
75
+ },
76
+ "configs": {
77
+ "ai2d": {
78
+ "task": "ai2d",
79
+ "dataset_path": "lmms-lab/ai2d",
80
+ "dataset_kwargs": {
81
+ "token": true
82
+ },
83
+ "test_split": "test",
84
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7fa9f0e6cca0>",
85
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7fa9f0e6cf70>",
86
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7fa9f0e72550>",
87
+ "description": "",
88
+ "target_delimiter": " ",
89
+ "fewshot_delimiter": "\n\n",
90
+ "metric_list": [
91
+ {
92
+ "metric": "exact_match",
93
+ "aggregation": "mean",
94
+ "higher_is_better": true,
95
+ "ignore_case": true,
96
+ "ignore_punctuation": true
97
+ }
98
+ ],
99
+ "output_type": "generate_until",
100
+ "generation_kwargs": {
101
+ "max_new_tokens": 512,
102
+ "temperature": 0.0,
103
+ "do_sample": false,
104
+ "until": [
105
+ "\n\n"
106
+ ]
107
+ },
108
+ "repeats": 1,
109
+ "filter_list": [
110
+ {
111
+ "name": "flexible-extract",
112
+ "filter": [
113
+ {
114
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
115
+ "group_select": 0,
116
+ "ignore_case": true,
117
+ "ignore_punctuation": true,
118
+ "regex_pattern": "([A-Z])\\."
119
+ }
120
+ ]
121
+ }
122
+ ],
123
+ "should_decontaminate": false,
124
+ "metadata": [
125
+ {
126
+ "version": 0.0
127
+ }
128
+ ],
129
+ "model_specific_prompt_kwargs": {
130
+ "default": {
131
+ "prompt_format": "mcq",
132
+ "pre_prompt": "",
133
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
134
+ },
135
+ "gpt4v": {
136
+ "prompt_format": "mcq",
137
+ "pre_prompt": "",
138
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
139
+ },
140
+ "qwen_vl": {
141
+ "prompt_format": "qa",
142
+ "pre_prompt": "",
143
+ "post_prompt": " Answer:"
144
+ },
145
+ "xcomposer2_4khd": {
146
+ "prompt_format": "mcq_xcomposer",
147
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
148
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
149
+ }
150
+ },
151
+ "model_specific_target_kwargs": {
152
+ "default": "mcq",
153
+ "qwen_vl": "qa"
154
+ }
155
+ },
156
+ "infovqa_val": {
157
+ "task": "infovqa_val",
158
+ "dataset_path": "lmms-lab/DocVQA",
159
+ "dataset_name": "InfographicVQA",
160
+ "dataset_kwargs": {
161
+ "token": true
162
+ },
163
+ "test_split": "validation",
164
+ "doc_to_visual": "<function infovqa_doc_to_visual at 0x7faa233e3b80>",
165
+ "doc_to_text": "<function infovqa_doc_to_text at 0x7faa233e3e50>",
166
+ "doc_to_target": "answers",
167
+ "description": "",
168
+ "target_delimiter": " ",
169
+ "fewshot_delimiter": "\n\n",
170
+ "metric_list": [
171
+ {
172
+ "metric": "anls",
173
+ "aggregation": "mean",
174
+ "higher_is_better": true
175
+ }
176
+ ],
177
+ "output_type": "generate_until",
178
+ "generation_kwargs": {
179
+ "max_new_tokens": 32,
180
+ "temperature": 0.0,
181
+ "do_sample": false,
182
+ "until": [
183
+ "\n\n"
184
+ ]
185
+ },
186
+ "repeats": 1,
187
+ "should_decontaminate": false,
188
+ "model_specific_prompt_kwargs": {
189
+ "default": {
190
+ "pre_prompt": "",
191
+ "post_prompt": "\nAnswer the question using a single word or phrase."
192
+ }
193
+ }
194
+ },
195
+ "mmbench_en_dev": {
196
+ "task": "mmbench_en_dev",
197
+ "dataset_path": "lmms-lab/MMBench",
198
+ "dataset_name": "en",
199
+ "dataset_kwargs": {
200
+ "token": true
201
+ },
202
+ "test_split": "dev",
203
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7faa500aa5e0>",
204
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7faa500aaaf0>",
205
+ "doc_to_target": "answer",
206
+ "process_results": "<function mmbench_process_results at 0x7faa500b5040>",
207
+ "description": "",
208
+ "target_delimiter": " ",
209
+ "fewshot_delimiter": "\n\n",
210
+ "metric_list": [
211
+ {
212
+ "metric": "gpt_eval_score",
213
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7faa50120e50>",
214
+ "higher_is_better": true
215
+ },
216
+ {
217
+ "metric": "submission",
218
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7faa500aa3a0>",
219
+ "higher_is_better": true
220
+ }
221
+ ],
222
+ "output_type": "generate_until",
223
+ "generation_kwargs": {
224
+ "until": [
225
+ "ASSISTANT:"
226
+ ],
227
+ "max_new_tokens": 1024,
228
+ "temperature": 0.0,
229
+ "top_p": 1.0,
230
+ "num_beams": 1,
231
+ "do_sample": false,
232
+ "image_aspect_ratio": "original"
233
+ },
234
+ "repeats": 1,
235
+ "should_decontaminate": false,
236
+ "model_specific_prompt_kwargs": {
237
+ "default": {
238
+ "pre_prompt": "",
239
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
240
+ }
241
+ },
242
+ "model_specific_generation_kwargs": {
243
+ "llava": {
244
+ "image_aspect_ratio": "original"
245
+ }
246
+ }
247
+ },
248
+ "mme": {
249
+ "task": "mme",
250
+ "dataset_path": "lmms-lab/MME",
251
+ "dataset_kwargs": {
252
+ "token": false
253
+ },
254
+ "test_split": "test",
255
+ "doc_to_visual": "<function mme_doc_to_visual at 0x7faa51b87a60>",
256
+ "doc_to_text": "<function mme_doc_to_text at 0x7faa5012b310>",
257
+ "doc_to_target": "answer",
258
+ "process_results": "<function mme_process_results at 0x7faa5012b8b0>",
259
+ "description": "",
260
+ "target_delimiter": " ",
261
+ "fewshot_delimiter": "\n\n",
262
+ "metric_list": [
263
+ {
264
+ "metric": "mme_percetion_score",
265
+ "aggregation": "<function mme_aggregate_results at 0x7faa5012bdc0>",
266
+ "higher_is_better": true
267
+ },
268
+ {
269
+ "metric": "mme_cognition_score",
270
+ "aggregation": "<function mme_aggregate_results at 0x7faa50134280>",
271
+ "higher_is_better": true
272
+ }
273
+ ],
274
+ "output_type": "generate_until",
275
+ "generation_kwargs": {
276
+ "max_new_tokens": 16,
277
+ "temperature": 0.0,
278
+ "top_p": 1.0,
279
+ "num_beams": 1,
280
+ "do_sample": false,
281
+ "until": [
282
+ "\n\n"
283
+ ]
284
+ },
285
+ "repeats": 1,
286
+ "should_decontaminate": false,
287
+ "metadata": [
288
+ {
289
+ "version": 0.0
290
+ }
291
+ ],
292
+ "model_specific_prompt_kwargs": {
293
+ "default": {
294
+ "pre_prompt": "",
295
+ "post_prompt": "\nAnswer the question using a single word or phrase."
296
+ },
297
+ "gpt4v": {
298
+ "pre_prompt": "",
299
+ "post_prompt": "\nAnswer the question with Yes or No."
300
+ },
301
+ "qwen_vl": {
302
+ "pre_prompt": "",
303
+ "post_prompt": " Answer:"
304
+ },
305
+ "otterhd": {
306
+ "pre_prompt": "",
307
+ "post_prompt": " Answer:"
308
+ },
309
+ "xcomposer2_4khd": {
310
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
311
+ "post_prompt": " Answer this question briefly[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
312
+ }
313
+ }
314
+ },
315
+ "mmerealworld_lite": {
316
+ "task": "mmerealworld_lite",
317
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
318
+ "dataset_kwargs": {
319
+ "token": true
320
+ },
321
+ "test_split": "train",
322
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7fa9f1614e50>",
323
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7fa9f1619940>",
324
+ "doc_to_target": "answer",
325
+ "process_results": "<function mme_realworld_process_results at 0x7fa9f161c5e0>",
326
+ "description": "",
327
+ "target_delimiter": " ",
328
+ "fewshot_delimiter": "\n\n",
329
+ "metric_list": [
330
+ {
331
+ "metric": "mme_realworld_score",
332
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7fa9f1623160>",
333
+ "higher_is_better": true
334
+ }
335
+ ],
336
+ "output_type": "generate_until",
337
+ "generation_kwargs": {
338
+ "max_new_tokens": 16,
339
+ "temperature": 0.0,
340
+ "top_p": 1.0,
341
+ "num_beams": 1,
342
+ "do_sample": false,
343
+ "until": [
344
+ "\n\n"
345
+ ]
346
+ },
347
+ "repeats": 1,
348
+ "should_decontaminate": false,
349
+ "metadata": [
350
+ {
351
+ "version": 0.0
352
+ }
353
+ ],
354
+ "model_specific_prompt_kwargs": {
355
+ "default": {
356
+ "pre_prompt": "",
357
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
358
+ },
359
+ "gpt4v": {
360
+ "pre_prompt": "",
361
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
362
+ },
363
+ "xcomposer2_4khd": {
364
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
365
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
366
+ }
367
+ }
368
+ },
369
+ "mmmu_val": {
370
+ "task": "mmmu_val",
371
+ "dataset_path": "lmms-lab/MMMU",
372
+ "test_split": "validation",
373
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7faa03fb38b0>",
374
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7faa03eb45e0>",
375
+ "doc_to_target": "answer",
376
+ "process_results": "<function mmmu_process_results at 0x7faa03d544c0>",
377
+ "description": "",
378
+ "target_delimiter": " ",
379
+ "fewshot_delimiter": "\n\n",
380
+ "metric_list": [
381
+ {
382
+ "metric": "mmmu_acc",
383
+ "aggregation": "<function mmmu_aggregate_results at 0x7faa03cec430>",
384
+ "higher_is_better": true
385
+ }
386
+ ],
387
+ "output_type": "generate_until",
388
+ "generation_kwargs": {
389
+ "max_new_tokens": 128,
390
+ "until": [
391
+ "\n\n"
392
+ ],
393
+ "image_aspect_ratio": "original"
394
+ },
395
+ "repeats": 1,
396
+ "should_decontaminate": false,
397
+ "metadata": [
398
+ {
399
+ "version": 0.0
400
+ }
401
+ ],
402
+ "model_specific_generation_kwargs": {
403
+ "llava": {
404
+ "image_aspect_ratio": "original"
405
+ }
406
+ }
407
+ },
408
+ "mmstar": {
409
+ "task": "mmstar",
410
+ "dataset_path": "Lin-Chen/MMStar",
411
+ "dataset_kwargs": {
412
+ "token": true
413
+ },
414
+ "test_split": "val",
415
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7fa9f0dc8310>",
416
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7fa9f0dc8790>",
417
+ "doc_to_target": "answer",
418
+ "process_results": "<function mmstar_process_results at 0x7fa9f0dc8ca0>",
419
+ "description": "",
420
+ "target_delimiter": " ",
421
+ "fewshot_delimiter": "\n\n",
422
+ "metric_list": [
423
+ {
424
+ "metric": "coarse perception",
425
+ "aggregation": "<function mmstar_aggregate_results at 0x7fa9f0dd2160>",
426
+ "higher_is_better": true
427
+ },
428
+ {
429
+ "metric": "fine-grained perception",
430
+ "aggregation": "<function mmstar_aggregate_results at 0x7fa9f0dd2550>",
431
+ "higher_is_better": true
432
+ },
433
+ {
434
+ "metric": "instance reasoning",
435
+ "aggregation": "<function mmstar_aggregate_results at 0x7fa9f0dd2940>",
436
+ "higher_is_better": true
437
+ },
438
+ {
439
+ "metric": "logical reasoning",
440
+ "aggregation": "<function mmstar_aggregate_results at 0x7fa9f0dd2d30>",
441
+ "higher_is_better": true
442
+ },
443
+ {
444
+ "metric": "science & technology",
445
+ "aggregation": "<function mmstar_aggregate_results at 0x7fa9f0ddb160>",
446
+ "higher_is_better": true
447
+ },
448
+ {
449
+ "metric": "math",
450
+ "aggregation": "<function mmstar_aggregate_results at 0x7fa9f0ddb550>",
451
+ "higher_is_better": true
452
+ }
453
+ ],
454
+ "output_type": "generate_until",
455
+ "generation_kwargs": {
456
+ "until": [
457
+ "\n\n"
458
+ ],
459
+ "do_sample": false
460
+ },
461
+ "repeats": 1,
462
+ "should_decontaminate": false,
463
+ "metadata": [
464
+ {
465
+ "version": 0.0
466
+ }
467
+ ],
468
+ "model_specific_prompt_kwargs": {
469
+ "default": {
470
+ "pre_prompt": "",
471
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
472
+ }
473
+ }
474
+ },
475
+ "scienceqa_img": {
476
+ "task": "scienceqa_img",
477
+ "dataset_path": "lmms-lab/ScienceQA",
478
+ "dataset_name": "ScienceQA-IMG",
479
+ "dataset_kwargs": {
480
+ "token": true
481
+ },
482
+ "test_split": "test",
483
+ "doc_to_visual": "<function sqa_doc_to_visual at 0x7faa1bd06d30>",
484
+ "doc_to_text": "<function sqa_doc_to_text at 0x7faa1bd06f70>",
485
+ "doc_to_target": "<function sqa_doc_to_target at 0x7faa1bc5d3a0>",
486
+ "process_results": "<function sqa_process_results at 0x7faa1bc5d700>",
487
+ "description": "",
488
+ "target_delimiter": " ",
489
+ "fewshot_delimiter": "\n\n",
490
+ "metric_list": [
491
+ {
492
+ "metric": "exact_match",
493
+ "aggregation": "mean",
494
+ "higher_is_better": true,
495
+ "ignore_case": true,
496
+ "ignore_punctuation": true
497
+ }
498
+ ],
499
+ "output_type": "generate_until",
500
+ "generation_kwargs": {
501
+ "max_new_tokens": 16,
502
+ "temperature": 0.0,
503
+ "do_sample": false,
504
+ "until": [
505
+ "\n\n"
506
+ ],
507
+ "image_aspect_ratio": "original"
508
+ },
509
+ "repeats": 1,
510
+ "should_decontaminate": false,
511
+ "metadata": [
512
+ {
513
+ "version": 0.0
514
+ }
515
+ ],
516
+ "model_specific_prompt_kwargs": {
517
+ "default": {
518
+ "format": "default",
519
+ "pre_prompt": "",
520
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
521
+ },
522
+ "qwen_vl": {
523
+ "format": "qwen_vl"
524
+ },
525
+ "idefics2": {
526
+ "format": "default",
527
+ "pre_prompt": "",
528
+ "post_prompt": "\nAnswer:"
529
+ }
530
+ },
531
+ "model_specific_generation_kwargs": {
532
+ "llava": {
533
+ "image_aspect_ratio": "original"
534
+ }
535
+ }
536
+ },
537
+ "seedbench_2_plus": {
538
+ "task": "seedbench_2_plus",
539
+ "dataset_path": "doolayer/SEED-Bench-2-Plus",
540
+ "dataset_kwargs": {
541
+ "token": true
542
+ },
543
+ "test_split": "test",
544
+ "doc_to_visual": "<function seed_doc_to_visual at 0x7faa19e13e50>",
545
+ "doc_to_text": "<function seed_doc_to_text at 0x7faa19e3a4c0>",
546
+ "doc_to_target": "answer",
547
+ "process_results": "<function seed_process_result at 0x7faa19e3a9d0>",
548
+ "description": "",
549
+ "target_delimiter": " ",
550
+ "fewshot_delimiter": "\n\n",
551
+ "metric_list": [
552
+ {
553
+ "metric": "seedbench_2_plus_Chart",
554
+ "aggregation": "<function seed_aggregation_result at 0x7faa19e3aee0>",
555
+ "higher_is_better": true
556
+ },
557
+ {
558
+ "metric": "seedbench_2_plus_Map",
559
+ "aggregation": "<function seed_aggregation_result at 0x7faa19c143a0>",
560
+ "higher_is_better": true
561
+ },
562
+ {
563
+ "metric": "seedbench_2_plus_Web",
564
+ "aggregation": "<function seed_aggregation_result at 0x7faa19c14820>",
565
+ "higher_is_better": true
566
+ },
567
+ {
568
+ "metric": "seedbench_2_plus_all",
569
+ "aggregation": "<function seed_aggregation_result at 0x7faa19c14ca0>",
570
+ "higher_is_better": true
571
+ }
572
+ ],
573
+ "output_type": "generate_until",
574
+ "generation_kwargs": {
575
+ "until": [
576
+ "ASSISTANT:"
577
+ ],
578
+ "max_new_tokens": 16,
579
+ "image_aspect_ratio": "original"
580
+ },
581
+ "repeats": 1,
582
+ "should_decontaminate": false,
583
+ "metadata": [
584
+ {
585
+ "version": 0.0
586
+ }
587
+ ],
588
+ "model_specific_prompt_kwargs": {
589
+ "llava": {
590
+ "img_token": "<image>",
591
+ "post_prompt": "Answer with the option's letter from the given choices directly."
592
+ },
593
+ "gpt4V": {
594
+ "img_token": "<image>",
595
+ "post_prompt": "Answer with the option's letter from the given choices directly."
596
+ },
597
+ "default": {
598
+ "img_token": "<image>",
599
+ "post_prompt": "Answer with the option's letter from the given choices directly."
600
+ }
601
+ }
602
+ },
603
+ "textvqa_val": {
604
+ "task": "textvqa_val",
605
+ "dataset_path": "lmms-lab/textvqa",
606
+ "test_split": "validation",
607
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7faa1d1ca940>",
608
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7faa1d1cadc0>",
609
+ "doc_to_target": "answer",
610
+ "process_results": "<function textvqa_process_results at 0x7faa1d17f0d0>",
611
+ "description": "",
612
+ "target_delimiter": " ",
613
+ "fewshot_delimiter": "\n\n",
614
+ "metric_list": [
615
+ {
616
+ "metric": "exact_match",
617
+ "aggregation": "mean",
618
+ "higher_is_better": true,
619
+ "ignore_case": true,
620
+ "ignore_punctuation": true
621
+ },
622
+ {
623
+ "metric": "submission",
624
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7faa1d1ca700>",
625
+ "higher_is_better": true
626
+ }
627
+ ],
628
+ "output_type": "generate_until",
629
+ "generation_kwargs": {
630
+ "until": [
631
+ "ASSISTANT:"
632
+ ]
633
+ },
634
+ "repeats": 1,
635
+ "should_decontaminate": false,
636
+ "model_specific_prompt_kwargs": {
637
+ "default": {
638
+ "pre_prompt": "",
639
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
640
+ "ocr": false
641
+ },
642
+ "qwen_vl": {
643
+ "pre_prompt": "",
644
+ "post_prompt": " Answer:"
645
+ }
646
+ }
647
+ }
648
+ },
649
+ "versions": {
650
+ "ai2d": "Yaml",
651
+ "infovqa_val": "Yaml",
652
+ "mmbench_en_dev": "Yaml",
653
+ "mme": "Yaml",
654
+ "mmerealworld_lite": "Yaml",
655
+ "mmmu_val": "Yaml",
656
+ "mmstar": "Yaml",
657
+ "scienceqa_img": "Yaml",
658
+ "seedbench_2_plus": "Yaml",
659
+ "textvqa_val": "Yaml"
660
+ },
661
+ "n-shot": {
662
+ "ai2d": 0,
663
+ "infovqa_val": 0,
664
+ "mmbench_en_dev": 0,
665
+ "mme": 0,
666
+ "mmerealworld_lite": 0,
667
+ "mmmu_val": 0,
668
+ "mmstar": 0,
669
+ "scienceqa_img": 0,
670
+ "seedbench_2_plus": 0,
671
+ "textvqa_val": 0
672
+ },
673
+ "model_configs": {
674
+ "model": "llava",
675
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/Full_xmoe/checkpoint-6893,conv_template=phi35",
676
+ "batch_size": "1",
677
+ "device": null,
678
+ "limit": null,
679
+ "bootstrap_iters": 100000,
680
+ "gen_kwargs": ""
681
+ },
682
+ "git_hash": "289c7fe5"
683
+ }
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/scienceqa_img.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/seedbench_2_plus.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.6451890034364262, "category_acc": {"action_recognition": 0.8518518518518519, "attribute_comparison": 0.45454545454545453, "attribute_recognition": 0.7837837837837838, "celebrity_recognition": 0.7575757575757576, "function_reasoning": 0.8354430379746836, "future_prediction": 0.475, "identity_reasoning": 0.9555555555555556, "image_emotion": 0.78, "image_quality": 0.24528301886792453, "image_scene": 0.9519230769230769, "image_style": 0.8490566037735849, "image_topic": 0.8333333333333334, "nature_relation": 0.4791666666666667, "object_localization": 0.2962962962962963, "ocr": 0.6666666666666666, "physical_property_reasoning": 0.4266666666666667, "physical_relation": 0.5833333333333334, "social_relation": 0.7674418604651163, "spatial_relationship": 0.2, "structuralized_imagetext_understanding": 0.47435897435897434}, "l2_category_acc": {"attribute_reasoning": 0.7085427135678392, "coarse_perception": 0.7635135135135135, "finegrained_perception (cross-instance)": 0.5244755244755245, "finegrained_perception (instance-level)": 0.6245733788395904, "logic_reasoning": 0.4745762711864407, "relation_reasoning": 0.6086956521739131}}
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c07e7ebd9c61863a3910baa85a66953103aa61f6b3134d31fe4a77e861c6103c
3
+ size 865074
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/submissions/textvqa_submission_2025-06-28-20-16-15.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2054_llava...l_mme_llava_model_args_1b5567/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d9b3b448c2e6b2346c75ae18ab583f8e1bde17a03a337899f59e26afa96f342
3
+ size 13143266
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2117_llava...bench_llava_model_args_1b5567/ocrbench.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2117_llava...bench_llava_model_args_1b5567/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/Full_xmoe/checkpoint-6893/logs/0628_2117_llava...bench_llava_model_args_1b5567/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done