Upload folder using huggingface_hub

#823
This view is limited to 50 files because it contains too many changes.  See the raw diff here.
Files changed (50) hide show
  1. .gitattributes +5 -0
  2. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/ai2d.json +0 -0
  3. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/infovqa_val.json +3 -0
  4. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmbench_en_dev.json +3 -0
  5. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mme.json +0 -0
  6. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmerealworld_lite.json +3 -0
  7. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmmu_pro_standard.json +0 -0
  8. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmmu_val.json +0 -0
  9. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmstar.json +0 -0
  10. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank0_metric_eval_done.txt +1 -0
  11. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank1_metric_eval_done.txt +1 -0
  12. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank2_metric_eval_done.txt +1 -0
  13. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank3_metric_eval_done.txt +1 -0
  14. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/results.json +729 -0
  15. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/scienceqa_img.json +0 -0
  16. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/seedbench_2_plus.json +0 -0
  17. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/mmbench_en_dev_results.json +1 -0
  18. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/mmbench_en_dev_results.xlsx +3 -0
  19. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/textvqa_submission_2025-07-17-11-00-08.json +0 -0
  20. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/textvqa_val.json +3 -0
  21. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/gpt_response/hallusion_output_vd_model.json +0 -0
  22. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/gpt_response/hallusion_output_vs_model.json +0 -0
  23. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/hallusion_bench_image.json +0 -0
  24. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/mathvista_testmini.json +0 -0
  25. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank0_metric_eval_done.txt +1 -0
  26. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank1_metric_eval_done.txt +1 -0
  27. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank2_metric_eval_done.txt +1 -0
  28. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank3_metric_eval_done.txt +1 -0
  29. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/results.json +146 -0
  30. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/submissions/mathvista_testmini_scores.json +0 -0
  31. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/ocrbench.json +0 -0
  32. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank0_metric_eval_done.txt +1 -0
  33. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank1_metric_eval_done.txt +1 -0
  34. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank2_metric_eval_done.txt +1 -0
  35. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank3_metric_eval_done.txt +1 -0
  36. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/results.json +67 -0
  37. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/results/ocrbench_results.txt +18 -0
  38. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json +0 -0
  39. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank0_metric_eval_done.txt +1 -0
  40. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank1_metric_eval_done.txt +1 -0
  41. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank2_metric_eval_done.txt +1 -0
  42. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank3_metric_eval_done.txt +1 -0
  43. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/results.json +106 -0
  44. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json +0 -0
  45. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank0_metric_eval_done.txt +1 -0
  46. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank1_metric_eval_done.txt +1 -0
  47. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank2_metric_eval_done.txt +1 -0
  48. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank3_metric_eval_done.txt +1 -0
  49. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/results.json +106 -0
  50. sft/665K36/revise_Full_smoe_sharev3/logs/0717_1747_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json +0 -0
.gitattributes CHANGED
@@ -380,3 +380,8 @@ sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model
380
  sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model_args_fe4e53/mme.json filter=lfs diff=lfs merge=lfs -text
381
  sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model_args_fe4e53/mmmu_val.json filter=lfs diff=lfs merge=lfs -text
382
  sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model_args_fe4e53/mmstar.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
380
  sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model_args_fe4e53/mme.json filter=lfs diff=lfs merge=lfs -text
381
  sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model_args_fe4e53/mmmu_val.json filter=lfs diff=lfs merge=lfs -text
382
  sft/665K36/revise_Full_smoe_sharev3/analysts/0717_2301_llava...mstar_llava_model_args_fe4e53/mmstar.json filter=lfs diff=lfs merge=lfs -text
383
+ sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
384
+ sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmbench_en_dev.json filter=lfs diff=lfs merge=lfs -text
385
+ sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmerealworld_lite.json filter=lfs diff=lfs merge=lfs -text
386
+ sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
387
+ sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/infovqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29a7d0a6b377e2db41a41138f9c0165be53761898d5871f29d3ba09e749d3082
3
+ size 576426884
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmbench_en_dev.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1160459a8adad31add5915f0c989f40dd9d2a73dedeb89365eed11ddac2b0af
3
+ size 14568288
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mme.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmerealworld_lite.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d071cdf852b01c5d7548b8130aada4f8e15fda5c595a5a840ddb48d3b6bd24ae
3
+ size 1994104321
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmmu_pro_standard.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmmu_val.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/mmstar.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/results.json ADDED
@@ -0,0 +1,729 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.645401554404145,
5
+ "exact_match_stderr,flexible-extract": 0.008610236936498226,
6
+ "alias": "ai2d"
7
+ },
8
+ "infovqa_val": {
9
+ "anls,none": 0.26811852909675116,
10
+ "anls_stderr,none": 0.008371528778539056,
11
+ "alias": "infovqa_val"
12
+ },
13
+ "mmbench_en_dev": {
14
+ "gpt_eval_score,none": 72.42268041237114,
15
+ "gpt_eval_score_stderr,none": "N/A",
16
+ "submission,none": null,
17
+ "submission_stderr,none": "N/A",
18
+ "alias": "mmbench_en_dev"
19
+ },
20
+ "mme": {
21
+ "mme_cognition_score,none": 305.3571428571429,
22
+ "mme_cognition_score_stderr,none": "N/A",
23
+ "mme_percetion_score,none": 1400.8141256502602,
24
+ "mme_percetion_score_stderr,none": "N/A",
25
+ "alias": "mme"
26
+ },
27
+ "mmerealworld_lite": {
28
+ "mme_realworld_score,none": 0.31474726420010424,
29
+ "mme_realworld_score_stderr,none": "N/A",
30
+ "alias": "mmerealworld_lite"
31
+ },
32
+ "mmmu_pro_standard": {
33
+ "mmmu_acc,none": 0.25838,
34
+ "mmmu_acc_stderr,none": "N/A",
35
+ "alias": "mmmu_pro_standard"
36
+ },
37
+ "mmmu_val": {
38
+ "mmmu_acc,none": 0.42444,
39
+ "mmmu_acc_stderr,none": "N/A",
40
+ "alias": "mmmu_val"
41
+ },
42
+ "mmstar": {
43
+ "coarse perception,none": 0.7007139127935286,
44
+ "coarse perception_stderr,none": "N/A",
45
+ "fine-grained perception,none": 0.36133628101203635,
46
+ "fine-grained perception_stderr,none": "N/A",
47
+ "instance reasoning,none": 0.5332709723793451,
48
+ "instance reasoning_stderr,none": "N/A",
49
+ "logical reasoning,none": 0.37952487556447956,
50
+ "logical reasoning_stderr,none": "N/A",
51
+ "math,none": 0.2904916243428673,
52
+ "math_stderr,none": "N/A",
53
+ "science & technology,none": 0.2396233390154238,
54
+ "science & technology_stderr,none": "N/A",
55
+ "alias": "mmstar"
56
+ },
57
+ "scienceqa_img": {
58
+ "exact_match,none": 0.7481408031730292,
59
+ "exact_match_stderr,none": 0.009667747830565338,
60
+ "alias": "scienceqa_img"
61
+ },
62
+ "seedbench_2_plus": {
63
+ "seedbench_2_plus_Chart,none": 0.4765432098765432,
64
+ "seedbench_2_plus_Chart_stderr,none": "N/A",
65
+ "seedbench_2_plus_all,none": 0.47474747474747475,
66
+ "seedbench_2_plus_all_stderr,none": "N/A",
67
+ "seedbench_2_plus_Web,none": 0.509090909090909,
68
+ "seedbench_2_plus_Web_stderr,none": "N/A",
69
+ "seedbench_2_plus_Map,none": 0.4448574969021066,
70
+ "seedbench_2_plus_Map_stderr,none": "N/A",
71
+ "alias": "seedbench_2_plus"
72
+ },
73
+ "textvqa_val": {
74
+ "exact_match,none": 0.4206200003147125,
75
+ "exact_match_stderr,none": 0.0067633506191732045,
76
+ "submission,none": null,
77
+ "submission_stderr,none": "N/A",
78
+ "alias": "textvqa_val"
79
+ }
80
+ },
81
+ "configs": {
82
+ "ai2d": {
83
+ "task": "ai2d",
84
+ "dataset_path": "lmms-lab/ai2d",
85
+ "dataset_kwargs": {
86
+ "token": true
87
+ },
88
+ "test_split": "test",
89
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7f5391f408b0>",
90
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7f5391f40b80>",
91
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7f5391f48160>",
92
+ "description": "",
93
+ "target_delimiter": " ",
94
+ "fewshot_delimiter": "\n\n",
95
+ "metric_list": [
96
+ {
97
+ "metric": "exact_match",
98
+ "aggregation": "mean",
99
+ "higher_is_better": true,
100
+ "ignore_case": true,
101
+ "ignore_punctuation": true
102
+ }
103
+ ],
104
+ "output_type": "generate_until",
105
+ "generation_kwargs": {
106
+ "max_new_tokens": 512,
107
+ "temperature": 0.0,
108
+ "do_sample": false,
109
+ "until": [
110
+ "\n\n"
111
+ ]
112
+ },
113
+ "repeats": 1,
114
+ "filter_list": [
115
+ {
116
+ "name": "flexible-extract",
117
+ "filter": [
118
+ {
119
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
120
+ "group_select": 0,
121
+ "ignore_case": true,
122
+ "ignore_punctuation": true,
123
+ "regex_pattern": "([A-Z])\\."
124
+ }
125
+ ]
126
+ }
127
+ ],
128
+ "should_decontaminate": false,
129
+ "metadata": [
130
+ {
131
+ "version": 0.0
132
+ }
133
+ ],
134
+ "model_specific_prompt_kwargs": {
135
+ "default": {
136
+ "prompt_format": "mcq",
137
+ "pre_prompt": "",
138
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
139
+ },
140
+ "gpt4v": {
141
+ "prompt_format": "mcq",
142
+ "pre_prompt": "",
143
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
144
+ },
145
+ "qwen_vl": {
146
+ "prompt_format": "qa",
147
+ "pre_prompt": "",
148
+ "post_prompt": " Answer:"
149
+ },
150
+ "xcomposer2_4khd": {
151
+ "prompt_format": "mcq_xcomposer",
152
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
153
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
154
+ }
155
+ },
156
+ "model_specific_target_kwargs": {
157
+ "default": "mcq",
158
+ "qwen_vl": "qa"
159
+ }
160
+ },
161
+ "infovqa_val": {
162
+ "task": "infovqa_val",
163
+ "dataset_path": "lmms-lab/DocVQA",
164
+ "dataset_name": "InfographicVQA",
165
+ "dataset_kwargs": {
166
+ "token": true
167
+ },
168
+ "test_split": "validation",
169
+ "doc_to_visual": "<function infovqa_doc_to_visual at 0x7f53c4b55940>",
170
+ "doc_to_text": "<function infovqa_doc_to_text at 0x7f53c4b55c10>",
171
+ "doc_to_target": "answers",
172
+ "description": "",
173
+ "target_delimiter": " ",
174
+ "fewshot_delimiter": "\n\n",
175
+ "metric_list": [
176
+ {
177
+ "metric": "anls",
178
+ "aggregation": "mean",
179
+ "higher_is_better": true
180
+ }
181
+ ],
182
+ "output_type": "generate_until",
183
+ "generation_kwargs": {
184
+ "max_new_tokens": 32,
185
+ "temperature": 0.0,
186
+ "do_sample": false,
187
+ "until": [
188
+ "\n\n"
189
+ ]
190
+ },
191
+ "repeats": 1,
192
+ "should_decontaminate": false,
193
+ "model_specific_prompt_kwargs": {
194
+ "default": {
195
+ "pre_prompt": "",
196
+ "post_prompt": "\nAnswer the question using a single word or phrase."
197
+ }
198
+ }
199
+ },
200
+ "mmbench_en_dev": {
201
+ "task": "mmbench_en_dev",
202
+ "dataset_path": "lmms-lab/MMBench",
203
+ "dataset_name": "en",
204
+ "dataset_kwargs": {
205
+ "token": true
206
+ },
207
+ "test_split": "dev",
208
+ "doc_to_visual": "<function mmbench_doc_to_visual at 0x7f53f11925e0>",
209
+ "doc_to_text": "<function mmbench_doc_to_text at 0x7f53f1192af0>",
210
+ "doc_to_target": "answer",
211
+ "process_results": "<function mmbench_process_results at 0x7f53f119e040>",
212
+ "description": "",
213
+ "target_delimiter": " ",
214
+ "fewshot_delimiter": "\n\n",
215
+ "metric_list": [
216
+ {
217
+ "metric": "gpt_eval_score",
218
+ "aggregation": "<function mmbench_aggregate_dev_results_eval at 0x7f53f1188e50>",
219
+ "higher_is_better": true
220
+ },
221
+ {
222
+ "metric": "submission",
223
+ "aggregation": "<function mmbench_aggregate_dev_results_submission at 0x7f53f11923a0>",
224
+ "higher_is_better": true
225
+ }
226
+ ],
227
+ "output_type": "generate_until",
228
+ "generation_kwargs": {
229
+ "until": [
230
+ "ASSISTANT:"
231
+ ],
232
+ "max_new_tokens": 1024,
233
+ "temperature": 0.0,
234
+ "top_p": 1.0,
235
+ "num_beams": 1,
236
+ "do_sample": false,
237
+ "image_aspect_ratio": "original"
238
+ },
239
+ "repeats": 1,
240
+ "should_decontaminate": false,
241
+ "model_specific_prompt_kwargs": {
242
+ "default": {
243
+ "pre_prompt": "",
244
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
245
+ }
246
+ },
247
+ "model_specific_generation_kwargs": {
248
+ "llava": {
249
+ "image_aspect_ratio": "original"
250
+ }
251
+ }
252
+ },
253
+ "mme": {
254
+ "task": "mme",
255
+ "dataset_path": "lmms-lab/MME",
256
+ "dataset_kwargs": {
257
+ "token": false
258
+ },
259
+ "test_split": "test",
260
+ "doc_to_visual": "<function mme_doc_to_visual at 0x7f53f2c6baf0>",
261
+ "doc_to_text": "<function mme_doc_to_text at 0x7f53f1213310>",
262
+ "doc_to_target": "answer",
263
+ "process_results": "<function mme_process_results at 0x7f53f12138b0>",
264
+ "description": "",
265
+ "target_delimiter": " ",
266
+ "fewshot_delimiter": "\n\n",
267
+ "metric_list": [
268
+ {
269
+ "metric": "mme_percetion_score",
270
+ "aggregation": "<function mme_aggregate_results at 0x7f53f1213dc0>",
271
+ "higher_is_better": true
272
+ },
273
+ {
274
+ "metric": "mme_cognition_score",
275
+ "aggregation": "<function mme_aggregate_results at 0x7f53f121a280>",
276
+ "higher_is_better": true
277
+ }
278
+ ],
279
+ "output_type": "generate_until",
280
+ "generation_kwargs": {
281
+ "max_new_tokens": 16,
282
+ "temperature": 0.0,
283
+ "top_p": 1.0,
284
+ "num_beams": 1,
285
+ "do_sample": false,
286
+ "until": [
287
+ "\n\n"
288
+ ]
289
+ },
290
+ "repeats": 1,
291
+ "should_decontaminate": false,
292
+ "metadata": [
293
+ {
294
+ "version": 0.0
295
+ }
296
+ ],
297
+ "model_specific_prompt_kwargs": {
298
+ "default": {
299
+ "pre_prompt": "",
300
+ "post_prompt": "\nAnswer the question using a single word or phrase."
301
+ },
302
+ "gpt4v": {
303
+ "pre_prompt": "",
304
+ "post_prompt": "\nAnswer the question with Yes or No."
305
+ },
306
+ "qwen_vl": {
307
+ "pre_prompt": "",
308
+ "post_prompt": " Answer:"
309
+ },
310
+ "otterhd": {
311
+ "pre_prompt": "",
312
+ "post_prompt": " Answer:"
313
+ },
314
+ "xcomposer2_4khd": {
315
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
316
+ "post_prompt": " Answer this question briefly[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
317
+ }
318
+ }
319
+ },
320
+ "mmerealworld_lite": {
321
+ "task": "mmerealworld_lite",
322
+ "dataset_path": "yifanzhang114/MME-RealWorld-lite-lmms-eval",
323
+ "dataset_kwargs": {
324
+ "token": true
325
+ },
326
+ "test_split": "train",
327
+ "doc_to_visual": "<function mme_realworld_doc_to_visual at 0x7f53926e9a60>",
328
+ "doc_to_text": "<function mme_realworld_doc_to_text at 0x7f53926ed550>",
329
+ "doc_to_target": "answer",
330
+ "process_results": "<function mme_realworld_process_results at 0x7f53926f31f0>",
331
+ "description": "",
332
+ "target_delimiter": " ",
333
+ "fewshot_delimiter": "\n\n",
334
+ "metric_list": [
335
+ {
336
+ "metric": "mme_realworld_score",
337
+ "aggregation": "<function mme_realworld_aggregate_results at 0x7f53926f3d30>",
338
+ "higher_is_better": true
339
+ }
340
+ ],
341
+ "output_type": "generate_until",
342
+ "generation_kwargs": {
343
+ "max_new_tokens": 16,
344
+ "temperature": 0.0,
345
+ "top_p": 1.0,
346
+ "num_beams": 1,
347
+ "do_sample": false,
348
+ "until": [
349
+ "\n\n"
350
+ ]
351
+ },
352
+ "repeats": 1,
353
+ "should_decontaminate": false,
354
+ "metadata": [
355
+ {
356
+ "version": 0.0
357
+ }
358
+ ],
359
+ "model_specific_prompt_kwargs": {
360
+ "default": {
361
+ "pre_prompt": "",
362
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
363
+ },
364
+ "gpt4v": {
365
+ "pre_prompt": "",
366
+ "post_prompt": "\nSelect the best answer to the above multiple-choice question based on the image. Respond with only the letter (A, B, C, D, or E) of the correct option."
367
+ },
368
+ "xcomposer2_4khd": {
369
+ "pre_prompt": "[UNUSED_TOKEN_146]user\n",
370
+ "post_prompt": " Answer this question with A, B, C, or D.[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\n"
371
+ }
372
+ }
373
+ },
374
+ "mmmu_pro_standard": {
375
+ "task": "mmmu_pro_standard",
376
+ "dataset_path": "MMMU/MMMU_Pro",
377
+ "dataset_name": "standard (10 options)",
378
+ "test_split": "test",
379
+ "doc_to_visual": "<function mmmu_pro_doc_to_visual at 0x7f53f115f940>",
380
+ "doc_to_text": "<function mmmu_pro_doc_to_text at 0x7f53f116c700>",
381
+ "doc_to_target": "{{answer}}",
382
+ "process_results": "<function mmmu_pro_process_results at 0x7f53f1172670>",
383
+ "description": "",
384
+ "target_delimiter": " ",
385
+ "fewshot_delimiter": "\n\n",
386
+ "metric_list": [
387
+ {
388
+ "metric": "mmmu_acc",
389
+ "aggregation": "<function mmmu_pro_aggregate_results at 0x7f53f117e5e0>",
390
+ "higher_is_better": true
391
+ }
392
+ ],
393
+ "output_type": "generate_until",
394
+ "generation_kwargs": {
395
+ "max_new_tokens": 256,
396
+ "until": [
397
+ "\n\n"
398
+ ]
399
+ },
400
+ "repeats": 1,
401
+ "should_decontaminate": false,
402
+ "metadata": {
403
+ "version": 0.0,
404
+ "interleaved_format": false
405
+ },
406
+ "model_specific_prompt_kwargs": {
407
+ "default": {
408
+ "pre_prompt": "",
409
+ "post_prompt": "Answer with the option letter from the given choices directly."
410
+ }
411
+ }
412
+ },
413
+ "mmmu_val": {
414
+ "task": "mmmu_val",
415
+ "dataset_path": "lmms-lab/MMMU",
416
+ "test_split": "validation",
417
+ "doc_to_visual": "<function mmmu_doc_to_visual at 0x7f53a412b8b0>",
418
+ "doc_to_text": "<function mmmu_doc_to_text at 0x7f53a40a95e0>",
419
+ "doc_to_target": "answer",
420
+ "process_results": "<function mmmu_process_results at 0x7f53a40764c0>",
421
+ "description": "",
422
+ "target_delimiter": " ",
423
+ "fewshot_delimiter": "\n\n",
424
+ "metric_list": [
425
+ {
426
+ "metric": "mmmu_acc",
427
+ "aggregation": "<function mmmu_aggregate_results at 0x7f53a3eaf430>",
428
+ "higher_is_better": true
429
+ }
430
+ ],
431
+ "output_type": "generate_until",
432
+ "generation_kwargs": {
433
+ "max_new_tokens": 128,
434
+ "until": [
435
+ "\n\n"
436
+ ],
437
+ "image_aspect_ratio": "original"
438
+ },
439
+ "repeats": 1,
440
+ "should_decontaminate": false,
441
+ "metadata": [
442
+ {
443
+ "version": 0.0
444
+ }
445
+ ],
446
+ "model_specific_generation_kwargs": {
447
+ "llava": {
448
+ "image_aspect_ratio": "original"
449
+ }
450
+ }
451
+ },
452
+ "mmstar": {
453
+ "task": "mmstar",
454
+ "dataset_path": "Lin-Chen/MMStar",
455
+ "dataset_kwargs": {
456
+ "token": true
457
+ },
458
+ "test_split": "val",
459
+ "doc_to_visual": "<function mmstar_doc_to_visual at 0x7f5391e97ee0>",
460
+ "doc_to_text": "<function mmstar_doc_to_text at 0x7f5391ea33a0>",
461
+ "doc_to_target": "answer",
462
+ "process_results": "<function mmstar_process_results at 0x7f5391ea38b0>",
463
+ "description": "",
464
+ "target_delimiter": " ",
465
+ "fewshot_delimiter": "\n\n",
466
+ "metric_list": [
467
+ {
468
+ "metric": "coarse perception",
469
+ "aggregation": "<function mmstar_aggregate_results at 0x7f5391ea3d30>",
470
+ "higher_is_better": true
471
+ },
472
+ {
473
+ "metric": "fine-grained perception",
474
+ "aggregation": "<function mmstar_aggregate_results at 0x7f5391e27160>",
475
+ "higher_is_better": true
476
+ },
477
+ {
478
+ "metric": "instance reasoning",
479
+ "aggregation": "<function mmstar_aggregate_results at 0x7f5391e27550>",
480
+ "higher_is_better": true
481
+ },
482
+ {
483
+ "metric": "logical reasoning",
484
+ "aggregation": "<function mmstar_aggregate_results at 0x7f5391e27940>",
485
+ "higher_is_better": true
486
+ },
487
+ {
488
+ "metric": "science & technology",
489
+ "aggregation": "<function mmstar_aggregate_results at 0x7f5391e27d30>",
490
+ "higher_is_better": true
491
+ },
492
+ {
493
+ "metric": "math",
494
+ "aggregation": "<function mmstar_aggregate_results at 0x7f5391e2e160>",
495
+ "higher_is_better": true
496
+ }
497
+ ],
498
+ "output_type": "generate_until",
499
+ "generation_kwargs": {
500
+ "until": [
501
+ "\n\n"
502
+ ],
503
+ "do_sample": false
504
+ },
505
+ "repeats": 1,
506
+ "should_decontaminate": false,
507
+ "metadata": [
508
+ {
509
+ "version": 0.0
510
+ }
511
+ ],
512
+ "model_specific_prompt_kwargs": {
513
+ "default": {
514
+ "pre_prompt": "",
515
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly"
516
+ }
517
+ }
518
+ },
519
+ "scienceqa_img": {
520
+ "task": "scienceqa_img",
521
+ "dataset_path": "lmms-lab/ScienceQA",
522
+ "dataset_name": "ScienceQA-IMG",
523
+ "dataset_kwargs": {
524
+ "token": true
525
+ },
526
+ "test_split": "test",
527
+ "doc_to_visual": "<function sqa_doc_to_visual at 0x7f53bcffdf70>",
528
+ "doc_to_text": "<function sqa_doc_to_text at 0x7f53bceb71f0>",
529
+ "doc_to_target": "<function sqa_doc_to_target at 0x7f53bceb75e0>",
530
+ "process_results": "<function sqa_process_results at 0x7f53bceb7940>",
531
+ "description": "",
532
+ "target_delimiter": " ",
533
+ "fewshot_delimiter": "\n\n",
534
+ "metric_list": [
535
+ {
536
+ "metric": "exact_match",
537
+ "aggregation": "mean",
538
+ "higher_is_better": true,
539
+ "ignore_case": true,
540
+ "ignore_punctuation": true
541
+ }
542
+ ],
543
+ "output_type": "generate_until",
544
+ "generation_kwargs": {
545
+ "max_new_tokens": 16,
546
+ "temperature": 0.0,
547
+ "do_sample": false,
548
+ "until": [
549
+ "\n\n"
550
+ ],
551
+ "image_aspect_ratio": "original"
552
+ },
553
+ "repeats": 1,
554
+ "should_decontaminate": false,
555
+ "metadata": [
556
+ {
557
+ "version": 0.0
558
+ }
559
+ ],
560
+ "model_specific_prompt_kwargs": {
561
+ "default": {
562
+ "format": "default",
563
+ "pre_prompt": "",
564
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
565
+ },
566
+ "qwen_vl": {
567
+ "format": "qwen_vl"
568
+ },
569
+ "idefics2": {
570
+ "format": "default",
571
+ "pre_prompt": "",
572
+ "post_prompt": "\nAnswer:"
573
+ }
574
+ },
575
+ "model_specific_generation_kwargs": {
576
+ "llava": {
577
+ "image_aspect_ratio": "original"
578
+ }
579
+ }
580
+ },
581
+ "seedbench_2_plus": {
582
+ "task": "seedbench_2_plus",
583
+ "dataset_path": "doolayer/SEED-Bench-2-Plus",
584
+ "dataset_kwargs": {
585
+ "token": true
586
+ },
587
+ "test_split": "test",
588
+ "doc_to_visual": "<function seed_doc_to_visual at 0x7f53b9d8a4c0>",
589
+ "doc_to_text": "<function seed_doc_to_text at 0x7f53b9d8aaf0>",
590
+ "doc_to_target": "answer",
591
+ "process_results": "<function seed_process_result at 0x7f53b9ab5040>",
592
+ "description": "",
593
+ "target_delimiter": " ",
594
+ "fewshot_delimiter": "\n\n",
595
+ "metric_list": [
596
+ {
597
+ "metric": "seedbench_2_plus_Chart",
598
+ "aggregation": "<function seed_aggregation_result at 0x7f53b9ab5550>",
599
+ "higher_is_better": true
600
+ },
601
+ {
602
+ "metric": "seedbench_2_plus_Map",
603
+ "aggregation": "<function seed_aggregation_result at 0x7f53b9ab59d0>",
604
+ "higher_is_better": true
605
+ },
606
+ {
607
+ "metric": "seedbench_2_plus_Web",
608
+ "aggregation": "<function seed_aggregation_result at 0x7f53b9ab5e50>",
609
+ "higher_is_better": true
610
+ },
611
+ {
612
+ "metric": "seedbench_2_plus_all",
613
+ "aggregation": "<function seed_aggregation_result at 0x7f53b93d5310>",
614
+ "higher_is_better": true
615
+ }
616
+ ],
617
+ "output_type": "generate_until",
618
+ "generation_kwargs": {
619
+ "until": [
620
+ "ASSISTANT:"
621
+ ],
622
+ "max_new_tokens": 16,
623
+ "image_aspect_ratio": "original"
624
+ },
625
+ "repeats": 1,
626
+ "should_decontaminate": false,
627
+ "metadata": [
628
+ {
629
+ "version": 0.0
630
+ }
631
+ ],
632
+ "model_specific_prompt_kwargs": {
633
+ "llava": {
634
+ "img_token": "<image>",
635
+ "post_prompt": "Answer with the option's letter from the given choices directly."
636
+ },
637
+ "gpt4V": {
638
+ "img_token": "<image>",
639
+ "post_prompt": "Answer with the option's letter from the given choices directly."
640
+ },
641
+ "default": {
642
+ "img_token": "<image>",
643
+ "post_prompt": "Answer with the option's letter from the given choices directly."
644
+ }
645
+ }
646
+ },
647
+ "textvqa_val": {
648
+ "task": "textvqa_val",
649
+ "dataset_path": "lmms-lab/textvqa",
650
+ "test_split": "validation",
651
+ "doc_to_visual": "<function textvqa_doc_to_visual at 0x7f53be318b80>",
652
+ "doc_to_text": "<function textvqa_doc_to_text at 0x7f53be333040>",
653
+ "doc_to_target": "answer",
654
+ "process_results": "<function textvqa_process_results at 0x7f53be333310>",
655
+ "description": "",
656
+ "target_delimiter": " ",
657
+ "fewshot_delimiter": "\n\n",
658
+ "metric_list": [
659
+ {
660
+ "metric": "exact_match",
661
+ "aggregation": "mean",
662
+ "higher_is_better": true,
663
+ "ignore_case": true,
664
+ "ignore_punctuation": true
665
+ },
666
+ {
667
+ "metric": "submission",
668
+ "aggregation": "<function textvqa_aggregate_submissions at 0x7f53be318940>",
669
+ "higher_is_better": true
670
+ }
671
+ ],
672
+ "output_type": "generate_until",
673
+ "generation_kwargs": {
674
+ "until": [
675
+ "ASSISTANT:"
676
+ ]
677
+ },
678
+ "repeats": 1,
679
+ "should_decontaminate": false,
680
+ "model_specific_prompt_kwargs": {
681
+ "default": {
682
+ "pre_prompt": "",
683
+ "post_prompt": "\nAnswer the question using a single word or phrase.",
684
+ "ocr": false
685
+ },
686
+ "qwen_vl": {
687
+ "pre_prompt": "",
688
+ "post_prompt": " Answer:"
689
+ }
690
+ }
691
+ }
692
+ },
693
+ "versions": {
694
+ "ai2d": "Yaml",
695
+ "infovqa_val": "Yaml",
696
+ "mmbench_en_dev": "Yaml",
697
+ "mme": "Yaml",
698
+ "mmerealworld_lite": "Yaml",
699
+ "mmmu_pro_standard": "Yaml",
700
+ "mmmu_val": "Yaml",
701
+ "mmstar": "Yaml",
702
+ "scienceqa_img": "Yaml",
703
+ "seedbench_2_plus": "Yaml",
704
+ "textvqa_val": "Yaml"
705
+ },
706
+ "n-shot": {
707
+ "ai2d": 0,
708
+ "infovqa_val": 0,
709
+ "mmbench_en_dev": 0,
710
+ "mme": 0,
711
+ "mmerealworld_lite": 0,
712
+ "mmmu_pro_standard": 0,
713
+ "mmmu_val": 0,
714
+ "mmstar": 0,
715
+ "scienceqa_img": 0,
716
+ "seedbench_2_plus": 0,
717
+ "textvqa_val": 0
718
+ },
719
+ "model_configs": {
720
+ "model": "llava",
721
+ "model_args": "pretrained=/cm/archive/namnv78_new/revise_checkpoints/Xphi35-siglip224/SMOE/665K36/revise_Full_smoe_sharev3,conv_template=phi35",
722
+ "batch_size": "1",
723
+ "device": null,
724
+ "limit": null,
725
+ "bootstrap_iters": 100000,
726
+ "gen_kwargs": ""
727
+ },
728
+ "git_hash": "289c7fe5"
729
+ }
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/scienceqa_img.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/seedbench_2_plus.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/mmbench_en_dev_results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"overall_acc": 0.7242268041237113, "category_acc": {"action_recognition": 0.8888888888888888, "attribute_comparison": 0.7272727272727273, "attribute_recognition": 0.8648648648648649, "celebrity_recognition": 0.7878787878787878, "function_reasoning": 0.8607594936708861, "future_prediction": 0.525, "identity_reasoning": 0.9777777777777777, "image_emotion": 0.82, "image_quality": 0.5094339622641509, "image_scene": 0.9615384615384616, "image_style": 0.9056603773584906, "image_topic": 0.8611111111111112, "nature_relation": 0.6041666666666666, "object_localization": 0.5308641975308642, "ocr": 0.717948717948718, "physical_property_reasoning": 0.5866666666666667, "physical_relation": 0.4583333333333333, "social_relation": 0.7906976744186046, "spatial_relationship": 0.3333333333333333, "structuralized_imagetext_understanding": 0.47435897435897434}, "l2_category_acc": {"attribute_reasoning": 0.7839195979899497, "coarse_perception": 0.8344594594594594, "finegrained_perception (cross-instance)": 0.6643356643356644, "finegrained_perception (instance-level)": 0.726962457337884, "logic_reasoning": 0.4915254237288136, "relation_reasoning": 0.6434782608695652}}
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/mmbench_en_dev_results.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c865ae8c4bebda3a05bc5d1d735de59e432ba4ef891218b9b6aba6f23fd33d8f
3
+ size 865128
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/submissions/textvqa_submission_2025-07-17-11-00-08.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1127_llava...l_mme_llava_model_args_5a2714/textvqa_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2008f45cb5e06d281e3c245fa4123d5cb8f5f42dbbceecd0c544301fea7fc47
3
+ size 13146759
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/gpt_response/hallusion_output_vd_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/gpt_response/hallusion_output_vs_model.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/hallusion_bench_image.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/mathvista_testmini.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/results.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hallusion_bench_image": {
4
+ "aAcc,none": 41.4301,
5
+ "aAcc_stderr,none": "N/A",
6
+ "fAcc,none": 17.052,
7
+ "fAcc_stderr,none": "N/A",
8
+ "qAcc,none": 11.4286,
9
+ "qAcc_stderr,none": "N/A",
10
+ "alias": "hallusion_bench_image"
11
+ },
12
+ "mathvista_testmini": {
13
+ "gpt_eval_score,none": 30.6,
14
+ "gpt_eval_score_stderr,none": "N/A",
15
+ "alias": "mathvista_testmini"
16
+ }
17
+ },
18
+ "configs": {
19
+ "hallusion_bench_image": {
20
+ "task": "hallusion_bench_image",
21
+ "dataset_path": "lmms-lab/HallusionBench",
22
+ "dataset_kwargs": {
23
+ "token": true
24
+ },
25
+ "test_split": "image",
26
+ "doc_to_visual": "<function hb_doc_to_visual at 0x7f66394cf4c0>",
27
+ "doc_to_text": "<function hb_doc_to_text at 0x7f66394cfc10>",
28
+ "doc_to_target": "gt_answer_details",
29
+ "process_results": "<function hb_process_results at 0x7f6639462430>",
30
+ "description": "",
31
+ "target_delimiter": " ",
32
+ "fewshot_delimiter": "\n\n",
33
+ "metric_list": [
34
+ {
35
+ "metric": "aAcc",
36
+ "aggregation": "<function hb_aggregation_result_aAcc at 0x7f6639462dc0>",
37
+ "higher_is_better": true
38
+ },
39
+ {
40
+ "metric": "qAcc",
41
+ "aggregation": "<function hb_aggregation_result_qAcc at 0x7f663941b430>",
42
+ "higher_is_better": true
43
+ },
44
+ {
45
+ "metric": "fAcc",
46
+ "aggregation": "<function hb_aggregation_result_fAcc at 0x7f663941bc10>",
47
+ "higher_is_better": true
48
+ }
49
+ ],
50
+ "output_type": "generate_until",
51
+ "generation_kwargs": {
52
+ "max_new_tokens": 128,
53
+ "temperature": 0.0,
54
+ "top_p": 1.0,
55
+ "num_beams": 1,
56
+ "do_sample": false,
57
+ "until": [
58
+ "\n\n"
59
+ ]
60
+ },
61
+ "repeats": 1,
62
+ "should_decontaminate": false,
63
+ "metadata": [
64
+ {
65
+ "version": 0.0
66
+ }
67
+ ],
68
+ "model_specific_prompt_kwargs": {
69
+ "default": {
70
+ "pre_prompt": "",
71
+ "post_prompt": ""
72
+ }
73
+ }
74
+ },
75
+ "mathvista_testmini": {
76
+ "task": "mathvista_testmini",
77
+ "dataset_path": "AI4Math/MathVista",
78
+ "dataset_kwargs": {
79
+ "token": true
80
+ },
81
+ "test_split": "testmini",
82
+ "doc_to_visual": "<function mathvista_doc_to_visual at 0x7f660f16a8b0>",
83
+ "doc_to_text": "<function mathvista_doc_to_text at 0x7f660f172f70>",
84
+ "doc_to_target": "answer",
85
+ "process_results": "<function mathvista_process_results at 0x7f660f180670>",
86
+ "description": "",
87
+ "target_delimiter": " ",
88
+ "fewshot_delimiter": "\n\n",
89
+ "metric_list": [
90
+ {
91
+ "metric": "gpt_eval_score",
92
+ "aggregation": "<function mathvista_aggregate_results at 0x7f660f18bd30>",
93
+ "higher_is_better": true
94
+ }
95
+ ],
96
+ "output_type": "generate_until",
97
+ "generation_kwargs": {
98
+ "until": [
99
+ "ASSISTANT:"
100
+ ],
101
+ "max_new_tokens": 1024,
102
+ "temperature": 0.0,
103
+ "top_p": 1.0,
104
+ "num_beams": 1,
105
+ "do_sample": false,
106
+ "image_aspect_ratio": "original"
107
+ },
108
+ "repeats": 1,
109
+ "should_decontaminate": false,
110
+ "model_specific_prompt_kwargs": {
111
+ "default": {
112
+ "shot_type": "format-prompt",
113
+ "shot": 0,
114
+ "use_caption": false,
115
+ "use_ocr": false
116
+ },
117
+ "phi3v": {
118
+ "shot_type": "solution"
119
+ }
120
+ },
121
+ "model_specific_generation_kwargs": {
122
+ "llava": {
123
+ "image_aspect_ratio": "original"
124
+ }
125
+ }
126
+ }
127
+ },
128
+ "versions": {
129
+ "hallusion_bench_image": "Yaml",
130
+ "mathvista_testmini": "Yaml"
131
+ },
132
+ "n-shot": {
133
+ "hallusion_bench_image": 0,
134
+ "mathvista_testmini": 0
135
+ },
136
+ "model_configs": {
137
+ "model": "llava",
138
+ "model_args": "pretrained=/cm/archive/namnv78_new/revise_checkpoints/Xphi35-siglip224/SMOE/665K36/revise_Full_smoe_sharev3,conv_template=phi35",
139
+ "batch_size": "1",
140
+ "device": null,
141
+ "limit": null,
142
+ "bootstrap_iters": 100000,
143
+ "gen_kwargs": ""
144
+ },
145
+ "git_hash": "289c7fe5"
146
+ }
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1201_llava...image_llava_model_args_5a2714/submissions/mathvista_testmini_scores.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/ocrbench.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/results.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ocrbench": {
4
+ "ocrbench_accuracy,none": 0.326,
5
+ "ocrbench_accuracy_stderr,none": "N/A",
6
+ "alias": "ocrbench"
7
+ }
8
+ },
9
+ "configs": {
10
+ "ocrbench": {
11
+ "task": "ocrbench",
12
+ "dataset_path": "echo840/OCRBench",
13
+ "dataset_kwargs": {
14
+ "token": true
15
+ },
16
+ "test_split": "test",
17
+ "doc_to_visual": "<function ocrbench_doc_to_visual at 0x7fd6ac655f70>",
18
+ "doc_to_text": "<function ocrbench_doc_to_text at 0x7fd69a22e550>",
19
+ "doc_to_target": "answer",
20
+ "process_results": "<function ocrbench_process_results at 0x7fd69a22e8b0>",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "ocrbench_accuracy",
27
+ "aggregation": "<function ocrbench_aggregate_accuracy at 0x7fd69a22ec10>",
28
+ "higher_is_better": true
29
+ }
30
+ ],
31
+ "output_type": "generate_until",
32
+ "generation_kwargs": {
33
+ "max_new_tokens": 128,
34
+ "temperature": 0.0,
35
+ "top_p": 1.0,
36
+ "num_beams": 1,
37
+ "do_sample": false,
38
+ "until": [
39
+ "\n\n"
40
+ ]
41
+ },
42
+ "repeats": 1,
43
+ "should_decontaminate": false,
44
+ "metadata": [
45
+ {
46
+ "version": 0.0
47
+ }
48
+ ]
49
+ }
50
+ },
51
+ "versions": {
52
+ "ocrbench": "Yaml"
53
+ },
54
+ "n-shot": {
55
+ "ocrbench": 0
56
+ },
57
+ "model_configs": {
58
+ "model": "llava",
59
+ "model_args": "pretrained=/cm/archive/namnv78_new/revise_checkpoints/Xphi35-siglip224/SMOE/665K36/revise_Full_smoe_sharev3,conv_template=phi35",
60
+ "batch_size": "1",
61
+ "device": null,
62
+ "limit": null,
63
+ "bootstrap_iters": 100000,
64
+ "gen_kwargs": ""
65
+ },
66
+ "git_hash": "289c7fe5"
67
+ }
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1220_llava...bench_llava_model_args_5a2714/results/ocrbench_results.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ######################### OCRBench #############################
2
+ Text Recognition(Total 300): 183
3
+ ---------------- Details of Recognition Score ------------------
4
+ Regular Text Recognition(Total 50): 45
5
+ Irregular Text Recognition(Total 50): 43
6
+ Artistic Text Recognition(Total 50): 46
7
+ Handwriting Recognition(Total 50): 28
8
+ Digit String Recognition(Total 50): 13
9
+ Non-Semantic Text Recognition(Total 50): 8
10
+ ----------------------------------------------------------------
11
+ Scene Text-centric VQA(Total 200): 114
12
+ ----------------------------------------------------------------
13
+ Doc-oriented VQA(Total 200): 20
14
+ ----------------------------------------------------------------
15
+ Key Information Extraction(Total 200): 9
16
+ Handwritten Mathematical Expression Recognition(Total 100): 0
17
+ --------------------- Final Score ------------------------------
18
+ Final Score(Total 1000): 326
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1739_llava_v1.5_ai2d_llava_model_args_5a2714/results.json ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6463730569948186,
5
+ "exact_match_stderr,flexible-extract": 0.008604903043803526,
6
+ "alias": "ai2d"
7
+ }
8
+ },
9
+ "configs": {
10
+ "ai2d": {
11
+ "task": "ai2d",
12
+ "dataset_path": "lmms-lab/ai2d",
13
+ "dataset_kwargs": {
14
+ "token": true
15
+ },
16
+ "test_split": "test",
17
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7f74c14778b0>",
18
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7f74c1477b80>",
19
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7f74c147f160>",
20
+ "description": "",
21
+ "target_delimiter": " ",
22
+ "fewshot_delimiter": "\n\n",
23
+ "metric_list": [
24
+ {
25
+ "metric": "exact_match",
26
+ "aggregation": "mean",
27
+ "higher_is_better": true,
28
+ "ignore_case": true,
29
+ "ignore_punctuation": true
30
+ }
31
+ ],
32
+ "output_type": "generate_until",
33
+ "generation_kwargs": {
34
+ "max_new_tokens": 512,
35
+ "temperature": 0.0,
36
+ "do_sample": false,
37
+ "until": [
38
+ "\n\n"
39
+ ]
40
+ },
41
+ "repeats": 1,
42
+ "filter_list": [
43
+ {
44
+ "name": "flexible-extract",
45
+ "filter": [
46
+ {
47
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
48
+ "group_select": 0,
49
+ "ignore_case": true,
50
+ "ignore_punctuation": true,
51
+ "regex_pattern": "([A-Z])\\."
52
+ }
53
+ ]
54
+ }
55
+ ],
56
+ "should_decontaminate": false,
57
+ "metadata": [
58
+ {
59
+ "version": 0.0
60
+ }
61
+ ],
62
+ "model_specific_prompt_kwargs": {
63
+ "default": {
64
+ "prompt_format": "mcq",
65
+ "pre_prompt": "",
66
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
67
+ },
68
+ "gpt4v": {
69
+ "prompt_format": "mcq",
70
+ "pre_prompt": "",
71
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
72
+ },
73
+ "qwen_vl": {
74
+ "prompt_format": "qa",
75
+ "pre_prompt": "",
76
+ "post_prompt": " Answer:"
77
+ },
78
+ "xcomposer2_4khd": {
79
+ "prompt_format": "mcq_xcomposer",
80
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
81
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
82
+ }
83
+ },
84
+ "model_specific_target_kwargs": {
85
+ "default": "mcq",
86
+ "qwen_vl": "qa"
87
+ }
88
+ }
89
+ },
90
+ "versions": {
91
+ "ai2d": "Yaml"
92
+ },
93
+ "n-shot": {
94
+ "ai2d": 0
95
+ },
96
+ "model_configs": {
97
+ "model": "llava",
98
+ "model_args": "pretrained=/cm/archive/namnv78_new/revise_checkpoints/Xphi35-siglip224/SMOE/665K36/revise_Full_smoe_sharev3,conv_template=phi35",
99
+ "batch_size": "1",
100
+ "device": null,
101
+ "limit": null,
102
+ "bootstrap_iters": 100000,
103
+ "gen_kwargs": ""
104
+ },
105
+ "git_hash": "289c7fe5"
106
+ }
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1743_llava_v1.5_ai2d_llava_model_args_5a2714/results.json ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ai2d": {
4
+ "exact_match,flexible-extract": 0.6541450777202072,
5
+ "exact_match_stderr,flexible-extract": 0.008560826682457891,
6
+ "alias": "ai2d"
7
+ }
8
+ },
9
+ "configs": {
10
+ "ai2d": {
11
+ "task": "ai2d",
12
+ "dataset_path": "lmms-lab/ai2d",
13
+ "dataset_kwargs": {
14
+ "token": true
15
+ },
16
+ "test_split": "test",
17
+ "doc_to_visual": "<function ai2d_doc_to_visual at 0x7f1b188ff940>",
18
+ "doc_to_text": "<function ai2d_doc_to_text at 0x7f1b188ffc10>",
19
+ "doc_to_target": "<function ai2d_doc_to_target at 0x7f1b189061f0>",
20
+ "description": "",
21
+ "target_delimiter": " ",
22
+ "fewshot_delimiter": "\n\n",
23
+ "metric_list": [
24
+ {
25
+ "metric": "exact_match",
26
+ "aggregation": "mean",
27
+ "higher_is_better": true,
28
+ "ignore_case": true,
29
+ "ignore_punctuation": true
30
+ }
31
+ ],
32
+ "output_type": "generate_until",
33
+ "generation_kwargs": {
34
+ "max_new_tokens": 512,
35
+ "temperature": 0.0,
36
+ "do_sample": false,
37
+ "until": [
38
+ "\n\n"
39
+ ]
40
+ },
41
+ "repeats": 1,
42
+ "filter_list": [
43
+ {
44
+ "name": "flexible-extract",
45
+ "filter": [
46
+ {
47
+ "function": "<class 'utils.MultiChoiceRegexFilter'>",
48
+ "group_select": 0,
49
+ "ignore_case": true,
50
+ "ignore_punctuation": true,
51
+ "regex_pattern": "([A-Z])\\."
52
+ }
53
+ ]
54
+ }
55
+ ],
56
+ "should_decontaminate": false,
57
+ "metadata": [
58
+ {
59
+ "version": 0.0
60
+ }
61
+ ],
62
+ "model_specific_prompt_kwargs": {
63
+ "default": {
64
+ "prompt_format": "mcq",
65
+ "pre_prompt": "",
66
+ "post_prompt": "\nAnswer with the option's letter from the given choices directly."
67
+ },
68
+ "gpt4v": {
69
+ "prompt_format": "mcq",
70
+ "pre_prompt": "",
71
+ "post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
72
+ },
73
+ "qwen_vl": {
74
+ "prompt_format": "qa",
75
+ "pre_prompt": "",
76
+ "post_prompt": " Answer:"
77
+ },
78
+ "xcomposer2_4khd": {
79
+ "prompt_format": "mcq_xcomposer",
80
+ "pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
81
+ "post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
82
+ }
83
+ },
84
+ "model_specific_target_kwargs": {
85
+ "default": "mcq",
86
+ "qwen_vl": "qa"
87
+ }
88
+ }
89
+ },
90
+ "versions": {
91
+ "ai2d": "Yaml"
92
+ },
93
+ "n-shot": {
94
+ "ai2d": 0
95
+ },
96
+ "model_configs": {
97
+ "model": "llava",
98
+ "model_args": "pretrained=/cm/archive/namnv78_new/revise_checkpoints/Xphi35-siglip224/SMOE/665K36/revise_Full_smoe_sharev3,conv_template=phi35",
99
+ "batch_size": "1",
100
+ "device": null,
101
+ "limit": null,
102
+ "bootstrap_iters": 100000,
103
+ "gen_kwargs": ""
104
+ },
105
+ "git_hash": "289c7fe5"
106
+ }
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1747_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json ADDED
The diff for this file is too large to render. See raw diff