pltops commited on
Commit
18ec871
·
verified ·
1 Parent(s): 0c3a5d3

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +10 -0
  2. llava_base/.ipynb_checkpoints/llava_base_MME_score-checkpoint.csv +2 -0
  3. llava_base/.ipynb_checkpoints/llava_base_POPE_score-checkpoint.csv +5 -0
  4. llava_base/.ipynb_checkpoints/llava_base_SEEDBench_IMG_acc-checkpoint.csv +2 -0
  5. llava_base/.ipynb_checkpoints/llava_base_ScienceQA_VAL_acc-checkpoint.csv +2 -0
  6. llava_base/.ipynb_checkpoints/llava_base_TextVQA_VAL_acc-checkpoint.csv +2 -0
  7. llava_base/llava_base_AI2D_TEST.xlsx +0 -0
  8. llava_base/llava_base_AI2D_TEST_acc.csv +2 -0
  9. llava_base/llava_base_AI2D_TEST_openai_result.pkl +3 -0
  10. llava_base/llava_base_AI2D_TEST_openai_result.xlsx +0 -0
  11. llava_base/llava_base_HallusionBench.xlsx +0 -0
  12. llava_base/llava_base_HallusionBench_PREV.pkl +3 -0
  13. llava_base/llava_base_HallusionBench_auxmatch.xlsx +0 -0
  14. llava_base/llava_base_HallusionBench_score.csv +13 -0
  15. llava_base/llava_base_MME.xlsx +0 -0
  16. llava_base/llava_base_MME_auxmatch.xlsx +0 -0
  17. llava_base/llava_base_POPE.xlsx +0 -0
  18. llava_base/llava_base_POPE_PREV.pkl +3 -0
  19. llava_base/llava_base_POPE_auxmatch.xlsx +0 -0
  20. llava_base/llava_base_POPE_score.csv +1 -1
  21. llava_base/llava_base_Q-Bench1_VAL.xlsx +0 -0
  22. llava_base/llava_base_Q-Bench1_VAL_acc.csv +2 -0
  23. llava_base/llava_base_Q-Bench1_VAL_openai_result.pkl +3 -0
  24. llava_base/llava_base_Q-Bench1_VAL_openai_result.xlsx +0 -0
  25. llava_base/llava_base_RealWorldQA.xlsx +0 -0
  26. llava_base/llava_base_RealWorldQA_PREV.pkl +3 -0
  27. llava_base/llava_base_RealWorldQA_acc.csv +2 -0
  28. llava_base/llava_base_RealWorldQA_openai_result.pkl +3 -0
  29. llava_base/llava_base_RealWorldQA_openai_result.xlsx +0 -0
  30. llava_base/llava_base_SEEDBench_IMG.xlsx +0 -0
  31. llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx +2 -2
  32. llava_base/llava_base_ScienceQA_VAL.xlsx +0 -0
  33. llava_base/llava_base_ScienceQA_VAL_PREV.pkl +3 -0
  34. llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx +0 -0
  35. llava_base/llava_base_TextVQA_VAL.xlsx +0 -0
  36. llava_base/llava_base_TextVQA_VAL_PREV.pkl +3 -0
  37. llava_base_v2/.ipynb_checkpoints/llava_base_v2_HallusionBench_score-checkpoint.csv +13 -0
  38. llava_base_v2/.ipynb_checkpoints/llava_base_v2_MME_score-checkpoint.csv +2 -0
  39. llava_base_v2/.ipynb_checkpoints/llava_base_v2_ScienceQA_VAL_acc-checkpoint.csv +2 -0
  40. llava_base_v2/llava_base_v2_AI2D_TEST.xlsx +0 -0
  41. llava_base_v2/llava_base_v2_AI2D_TEST_acc.csv +2 -0
  42. llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.pkl +3 -0
  43. llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.xlsx +0 -0
  44. llava_base_v2/llava_base_v2_HallusionBench.xlsx +0 -0
  45. llava_base_v2/llava_base_v2_HallusionBench_PREV.pkl +3 -0
  46. llava_base_v2/llava_base_v2_HallusionBench_auxmatch.xlsx +0 -0
  47. llava_base_v2/llava_base_v2_HallusionBench_score.csv +13 -0
  48. llava_base_v2/llava_base_v2_MME.xlsx +0 -0
  49. llava_base_v2/llava_base_v2_MME_PREV.pkl +3 -0
  50. llava_base_v2/llava_base_v2_MME_auxmatch.xlsx +0 -0
.gitattributes CHANGED
@@ -37,3 +37,13 @@ llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge
37
  llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
38
  llava_moe_e4t2/llava_moe_e4t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
39
  llava_moe_e5t3/llava_moe_e5t3_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
37
  llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
38
  llava_moe_e4t2/llava_moe_e4t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
39
  llava_moe_e5t3/llava_moe_e5t3_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
40
+ llava_moe_e8t2/llava_moe_e8t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
41
+ llava_phi_c01/llava_phi_c01_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
42
+ llava_phi_c02/llava_phi_c02_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
43
+ llava_phi_c03/llava_phi_c03_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
44
+ llava_phi_c05/llava_phi_c05_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
45
+ llava_phi_c06/llava_phi_c06_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
46
+ llava_phi_c09/llava_phi_c09_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
47
+ llava_phi_c10/llava_phi_c10_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
48
+ vicuna7b_moe_llava/vicuna7b_moe_llava_SEEDBench_IMG.xlsx filter=lfs diff=lfs merge=lfs -text
49
+ vicuna7b_moe_llava/vicuna7b_moe_llava_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
llava_base/.ipynb_checkpoints/llava_base_MME_score-checkpoint.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
2
+ "1259.7040816326532","214.28571428571428","130.0","108.0","105.0","22.5","173.33333333333334","104.28571428571428","125.0","143.33333333333331","75.0","12.5","133.33333333333334","110.20408163265304","156.5","75.0"
llava_base/.ipynb_checkpoints/llava_base_POPE_score-checkpoint.csv ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ "split","Overall","acc","precision","recall"
2
+ "Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
3
+ "popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
4
+ "random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
5
+ "adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
llava_base/.ipynb_checkpoints/llava_base_SEEDBench_IMG_acc-checkpoint.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Instance Attributes","Instance Identity","Instance Interaction","Instance Location","Instances Counting","Scene Understanding","Spatial Relation","Text Understanding","Visual Reasoning"
2
+ "none","0.5265598650927488","0.5091417509141751","0.590387766247952","0.6597938144329897","0.5040899795501023","0.27257866775643647","0.7178594046865104","0.4337899543378995","0.2857142857142857","0.743202416918429"
llava_base/.ipynb_checkpoints/llava_base_ScienceQA_VAL_acc-checkpoint.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Adaptations","Adaptations and natural selection","Age of Exploration","Ancient Egypt and Kush","Ancient Mesopotamia","Animals","Astronomy","Atoms and molecules","Basic economic principles","Chemical reactions","Cities","Classification","Classification and scientific names","Climate change","Colonial America","Context clues","Descriptive details","Designing experiments","Domain-specific vocabulary","Early 19th century American history","Early Americas","Earth events","Ecological interactions","Ecosystems","Engineering practices","English colonies in North America","Force and motion","Fossils","Genes to traits","Geography","Government","Independent reading comprehension","Informational texts: level 1","Magnets","Maps","Materials","Medieval Asia","Natural resources and human impacts","Oceania: geography","Oceans and continents","Oceans and continents ","Particle motion and energy","Persuasive strategies","Physical Geography","Plant reproduction","Plants","Plate tectonics","Read-alone texts","Rocks and minerals","Rome and the Byzantine Empire","Scientific names","Solutions","State capitals","States","States of matter","The American Revolution","The Americas: geography","The Antebellum period","The Civil War and Reconstruction","The Silk Road","Thermal energy","Velocity, acceleration, and forces","Visual elements","Water cycle","Weather and climate","World religions"
2
+ "val","0.5970433953266572","0.625","1.0","1.0","1.0","1.0","1.0","0.5869565217391305","0.6190476190476191","0.53125","1.0","0.125","0.5833333333333334","0.8723404255319149","1.0","0.21428571428571427","0.8","1.0","0.9032258064516129","0.6","0.0","0.5","1.0","0.52","0.7611940298507462","0.9811320754716981","0.1794871794871795","0.8","0.29411764705882354","0.1276595744680851","0.4225352112676056","1.0","1.0","0.8","0.23308270676691728","0.3387096774193548","0.646551724137931","0.5","0.5","0.19148936170212766","0.9","0.5","0.5147058823529411","0.5555555555555556","0.6046511627906976","0.6","0.8","0.75","1.0","0.926829268292683","0.5","0.8571428571428571","0.373134328358209","0.949685534591195","0.3333333333333333","0.8947368421052632","0.6666666666666666","0.13333333333333333","1.0","0.0","1.0","0.4444444444444444","0.19607843137254902","0.5555555555555556","1.0","0.34375","0.6666666666666666"
llava_base/.ipynb_checkpoints/llava_base_TextVQA_VAL_acc-checkpoint.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "Overall"
2
+ "15.498000000000001"
llava_base/llava_base_AI2D_TEST.xlsx ADDED
Binary file (244 kB). View file
 
llava_base/llava_base_AI2D_TEST_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","atomStructure","eclipses","faultsEarthquakes","foodChainsWebs","lifeCycles","moonPhaseEquinox","partsOfA","partsOfTheEarth","photosynthesisRespiration","rockCycle","rockStrata","solarSystem","typesOf","volcano","waterCNPCycle"
2
+ "none","0.4413860103626943","0.375","0.5806451612903226","0.5357142857142857","0.4750430292598967","0.37170263788968827","0.3249097472924188","0.5133470225872689","0.4807692307692308","0.379746835443038","0.34328358208955223","0.5121951219512195","0.6111111111111112","0.3848396501457726","0.5625","0.4090909090909091"
llava_base/llava_base_AI2D_TEST_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43a91a41f4cd9c52ccdd9bc8e420b5d453663fc641f4e8c42089701d5608dbd5
3
+ size 241186
llava_base/llava_base_AI2D_TEST_openai_result.xlsx ADDED
Binary file (264 kB). View file
 
llava_base/llava_base_HallusionBench.xlsx ADDED
Binary file (80.3 kB). View file
 
llava_base/llava_base_HallusionBench_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c368882dba8fcf9cd24d9d56b278a37bcee7b46e25e3419009d638c9f4bb119a
3
+ size 127440
llava_base/llava_base_HallusionBench_auxmatch.xlsx ADDED
Binary file (85.9 kB). View file
 
llava_base/llava_base_HallusionBench_score.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ "split","aAcc","fAcc","qAcc"
2
+ "Overall","33.64879074658255","15.31791907514451","12.087912087912088"
3
+ "VD","38.40947546531303","21.304347826086957","15.884476534296029"
4
+ "VS","25.833333333333336","3.4482758620689653","6.179775280898876"
5
+ "VD_video","15.294117647058824","2.083333333333333","1.4492753623188406"
6
+ "VS_chart","22.30769230769231","0.0","9.210526315789473"
7
+ "VD_illusion","48.61111111111111","22.58064516129032","11.11111111111111"
8
+ "VD_figure","50.0","29.268292682926827","23.076923076923077"
9
+ "VS_table","23.214285714285715","0.0","0.0"
10
+ "VD_ocr","59.55056179775281","46.51162790697674","46.51162790697674"
11
+ "VS_map","29.6875","0.0","3.125"
12
+ "VS_ocr","35.18518518518518","15.384615384615385","11.11111111111111"
13
+ "VD_math","35.18518518518518","5.555555555555555","11.11111111111111"
llava_base/llava_base_MME.xlsx CHANGED
Binary files a/llava_base/llava_base_MME.xlsx and b/llava_base/llava_base_MME.xlsx differ
 
llava_base/llava_base_MME_auxmatch.xlsx CHANGED
Binary files a/llava_base/llava_base_MME_auxmatch.xlsx and b/llava_base/llava_base_MME_auxmatch.xlsx differ
 
llava_base/llava_base_POPE.xlsx CHANGED
Binary files a/llava_base/llava_base_POPE.xlsx and b/llava_base/llava_base_POPE.xlsx differ
 
llava_base/llava_base_POPE_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d2867aacdff31a0e6b1459610a9123dc7e91a3de36c37d1be73e8aff3a089fe
3
+ size 42809
llava_base/llava_base_POPE_auxmatch.xlsx CHANGED
Binary files a/llava_base/llava_base_POPE_auxmatch.xlsx and b/llava_base/llava_base_POPE_auxmatch.xlsx differ
 
llava_base/llava_base_POPE_score.csv CHANGED
@@ -1,5 +1,5 @@
1
  "split","Overall","acc","precision","recall"
2
  "Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
3
- "adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
4
  "popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
 
5
  "random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
 
1
  "split","Overall","acc","precision","recall"
2
  "Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
 
3
  "popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
4
+ "adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
5
  "random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
llava_base/llava_base_Q-Bench1_VAL.xlsx ADDED
Binary file (88.7 kB). View file
 
llava_base/llava_base_Q-Bench1_VAL_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","type_0_concern_0","type_0_concern_1","type_0_concern_2","type_0_concern_3","type_1_concern_0","type_1_concern_1","type_1_concern_2","type_1_concern_3","type_2_concern_0","type_2_concern_1","type_2_concern_2","type_2_concern_3"
2
+ "val","0.5351170568561873","0.526595744680851","0.5485714285714286","0.5128205128205128","0.6571428571428571","0.3","0.6696428571428571","0.53","0.6","0.5113636363636364","0.5793103448275863","0.4942528735632184","0.6470588235294118"
llava_base/llava_base_Q-Bench1_VAL_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28991b05fdf580881649ed2e8e681aca2cc72490b747178deefa9b7d37952698
3
+ size 93585
llava_base/llava_base_Q-Bench1_VAL_openai_result.xlsx ADDED
Binary file (97.8 kB). View file
 
llava_base/llava_base_RealWorldQA.xlsx ADDED
Binary file (46.5 kB). View file
 
llava_base/llava_base_RealWorldQA_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:053811597f0f10d8022535ce9ec32b6e3ab8daeb44f8694c5b7932e2083e60e3
3
+ size 3826
llava_base/llava_base_RealWorldQA_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall"
2
+ "none","0.39738562091503266"
llava_base/llava_base_RealWorldQA_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a857ca6130723ceff478d8434ebfe6b22ef785fd027403cc8276558652d9c22
3
+ size 59088
llava_base/llava_base_RealWorldQA_openai_result.xlsx ADDED
Binary file (52.5 kB). View file
 
llava_base/llava_base_SEEDBench_IMG.xlsx CHANGED
Binary files a/llava_base/llava_base_SEEDBench_IMG.xlsx and b/llava_base/llava_base_SEEDBench_IMG.xlsx differ
 
llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b848ae407b25ea999e8e77dac01c39f68d147660cfdc70f0a9bf556535bb133
3
- size 1061587
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bcdd6c6c6e008a04008cc8af05b6fa5ace113a2e291a0a06903c7ea3032761a
3
+ size 1061588
llava_base/llava_base_ScienceQA_VAL.xlsx CHANGED
Binary files a/llava_base/llava_base_ScienceQA_VAL.xlsx and b/llava_base/llava_base_ScienceQA_VAL.xlsx differ
 
llava_base/llava_base_ScienceQA_VAL_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:362aa0fb5597a075efb0989640491d8dabc7134d2eb21dc7f56783bcf976e4e2
3
+ size 10875
llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx CHANGED
Binary files a/llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx and b/llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx differ
 
llava_base/llava_base_TextVQA_VAL.xlsx CHANGED
Binary files a/llava_base/llava_base_TextVQA_VAL.xlsx and b/llava_base/llava_base_TextVQA_VAL.xlsx differ
 
llava_base/llava_base_TextVQA_VAL_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9596b33c0105a0daac9fd01eb1470408225ad2c140362535ae27051b5910cd15
3
+ size 47012
llava_base_v2/.ipynb_checkpoints/llava_base_v2_HallusionBench_score-checkpoint.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ "split","aAcc","fAcc","qAcc"
2
+ "Overall","44.05888538380652","18.20809248554913","15.604395604395604"
3
+ "VD","48.73096446700508","23.47826086956522","19.855595667870034"
4
+ "VS","36.388888888888886","7.758620689655173","8.98876404494382"
5
+ "VS_map","43.75","13.636363636363635","6.25"
6
+ "VD_illusion","54.861111111111114","20.967741935483872","15.277777777777779"
7
+ "VD_ocr","57.30337078651685","39.53488372093023","30.23255813953488"
8
+ "VD_math","52.77777777777778","19.444444444444446","27.77777777777778"
9
+ "VD_video","30.0","2.083333333333333","5.797101449275362"
10
+ "VD_figure","62.5","39.02439024390244","30.76923076923077"
11
+ "VS_chart","29.230769230769234","2.5","10.526315789473683"
12
+ "VS_ocr","40.74074074074074","15.384615384615385","11.11111111111111"
13
+ "VS_table","38.392857142857146","3.571428571428571","6.976744186046512"
llava_base_v2/.ipynb_checkpoints/llava_base_v2_MME_score-checkpoint.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
2
+ "1268.2366946778711","265.0","95.0","110.5","69.11764705882354","57.5","185.0","115.0","153.33333333333334","190.0","56.0","47.5","128.33333333333334","130.95238095238096","150.0","45.0"
llava_base_v2/.ipynb_checkpoints/llava_base_v2_ScienceQA_VAL_acc-checkpoint.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Adaptations","Adaptations and natural selection","Age of Exploration","Ancient Egypt and Kush","Ancient Mesopotamia","Animals","Astronomy","Atoms and molecules","Basic economic principles","Chemical reactions","Cities","Classification","Classification and scientific names","Climate change","Colonial America","Context clues","Descriptive details","Designing experiments","Domain-specific vocabulary","Early 19th century American history","Early Americas","Earth events","Ecological interactions","Ecosystems","Engineering practices","English colonies in North America","Force and motion","Fossils","Genes to traits","Geography","Government","Independent reading comprehension","Informational texts: level 1","Magnets","Maps","Materials","Medieval Asia","Natural resources and human impacts","Oceania: geography","Oceans and continents","Oceans and continents ","Particle motion and energy","Persuasive strategies","Physical Geography","Plant reproduction","Plants","Plate tectonics","Read-alone texts","Rocks and minerals","Rome and the Byzantine Empire","Scientific names","Solutions","State capitals","States","States of matter","The American Revolution","The Americas: geography","The Antebellum period","The Civil War and Reconstruction","The Silk Road","Thermal energy","Velocity, acceleration, and forces","Visual elements","Water cycle","Weather and climate","World religions"
2
+ "val","0.6547448736289938","0.84375","1.0","1.0","1.0","0.5","1.0","0.5","0.6904761904761905","0.59375","1.0","0.375","0.7619047619047619","0.9361702127659575","1.0","0.5238095238095238","0.8","1.0","1.0","0.6","0.0","0.5","1.0","0.6","0.7164179104477612","1.0","0.4358974358974359","0.68","0.47058823529411764","0.0851063829787234","0.4647887323943662","1.0","1.0","1.0","0.2781954887218045","0.5483870967741935","0.8017241379310345","0.5","0.5","0.2765957446808511","0.8","1.0","0.25","0.5555555555555556","0.5348837209302325","0.7","0.8","1.0","1.0","1.0","0.5","0.8571428571428571","0.417910447761194","0.9905660377358491","0.3333333333333333","0.7894736842105263","0.6666666666666666","0.3","1.0","0.0","1.0","0.6666666666666666","0.13725490196078433","0.7777777777777778","1.0","0.359375","1.0"
llava_base_v2/llava_base_v2_AI2D_TEST.xlsx ADDED
Binary file (243 kB). View file
 
llava_base_v2/llava_base_v2_AI2D_TEST_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","atomStructure","eclipses","faultsEarthquakes","foodChainsWebs","lifeCycles","moonPhaseEquinox","partsOfA","partsOfTheEarth","photosynthesisRespiration","rockCycle","rockStrata","solarSystem","typesOf","volcano","waterCNPCycle"
2
+ "none","0.5761010362694301","0.75","0.6451612903225806","0.5714285714285714","0.6600688468158348","0.4892086330935252","0.4296028880866426","0.5913757700205339","0.4807692307692308","0.4810126582278481","0.4626865671641791","0.5365853658536586","0.7777777777777778","0.5335276967930029","0.625","0.5"
llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19f30bb2c7308b8dc6c9aa69174a7179fd4a487dd041325d78f5c3c2752b1826
3
+ size 171966
llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.xlsx ADDED
Binary file (262 kB). View file
 
llava_base_v2/llava_base_v2_HallusionBench.xlsx ADDED
Binary file (78.5 kB). View file
 
llava_base_v2/llava_base_v2_HallusionBench_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3b93ac3fc083461d16975077f8f62141c24cc879641a88978408c9ce9160ed8
3
+ size 118397
llava_base_v2/llava_base_v2_HallusionBench_auxmatch.xlsx ADDED
Binary file (84 kB). View file
 
llava_base_v2/llava_base_v2_HallusionBench_score.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ "split","aAcc","fAcc","qAcc"
2
+ "Overall","44.05888538380652","18.20809248554913","15.604395604395604"
3
+ "VS","36.388888888888886","7.758620689655173","8.98876404494382"
4
+ "VD","48.73096446700508","23.47826086956522","19.855595667870034"
5
+ "VS_table","38.392857142857146","3.571428571428571","6.976744186046512"
6
+ "VD_math","52.77777777777778","19.444444444444446","27.77777777777778"
7
+ "VD_illusion","54.861111111111114","20.967741935483872","15.277777777777779"
8
+ "VS_chart","29.230769230769234","2.5","10.526315789473683"
9
+ "VS_ocr","40.74074074074074","15.384615384615385","11.11111111111111"
10
+ "VD_ocr","57.30337078651685","39.53488372093023","30.23255813953488"
11
+ "VD_video","30.0","2.083333333333333","5.797101449275362"
12
+ "VD_figure","62.5","39.02439024390244","30.76923076923077"
13
+ "VS_map","43.75","13.636363636363635","6.25"
llava_base_v2/llava_base_v2_MME.xlsx CHANGED
Binary files a/llava_base_v2/llava_base_v2_MME.xlsx and b/llava_base_v2/llava_base_v2_MME.xlsx differ
 
llava_base_v2/llava_base_v2_MME_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15c7dc51edd95ab4497f6d1380c721c10f51a74018472654929654631bce42c7
3
+ size 12783
llava_base_v2/llava_base_v2_MME_auxmatch.xlsx CHANGED
Binary files a/llava_base_v2/llava_base_v2_MME_auxmatch.xlsx and b/llava_base_v2/llava_base_v2_MME_auxmatch.xlsx differ