Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +10 -0
- llava_base/.ipynb_checkpoints/llava_base_MME_score-checkpoint.csv +2 -0
- llava_base/.ipynb_checkpoints/llava_base_POPE_score-checkpoint.csv +5 -0
- llava_base/.ipynb_checkpoints/llava_base_SEEDBench_IMG_acc-checkpoint.csv +2 -0
- llava_base/.ipynb_checkpoints/llava_base_ScienceQA_VAL_acc-checkpoint.csv +2 -0
- llava_base/.ipynb_checkpoints/llava_base_TextVQA_VAL_acc-checkpoint.csv +2 -0
- llava_base/llava_base_AI2D_TEST.xlsx +0 -0
- llava_base/llava_base_AI2D_TEST_acc.csv +2 -0
- llava_base/llava_base_AI2D_TEST_openai_result.pkl +3 -0
- llava_base/llava_base_AI2D_TEST_openai_result.xlsx +0 -0
- llava_base/llava_base_HallusionBench.xlsx +0 -0
- llava_base/llava_base_HallusionBench_PREV.pkl +3 -0
- llava_base/llava_base_HallusionBench_auxmatch.xlsx +0 -0
- llava_base/llava_base_HallusionBench_score.csv +13 -0
- llava_base/llava_base_MME.xlsx +0 -0
- llava_base/llava_base_MME_auxmatch.xlsx +0 -0
- llava_base/llava_base_POPE.xlsx +0 -0
- llava_base/llava_base_POPE_PREV.pkl +3 -0
- llava_base/llava_base_POPE_auxmatch.xlsx +0 -0
- llava_base/llava_base_POPE_score.csv +1 -1
- llava_base/llava_base_Q-Bench1_VAL.xlsx +0 -0
- llava_base/llava_base_Q-Bench1_VAL_acc.csv +2 -0
- llava_base/llava_base_Q-Bench1_VAL_openai_result.pkl +3 -0
- llava_base/llava_base_Q-Bench1_VAL_openai_result.xlsx +0 -0
- llava_base/llava_base_RealWorldQA.xlsx +0 -0
- llava_base/llava_base_RealWorldQA_PREV.pkl +3 -0
- llava_base/llava_base_RealWorldQA_acc.csv +2 -0
- llava_base/llava_base_RealWorldQA_openai_result.pkl +3 -0
- llava_base/llava_base_RealWorldQA_openai_result.xlsx +0 -0
- llava_base/llava_base_SEEDBench_IMG.xlsx +0 -0
- llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx +2 -2
- llava_base/llava_base_ScienceQA_VAL.xlsx +0 -0
- llava_base/llava_base_ScienceQA_VAL_PREV.pkl +3 -0
- llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx +0 -0
- llava_base/llava_base_TextVQA_VAL.xlsx +0 -0
- llava_base/llava_base_TextVQA_VAL_PREV.pkl +3 -0
- llava_base_v2/.ipynb_checkpoints/llava_base_v2_HallusionBench_score-checkpoint.csv +13 -0
- llava_base_v2/.ipynb_checkpoints/llava_base_v2_MME_score-checkpoint.csv +2 -0
- llava_base_v2/.ipynb_checkpoints/llava_base_v2_ScienceQA_VAL_acc-checkpoint.csv +2 -0
- llava_base_v2/llava_base_v2_AI2D_TEST.xlsx +0 -0
- llava_base_v2/llava_base_v2_AI2D_TEST_acc.csv +2 -0
- llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.pkl +3 -0
- llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.xlsx +0 -0
- llava_base_v2/llava_base_v2_HallusionBench.xlsx +0 -0
- llava_base_v2/llava_base_v2_HallusionBench_PREV.pkl +3 -0
- llava_base_v2/llava_base_v2_HallusionBench_auxmatch.xlsx +0 -0
- llava_base_v2/llava_base_v2_HallusionBench_score.csv +13 -0
- llava_base_v2/llava_base_v2_MME.xlsx +0 -0
- llava_base_v2/llava_base_v2_MME_PREV.pkl +3 -0
- llava_base_v2/llava_base_v2_MME_auxmatch.xlsx +0 -0
.gitattributes
CHANGED
|
@@ -37,3 +37,13 @@ llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge
|
|
| 37 |
llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 38 |
llava_moe_e4t2/llava_moe_e4t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 39 |
llava_moe_e5t3/llava_moe_e5t3_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 38 |
llava_moe_e4t2/llava_moe_e4t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 39 |
llava_moe_e5t3/llava_moe_e5t3_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
llava_moe_e8t2/llava_moe_e8t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
llava_phi_c01/llava_phi_c01_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
llava_phi_c02/llava_phi_c02_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
llava_phi_c03/llava_phi_c03_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
llava_phi_c05/llava_phi_c05_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
llava_phi_c06/llava_phi_c06_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
llava_phi_c09/llava_phi_c09_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
llava_phi_c10/llava_phi_c10_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
vicuna7b_moe_llava/vicuna7b_moe_llava_SEEDBench_IMG.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
vicuna7b_moe_llava/vicuna7b_moe_llava_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
|
llava_base/.ipynb_checkpoints/llava_base_MME_score-checkpoint.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
|
| 2 |
+
"1259.7040816326532","214.28571428571428","130.0","108.0","105.0","22.5","173.33333333333334","104.28571428571428","125.0","143.33333333333331","75.0","12.5","133.33333333333334","110.20408163265304","156.5","75.0"
|
llava_base/.ipynb_checkpoints/llava_base_POPE_score-checkpoint.csv
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","acc","precision","recall"
|
| 2 |
+
"Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
|
| 3 |
+
"popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
|
| 4 |
+
"random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
|
| 5 |
+
"adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
|
llava_base/.ipynb_checkpoints/llava_base_SEEDBench_IMG_acc-checkpoint.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","Instance Attributes","Instance Identity","Instance Interaction","Instance Location","Instances Counting","Scene Understanding","Spatial Relation","Text Understanding","Visual Reasoning"
|
| 2 |
+
"none","0.5265598650927488","0.5091417509141751","0.590387766247952","0.6597938144329897","0.5040899795501023","0.27257866775643647","0.7178594046865104","0.4337899543378995","0.2857142857142857","0.743202416918429"
|
llava_base/.ipynb_checkpoints/llava_base_ScienceQA_VAL_acc-checkpoint.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","Adaptations","Adaptations and natural selection","Age of Exploration","Ancient Egypt and Kush","Ancient Mesopotamia","Animals","Astronomy","Atoms and molecules","Basic economic principles","Chemical reactions","Cities","Classification","Classification and scientific names","Climate change","Colonial America","Context clues","Descriptive details","Designing experiments","Domain-specific vocabulary","Early 19th century American history","Early Americas","Earth events","Ecological interactions","Ecosystems","Engineering practices","English colonies in North America","Force and motion","Fossils","Genes to traits","Geography","Government","Independent reading comprehension","Informational texts: level 1","Magnets","Maps","Materials","Medieval Asia","Natural resources and human impacts","Oceania: geography","Oceans and continents","Oceans and continents ","Particle motion and energy","Persuasive strategies","Physical Geography","Plant reproduction","Plants","Plate tectonics","Read-alone texts","Rocks and minerals","Rome and the Byzantine Empire","Scientific names","Solutions","State capitals","States","States of matter","The American Revolution","The Americas: geography","The Antebellum period","The Civil War and Reconstruction","The Silk Road","Thermal energy","Velocity, acceleration, and forces","Visual elements","Water cycle","Weather and climate","World religions"
|
| 2 |
+
"val","0.5970433953266572","0.625","1.0","1.0","1.0","1.0","1.0","0.5869565217391305","0.6190476190476191","0.53125","1.0","0.125","0.5833333333333334","0.8723404255319149","1.0","0.21428571428571427","0.8","1.0","0.9032258064516129","0.6","0.0","0.5","1.0","0.52","0.7611940298507462","0.9811320754716981","0.1794871794871795","0.8","0.29411764705882354","0.1276595744680851","0.4225352112676056","1.0","1.0","0.8","0.23308270676691728","0.3387096774193548","0.646551724137931","0.5","0.5","0.19148936170212766","0.9","0.5","0.5147058823529411","0.5555555555555556","0.6046511627906976","0.6","0.8","0.75","1.0","0.926829268292683","0.5","0.8571428571428571","0.373134328358209","0.949685534591195","0.3333333333333333","0.8947368421052632","0.6666666666666666","0.13333333333333333","1.0","0.0","1.0","0.4444444444444444","0.19607843137254902","0.5555555555555556","1.0","0.34375","0.6666666666666666"
|
llava_base/.ipynb_checkpoints/llava_base_TextVQA_VAL_acc-checkpoint.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"Overall"
|
| 2 |
+
"15.498000000000001"
|
llava_base/llava_base_AI2D_TEST.xlsx
ADDED
|
Binary file (244 kB). View file
|
|
|
llava_base/llava_base_AI2D_TEST_acc.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","atomStructure","eclipses","faultsEarthquakes","foodChainsWebs","lifeCycles","moonPhaseEquinox","partsOfA","partsOfTheEarth","photosynthesisRespiration","rockCycle","rockStrata","solarSystem","typesOf","volcano","waterCNPCycle"
|
| 2 |
+
"none","0.4413860103626943","0.375","0.5806451612903226","0.5357142857142857","0.4750430292598967","0.37170263788968827","0.3249097472924188","0.5133470225872689","0.4807692307692308","0.379746835443038","0.34328358208955223","0.5121951219512195","0.6111111111111112","0.3848396501457726","0.5625","0.4090909090909091"
|
llava_base/llava_base_AI2D_TEST_openai_result.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:43a91a41f4cd9c52ccdd9bc8e420b5d453663fc641f4e8c42089701d5608dbd5
|
| 3 |
+
size 241186
|
llava_base/llava_base_AI2D_TEST_openai_result.xlsx
ADDED
|
Binary file (264 kB). View file
|
|
|
llava_base/llava_base_HallusionBench.xlsx
ADDED
|
Binary file (80.3 kB). View file
|
|
|
llava_base/llava_base_HallusionBench_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c368882dba8fcf9cd24d9d56b278a37bcee7b46e25e3419009d638c9f4bb119a
|
| 3 |
+
size 127440
|
llava_base/llava_base_HallusionBench_auxmatch.xlsx
ADDED
|
Binary file (85.9 kB). View file
|
|
|
llava_base/llava_base_HallusionBench_score.csv
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","aAcc","fAcc","qAcc"
|
| 2 |
+
"Overall","33.64879074658255","15.31791907514451","12.087912087912088"
|
| 3 |
+
"VD","38.40947546531303","21.304347826086957","15.884476534296029"
|
| 4 |
+
"VS","25.833333333333336","3.4482758620689653","6.179775280898876"
|
| 5 |
+
"VD_video","15.294117647058824","2.083333333333333","1.4492753623188406"
|
| 6 |
+
"VS_chart","22.30769230769231","0.0","9.210526315789473"
|
| 7 |
+
"VD_illusion","48.61111111111111","22.58064516129032","11.11111111111111"
|
| 8 |
+
"VD_figure","50.0","29.268292682926827","23.076923076923077"
|
| 9 |
+
"VS_table","23.214285714285715","0.0","0.0"
|
| 10 |
+
"VD_ocr","59.55056179775281","46.51162790697674","46.51162790697674"
|
| 11 |
+
"VS_map","29.6875","0.0","3.125"
|
| 12 |
+
"VS_ocr","35.18518518518518","15.384615384615385","11.11111111111111"
|
| 13 |
+
"VD_math","35.18518518518518","5.555555555555555","11.11111111111111"
|
llava_base/llava_base_MME.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_MME.xlsx and b/llava_base/llava_base_MME.xlsx differ
|
|
|
llava_base/llava_base_MME_auxmatch.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_MME_auxmatch.xlsx and b/llava_base/llava_base_MME_auxmatch.xlsx differ
|
|
|
llava_base/llava_base_POPE.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_POPE.xlsx and b/llava_base/llava_base_POPE.xlsx differ
|
|
|
llava_base/llava_base_POPE_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7d2867aacdff31a0e6b1459610a9123dc7e91a3de36c37d1be73e8aff3a089fe
|
| 3 |
+
size 42809
|
llava_base/llava_base_POPE_auxmatch.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_POPE_auxmatch.xlsx and b/llava_base/llava_base_POPE_auxmatch.xlsx differ
|
|
|
llava_base/llava_base_POPE_score.csv
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
"split","Overall","acc","precision","recall"
|
| 2 |
"Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
|
| 3 |
-
"adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
|
| 4 |
"popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
|
|
|
|
| 5 |
"random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
|
|
|
|
| 1 |
"split","Overall","acc","precision","recall"
|
| 2 |
"Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
|
|
|
|
| 3 |
"popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
|
| 4 |
+
"adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
|
| 5 |
"random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
|
llava_base/llava_base_Q-Bench1_VAL.xlsx
ADDED
|
Binary file (88.7 kB). View file
|
|
|
llava_base/llava_base_Q-Bench1_VAL_acc.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","type_0_concern_0","type_0_concern_1","type_0_concern_2","type_0_concern_3","type_1_concern_0","type_1_concern_1","type_1_concern_2","type_1_concern_3","type_2_concern_0","type_2_concern_1","type_2_concern_2","type_2_concern_3"
|
| 2 |
+
"val","0.5351170568561873","0.526595744680851","0.5485714285714286","0.5128205128205128","0.6571428571428571","0.3","0.6696428571428571","0.53","0.6","0.5113636363636364","0.5793103448275863","0.4942528735632184","0.6470588235294118"
|
llava_base/llava_base_Q-Bench1_VAL_openai_result.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28991b05fdf580881649ed2e8e681aca2cc72490b747178deefa9b7d37952698
|
| 3 |
+
size 93585
|
llava_base/llava_base_Q-Bench1_VAL_openai_result.xlsx
ADDED
|
Binary file (97.8 kB). View file
|
|
|
llava_base/llava_base_RealWorldQA.xlsx
ADDED
|
Binary file (46.5 kB). View file
|
|
|
llava_base/llava_base_RealWorldQA_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:053811597f0f10d8022535ce9ec32b6e3ab8daeb44f8694c5b7932e2083e60e3
|
| 3 |
+
size 3826
|
llava_base/llava_base_RealWorldQA_acc.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall"
|
| 2 |
+
"none","0.39738562091503266"
|
llava_base/llava_base_RealWorldQA_openai_result.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6a857ca6130723ceff478d8434ebfe6b22ef785fd027403cc8276558652d9c22
|
| 3 |
+
size 59088
|
llava_base/llava_base_RealWorldQA_openai_result.xlsx
ADDED
|
Binary file (52.5 kB). View file
|
|
|
llava_base/llava_base_SEEDBench_IMG.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_SEEDBench_IMG.xlsx and b/llava_base/llava_base_SEEDBench_IMG.xlsx differ
|
|
|
llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9bcdd6c6c6e008a04008cc8af05b6fa5ace113a2e291a0a06903c7ea3032761a
|
| 3 |
+
size 1061588
|
llava_base/llava_base_ScienceQA_VAL.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_ScienceQA_VAL.xlsx and b/llava_base/llava_base_ScienceQA_VAL.xlsx differ
|
|
|
llava_base/llava_base_ScienceQA_VAL_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:362aa0fb5597a075efb0989640491d8dabc7134d2eb21dc7f56783bcf976e4e2
|
| 3 |
+
size 10875
|
llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx and b/llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx differ
|
|
|
llava_base/llava_base_TextVQA_VAL.xlsx
CHANGED
|
Binary files a/llava_base/llava_base_TextVQA_VAL.xlsx and b/llava_base/llava_base_TextVQA_VAL.xlsx differ
|
|
|
llava_base/llava_base_TextVQA_VAL_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9596b33c0105a0daac9fd01eb1470408225ad2c140362535ae27051b5910cd15
|
| 3 |
+
size 47012
|
llava_base_v2/.ipynb_checkpoints/llava_base_v2_HallusionBench_score-checkpoint.csv
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","aAcc","fAcc","qAcc"
|
| 2 |
+
"Overall","44.05888538380652","18.20809248554913","15.604395604395604"
|
| 3 |
+
"VD","48.73096446700508","23.47826086956522","19.855595667870034"
|
| 4 |
+
"VS","36.388888888888886","7.758620689655173","8.98876404494382"
|
| 5 |
+
"VS_map","43.75","13.636363636363635","6.25"
|
| 6 |
+
"VD_illusion","54.861111111111114","20.967741935483872","15.277777777777779"
|
| 7 |
+
"VD_ocr","57.30337078651685","39.53488372093023","30.23255813953488"
|
| 8 |
+
"VD_math","52.77777777777778","19.444444444444446","27.77777777777778"
|
| 9 |
+
"VD_video","30.0","2.083333333333333","5.797101449275362"
|
| 10 |
+
"VD_figure","62.5","39.02439024390244","30.76923076923077"
|
| 11 |
+
"VS_chart","29.230769230769234","2.5","10.526315789473683"
|
| 12 |
+
"VS_ocr","40.74074074074074","15.384615384615385","11.11111111111111"
|
| 13 |
+
"VS_table","38.392857142857146","3.571428571428571","6.976744186046512"
|
llava_base_v2/.ipynb_checkpoints/llava_base_v2_MME_score-checkpoint.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
|
| 2 |
+
"1268.2366946778711","265.0","95.0","110.5","69.11764705882354","57.5","185.0","115.0","153.33333333333334","190.0","56.0","47.5","128.33333333333334","130.95238095238096","150.0","45.0"
|
llava_base_v2/.ipynb_checkpoints/llava_base_v2_ScienceQA_VAL_acc-checkpoint.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","Adaptations","Adaptations and natural selection","Age of Exploration","Ancient Egypt and Kush","Ancient Mesopotamia","Animals","Astronomy","Atoms and molecules","Basic economic principles","Chemical reactions","Cities","Classification","Classification and scientific names","Climate change","Colonial America","Context clues","Descriptive details","Designing experiments","Domain-specific vocabulary","Early 19th century American history","Early Americas","Earth events","Ecological interactions","Ecosystems","Engineering practices","English colonies in North America","Force and motion","Fossils","Genes to traits","Geography","Government","Independent reading comprehension","Informational texts: level 1","Magnets","Maps","Materials","Medieval Asia","Natural resources and human impacts","Oceania: geography","Oceans and continents","Oceans and continents ","Particle motion and energy","Persuasive strategies","Physical Geography","Plant reproduction","Plants","Plate tectonics","Read-alone texts","Rocks and minerals","Rome and the Byzantine Empire","Scientific names","Solutions","State capitals","States","States of matter","The American Revolution","The Americas: geography","The Antebellum period","The Civil War and Reconstruction","The Silk Road","Thermal energy","Velocity, acceleration, and forces","Visual elements","Water cycle","Weather and climate","World religions"
|
| 2 |
+
"val","0.6547448736289938","0.84375","1.0","1.0","1.0","0.5","1.0","0.5","0.6904761904761905","0.59375","1.0","0.375","0.7619047619047619","0.9361702127659575","1.0","0.5238095238095238","0.8","1.0","1.0","0.6","0.0","0.5","1.0","0.6","0.7164179104477612","1.0","0.4358974358974359","0.68","0.47058823529411764","0.0851063829787234","0.4647887323943662","1.0","1.0","1.0","0.2781954887218045","0.5483870967741935","0.8017241379310345","0.5","0.5","0.2765957446808511","0.8","1.0","0.25","0.5555555555555556","0.5348837209302325","0.7","0.8","1.0","1.0","1.0","0.5","0.8571428571428571","0.417910447761194","0.9905660377358491","0.3333333333333333","0.7894736842105263","0.6666666666666666","0.3","1.0","0.0","1.0","0.6666666666666666","0.13725490196078433","0.7777777777777778","1.0","0.359375","1.0"
|
llava_base_v2/llava_base_v2_AI2D_TEST.xlsx
ADDED
|
Binary file (243 kB). View file
|
|
|
llava_base_v2/llava_base_v2_AI2D_TEST_acc.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","Overall","atomStructure","eclipses","faultsEarthquakes","foodChainsWebs","lifeCycles","moonPhaseEquinox","partsOfA","partsOfTheEarth","photosynthesisRespiration","rockCycle","rockStrata","solarSystem","typesOf","volcano","waterCNPCycle"
|
| 2 |
+
"none","0.5761010362694301","0.75","0.6451612903225806","0.5714285714285714","0.6600688468158348","0.4892086330935252","0.4296028880866426","0.5913757700205339","0.4807692307692308","0.4810126582278481","0.4626865671641791","0.5365853658536586","0.7777777777777778","0.5335276967930029","0.625","0.5"
|
llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:19f30bb2c7308b8dc6c9aa69174a7179fd4a487dd041325d78f5c3c2752b1826
|
| 3 |
+
size 171966
|
llava_base_v2/llava_base_v2_AI2D_TEST_openai_result.xlsx
ADDED
|
Binary file (262 kB). View file
|
|
|
llava_base_v2/llava_base_v2_HallusionBench.xlsx
ADDED
|
Binary file (78.5 kB). View file
|
|
|
llava_base_v2/llava_base_v2_HallusionBench_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c3b93ac3fc083461d16975077f8f62141c24cc879641a88978408c9ce9160ed8
|
| 3 |
+
size 118397
|
llava_base_v2/llava_base_v2_HallusionBench_auxmatch.xlsx
ADDED
|
Binary file (84 kB). View file
|
|
|
llava_base_v2/llava_base_v2_HallusionBench_score.csv
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"split","aAcc","fAcc","qAcc"
|
| 2 |
+
"Overall","44.05888538380652","18.20809248554913","15.604395604395604"
|
| 3 |
+
"VS","36.388888888888886","7.758620689655173","8.98876404494382"
|
| 4 |
+
"VD","48.73096446700508","23.47826086956522","19.855595667870034"
|
| 5 |
+
"VS_table","38.392857142857146","3.571428571428571","6.976744186046512"
|
| 6 |
+
"VD_math","52.77777777777778","19.444444444444446","27.77777777777778"
|
| 7 |
+
"VD_illusion","54.861111111111114","20.967741935483872","15.277777777777779"
|
| 8 |
+
"VS_chart","29.230769230769234","2.5","10.526315789473683"
|
| 9 |
+
"VS_ocr","40.74074074074074","15.384615384615385","11.11111111111111"
|
| 10 |
+
"VD_ocr","57.30337078651685","39.53488372093023","30.23255813953488"
|
| 11 |
+
"VD_video","30.0","2.083333333333333","5.797101449275362"
|
| 12 |
+
"VD_figure","62.5","39.02439024390244","30.76923076923077"
|
| 13 |
+
"VS_map","43.75","13.636363636363635","6.25"
|
llava_base_v2/llava_base_v2_MME.xlsx
CHANGED
|
Binary files a/llava_base_v2/llava_base_v2_MME.xlsx and b/llava_base_v2/llava_base_v2_MME.xlsx differ
|
|
|
llava_base_v2/llava_base_v2_MME_PREV.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15c7dc51edd95ab4497f6d1380c721c10f51a74018472654929654631bce42c7
|
| 3 |
+
size 12783
|
llava_base_v2/llava_base_v2_MME_auxmatch.xlsx
CHANGED
|
Binary files a/llava_base_v2/llava_base_v2_MME_auxmatch.xlsx and b/llava_base_v2/llava_base_v2_MME_auxmatch.xlsx differ
|
|
|