pltops commited on
Commit
475e737
·
verified ·
1 Parent(s): 7e4a80c

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. llava_base/01_MMMU_DEV_VAL.pkl +3 -0
  3. llava_base/llava_base_MMBench_DEV_EN.xlsx +0 -0
  4. llava_base/llava_base_MMBench_DEV_EN_acc.csv +2 -0
  5. llava_base/llava_base_MMBench_DEV_EN_openai_result.pkl +3 -0
  6. llava_base/llava_base_MMBench_DEV_EN_openai_result.xlsx +0 -0
  7. llava_base/llava_base_MME.xlsx +0 -0
  8. llava_base/llava_base_MME_PREV.pkl +3 -0
  9. llava_base/llava_base_MME_auxmatch.xlsx +0 -0
  10. llava_base/llava_base_MME_score.csv +2 -0
  11. llava_base/llava_base_MMVet.xlsx +0 -0
  12. llava_base/llava_base_MathVista_MINI.xlsx +0 -0
  13. llava_base/llava_base_POPE.xlsx +0 -0
  14. llava_base/llava_base_POPE_auxmatch.xlsx +0 -0
  15. llava_base/llava_base_POPE_score.csv +5 -0
  16. llava_base/llava_base_SEEDBench_IMG.xlsx +0 -0
  17. llava_base/llava_base_SEEDBench_IMG_PREV.pkl +3 -0
  18. llava_base/llava_base_SEEDBench_IMG_acc.csv +2 -0
  19. llava_base/llava_base_SEEDBench_IMG_openai_result.pkl +3 -0
  20. llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx +3 -0
  21. llava_base/llava_base_ScienceQA_VAL.xlsx +0 -0
  22. llava_base/llava_base_ScienceQA_VAL_acc.csv +2 -0
  23. llava_base/llava_base_ScienceQA_VAL_openai_result.pkl +3 -0
  24. llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx +0 -0
  25. llava_base/llava_base_TextVQA_VAL.xlsx +0 -0
  26. llava_base/llava_base_TextVQA_VAL_acc.csv +2 -0
  27. llava_base_v2/01_MMMU_DEV_VAL.pkl +3 -0
  28. llava_base_v2/llava_base_v2_MME.xlsx +0 -0
  29. llava_base_v2/llava_base_v2_MME_auxmatch.xlsx +0 -0
  30. llava_base_v2/llava_base_v2_MME_score.csv +2 -0
  31. llava_base_v2/llava_base_v2_MMVet.xlsx +0 -0
  32. llava_base_v2/llava_base_v2_MathVista_MINI.xlsx +0 -0
  33. llava_base_v2/llava_base_v2_POPE.xlsx +0 -0
  34. llava_base_v2/llava_base_v2_POPE_auxmatch.xlsx +0 -0
  35. llava_base_v2/llava_base_v2_POPE_score.csv +5 -0
  36. llava_base_v2/llava_base_v2_SEEDBench_IMG.xlsx +0 -0
  37. llava_base_v2/llava_base_v2_SEEDBench_IMG_acc.csv +2 -0
  38. llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.pkl +3 -0
  39. llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx +3 -0
  40. llava_base_v2/llava_base_v2_ScienceQA_VAL.xlsx +0 -0
  41. llava_base_v2/llava_base_v2_ScienceQA_VAL_acc.csv +2 -0
  42. llava_base_v2/llava_base_v2_ScienceQA_VAL_openai_result.pkl +3 -0
  43. llava_base_v2/llava_base_v2_ScienceQA_VAL_openai_result.xlsx +0 -0
  44. llava_base_v2/llava_base_v2_TextVQA_VAL.xlsx +0 -0
  45. llava_base_v2/llava_base_v2_TextVQA_VAL_acc.csv +2 -0
  46. llava_moe_e4t2/01_MMMU_DEV_VAL.pkl +3 -0
  47. llava_moe_e4t2/llava_moe_e4t2_MME.xlsx +0 -0
  48. llava_moe_e4t2/llava_moe_e4t2_MME_auxmatch.xlsx +0 -0
  49. llava_moe_e4t2/llava_moe_e4t2_MME_score.csv +2 -0
  50. llava_moe_e4t2/llava_moe_e4t2_MMVet.xlsx +0 -0
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
37
+ llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
38
+ llava_moe_e4t2/llava_moe_e4t2_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
39
+ llava_moe_e5t3/llava_moe_e5t3_SEEDBench_IMG_openai_result.xlsx filter=lfs diff=lfs merge=lfs -text
llava_base/01_MMMU_DEV_VAL.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da276d484dc732bf2299ec90aed1b96016502c1a06adb344cf4bf5b23ace0127
3
+ size 2689
llava_base/llava_base_MMBench_DEV_EN.xlsx ADDED
Binary file (501 kB). View file
 
llava_base/llava_base_MMBench_DEV_EN_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","AR","CP","FP-C","FP-S","LR","RR","action_recognition","attribute_comparison","attribute_recognition","celebrity_recognition","function_reasoning","future_prediction","identity_reasoning","image_emotion","image_quality","image_scene","image_style","image_topic","nature_relation","object_localization","ocr","physical_property_reasoning","physical_relation","social_relation","spatial_relationship","structuralized_imagetext_understanding"
2
+ "dev","0.6030927835051546","0.6884422110552764","0.7331081081081081","0.5804195804195804","0.5972696245733788","0.3305084745762712","0.4434782608695652","0.8703703703703703","0.5454545454545454","0.6891891891891891","0.7676767676767676","0.7468354430379747","0.4","0.9555555555555556","0.8","0.20754716981132076","0.9519230769230769","0.7358490566037735","0.7777777777777778","0.375","0.30864197530864196","0.5897435897435898","0.4666666666666667","0.25","0.627906976744186","0.26666666666666666","0.2948717948717949"
llava_base/llava_base_MMBench_DEV_EN_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06c63ec6ed5f9cdca6b4ccf69e141852ad475ee14284447c3b1d9452256faf47
3
+ size 55132
llava_base/llava_base_MMBench_DEV_EN_openai_result.xlsx ADDED
Binary file (238 kB). View file
 
llava_base/llava_base_MME.xlsx ADDED
Binary file (115 kB). View file
 
llava_base/llava_base_MME_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45cd4e356c0a3339e3c1181bc0c03bdb9f217b8f44bb754031a65f3aca410962
3
+ size 44296
llava_base/llava_base_MME_auxmatch.xlsx ADDED
Binary file (130 kB). View file
 
llava_base/llava_base_MME_score.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
2
+ "1259.7040816326532","214.28571428571428","130.0","108.0","105.0","22.5","173.33333333333334","104.28571428571428","125.0","143.33333333333331","75.0","12.5","133.33333333333334","110.20408163265304","156.5","75.0"
llava_base/llava_base_MMVet.xlsx ADDED
Binary file (67.9 kB). View file
 
llava_base/llava_base_MathVista_MINI.xlsx ADDED
Binary file (132 kB). View file
 
llava_base/llava_base_POPE.xlsx ADDED
Binary file (168 kB). View file
 
llava_base/llava_base_POPE_auxmatch.xlsx ADDED
Binary file (197 kB). View file
 
llava_base/llava_base_POPE_score.csv ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ "split","Overall","acc","precision","recall"
2
+ "Overall","81.02643856920683","75.3","97.20149253731343","69.46666666666667"
3
+ "adversarial","80.37022753567297","77.3","95.33394327538883","69.46666666666667"
4
+ "popular","81.12105877773453","74.0","97.47427502338635","69.46666666666667"
5
+ "random","81.59749412685983","74.6","98.86148007590133","69.46666666666667"
llava_base/llava_base_SEEDBench_IMG.xlsx ADDED
Binary file (979 kB). View file
 
llava_base/llava_base_SEEDBench_IMG_PREV.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebaed15a17f9354992351ba742fbe8f96a156f164028239caa1513dc0b4128bc
3
+ size 82582
llava_base/llava_base_SEEDBench_IMG_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Instance Attributes","Instance Identity","Instance Interaction","Instance Location","Instances Counting","Scene Understanding","Spatial Relation","Text Understanding","Visual Reasoning"
2
+ "none","0.5265598650927488","0.5091417509141751","0.590387766247952","0.6597938144329897","0.5040899795501023","0.27257866775643647","0.7178594046865104","0.4337899543378995","0.2857142857142857","0.743202416918429"
llava_base/llava_base_SEEDBench_IMG_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00aefe1f2727f4d934914a2bed5d384533c793f46ce6b634f6bf1132a32a765a
3
+ size 1052099
llava_base/llava_base_SEEDBench_IMG_openai_result.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b848ae407b25ea999e8e77dac01c39f68d147660cfdc70f0a9bf556535bb133
3
+ size 1061587
llava_base/llava_base_ScienceQA_VAL.xlsx ADDED
Binary file (343 kB). View file
 
llava_base/llava_base_ScienceQA_VAL_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Adaptations","Adaptations and natural selection","Age of Exploration","Ancient Egypt and Kush","Ancient Mesopotamia","Animals","Astronomy","Atoms and molecules","Basic economic principles","Chemical reactions","Cities","Classification","Classification and scientific names","Climate change","Colonial America","Context clues","Descriptive details","Designing experiments","Domain-specific vocabulary","Early 19th century American history","Early Americas","Earth events","Ecological interactions","Ecosystems","Engineering practices","English colonies in North America","Force and motion","Fossils","Genes to traits","Geography","Government","Independent reading comprehension","Informational texts: level 1","Magnets","Maps","Materials","Medieval Asia","Natural resources and human impacts","Oceania: geography","Oceans and continents","Oceans and continents ","Particle motion and energy","Persuasive strategies","Physical Geography","Plant reproduction","Plants","Plate tectonics","Read-alone texts","Rocks and minerals","Rome and the Byzantine Empire","Scientific names","Solutions","State capitals","States","States of matter","The American Revolution","The Americas: geography","The Antebellum period","The Civil War and Reconstruction","The Silk Road","Thermal energy","Velocity, acceleration, and forces","Visual elements","Water cycle","Weather and climate","World religions"
2
+ "val","0.5970433953266572","0.625","1.0","1.0","1.0","1.0","1.0","0.5869565217391305","0.6190476190476191","0.53125","1.0","0.125","0.5833333333333334","0.8723404255319149","1.0","0.21428571428571427","0.8","1.0","0.9032258064516129","0.6","0.0","0.5","1.0","0.52","0.7611940298507462","0.9811320754716981","0.1794871794871795","0.8","0.29411764705882354","0.1276595744680851","0.4225352112676056","1.0","1.0","0.8","0.23308270676691728","0.3387096774193548","0.646551724137931","0.5","0.5","0.19148936170212766","0.9","0.5","0.5147058823529411","0.5555555555555556","0.6046511627906976","0.6","0.8","0.75","1.0","0.926829268292683","0.5","0.8571428571428571","0.373134328358209","0.949685534591195","0.3333333333333333","0.8947368421052632","0.6666666666666666","0.13333333333333333","1.0","0.0","1.0","0.4444444444444444","0.19607843137254902","0.5555555555555556","1.0","0.34375","0.6666666666666666"
llava_base/llava_base_ScienceQA_VAL_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76de04b9a9defce7533e24f24b51d989b49d580b600c9c0c68cb882c3c2a865e
3
+ size 142786
llava_base/llava_base_ScienceQA_VAL_openai_result.xlsx ADDED
Binary file (361 kB). View file
 
llava_base/llava_base_TextVQA_VAL.xlsx ADDED
Binary file (328 kB). View file
 
llava_base/llava_base_TextVQA_VAL_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "Overall"
2
+ "15.498000000000001"
llava_base_v2/01_MMMU_DEV_VAL.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd91a125b37689d1f7f7177eacba081afacbbd955c0a4688d74182f854fcc93c
3
+ size 1432
llava_base_v2/llava_base_v2_MME.xlsx ADDED
Binary file (103 kB). View file
 
llava_base_v2/llava_base_v2_MME_auxmatch.xlsx ADDED
Binary file (117 kB). View file
 
llava_base_v2/llava_base_v2_MME_score.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
2
+ "1268.2366946778711","265.0","95.0","110.5","69.11764705882354","57.5","185.0","115.0","153.33333333333334","190.0","56.0","47.5","128.33333333333334","130.95238095238096","150.0","45.0"
llava_base_v2/llava_base_v2_MMVet.xlsx ADDED
Binary file (68.3 kB). View file
 
llava_base_v2/llava_base_v2_MathVista_MINI.xlsx ADDED
Binary file (128 kB). View file
 
llava_base_v2/llava_base_v2_POPE.xlsx ADDED
Binary file (162 kB). View file
 
llava_base_v2/llava_base_v2_POPE_auxmatch.xlsx ADDED
Binary file (186 kB). View file
 
llava_base_v2/llava_base_v2_POPE_score.csv ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ "split","Overall","acc","precision","recall"
2
+ "Overall","82.78221208665907","84.83333333333334","96.28647214854111","72.6"
3
+ "random","83.6405529953917","85.66666666666667","98.6413043478261","72.6"
4
+ "adversarial","81.8489289740699","83.89999999999999","93.7984496124031","72.6"
5
+ "popular","82.87671232876713","84.93333333333334","96.54255319148936","72.6"
llava_base_v2/llava_base_v2_SEEDBench_IMG.xlsx ADDED
Binary file (977 kB). View file
 
llava_base_v2/llava_base_v2_SEEDBench_IMG_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Instance Attributes","Instance Identity","Instance Interaction","Instance Location","Instances Counting","Scene Understanding","Spatial Relation","Text Understanding","Visual Reasoning"
2
+ "none","0.604061270376616","0.6530436653043665","0.6810486073184052","0.7216494845360825","0.6278118609406953","0.2709440130772374","0.7412919569347688","0.5190258751902588","0.2976190476190476","0.7854984894259819"
llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2c0daae5b0338b94adaa9002dd43990b7911427a4e3789d34dbb48c9e567210
3
+ size 895520
llava_base_v2/llava_base_v2_SEEDBench_IMG_openai_result.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1fc04a9084a0a9286ff0db0b50f97609316cb52d240b64c18401463689e9cc
3
+ size 1057192
llava_base_v2/llava_base_v2_ScienceQA_VAL.xlsx ADDED
Binary file (343 kB). View file
 
llava_base_v2/llava_base_v2_ScienceQA_VAL_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "split","Overall","Adaptations","Adaptations and natural selection","Age of Exploration","Ancient Egypt and Kush","Ancient Mesopotamia","Animals","Astronomy","Atoms and molecules","Basic economic principles","Chemical reactions","Cities","Classification","Classification and scientific names","Climate change","Colonial America","Context clues","Descriptive details","Designing experiments","Domain-specific vocabulary","Early 19th century American history","Early Americas","Earth events","Ecological interactions","Ecosystems","Engineering practices","English colonies in North America","Force and motion","Fossils","Genes to traits","Geography","Government","Independent reading comprehension","Informational texts: level 1","Magnets","Maps","Materials","Medieval Asia","Natural resources and human impacts","Oceania: geography","Oceans and continents","Oceans and continents ","Particle motion and energy","Persuasive strategies","Physical Geography","Plant reproduction","Plants","Plate tectonics","Read-alone texts","Rocks and minerals","Rome and the Byzantine Empire","Scientific names","Solutions","State capitals","States","States of matter","The American Revolution","The Americas: geography","The Antebellum period","The Civil War and Reconstruction","The Silk Road","Thermal energy","Velocity, acceleration, and forces","Visual elements","Water cycle","Weather and climate","World religions"
2
+ "val","0.6547448736289938","0.84375","1.0","1.0","1.0","0.5","1.0","0.5","0.6904761904761905","0.59375","1.0","0.375","0.7619047619047619","0.9361702127659575","1.0","0.5238095238095238","0.8","1.0","1.0","0.6","0.0","0.5","1.0","0.6","0.7164179104477612","1.0","0.4358974358974359","0.68","0.47058823529411764","0.0851063829787234","0.4647887323943662","1.0","1.0","1.0","0.2781954887218045","0.5483870967741935","0.8017241379310345","0.5","0.5","0.2765957446808511","0.8","1.0","0.25","0.5555555555555556","0.5348837209302325","0.7","0.8","1.0","1.0","1.0","0.5","0.8571428571428571","0.417910447761194","0.9905660377358491","0.3333333333333333","0.7894736842105263","0.6666666666666666","0.3","1.0","0.0","1.0","0.6666666666666666","0.13725490196078433","0.7777777777777778","1.0","0.359375","1.0"
llava_base_v2/llava_base_v2_ScienceQA_VAL_openai_result.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b210adcd797510b1b99b09948c943176b28ecb065130331bae80c133e5da35f4
3
+ size 120112
llava_base_v2/llava_base_v2_ScienceQA_VAL_openai_result.xlsx ADDED
Binary file (360 kB). View file
 
llava_base_v2/llava_base_v2_TextVQA_VAL.xlsx ADDED
Binary file (328 kB). View file
 
llava_base_v2/llava_base_v2_TextVQA_VAL_acc.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "Overall"
2
+ "16.566"
llava_moe_e4t2/01_MMMU_DEV_VAL.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e93beea652c9611dccccc06ea9595d3f36ce5869400c2a0678269c360dcbe11
3
+ size 1511
llava_moe_e4t2/llava_moe_e4t2_MME.xlsx ADDED
Binary file (104 kB). View file
 
llava_moe_e4t2/llava_moe_e4t2_MME_auxmatch.xlsx ADDED
Binary file (119 kB). View file
 
llava_moe_e4t2/llava_moe_e4t2_MME_score.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ "perception","reasoning","OCR","artwork","celebrity","code_reasoning","color","commonsense_reasoning","count","existence","landmark","numerical_calculation","position","posters","scene","text_translation"
2
+ "1280.2602040816328","259.2857142857143","115.0","106.0","80.0","57.5","180.0","104.28571428571428","140.0","185.0","66.75","47.5","123.33333333333333","142.1768707482993","142.0","50.0"
llava_moe_e4t2/llava_moe_e4t2_MMVet.xlsx ADDED
Binary file (68.2 kB). View file