5da582ad10fdabc083e14c315b828322bc348f47d5eaaf44d3a16a6b5d728b0e
Browse files- .gitattributes +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/ocrbench.json +0 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank0_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank1_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank2_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank3_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/results.json +67 -0
- sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/results/ocrbench_results.txt +18 -0
- sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/gqa.json +3 -0
- sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank0_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank1_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank2_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank3_metric_eval_done.txt +1 -0
- sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/results.json +79 -0
.gitattributes
CHANGED
|
@@ -198,3 +198,4 @@ sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/p
|
|
| 198 |
sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 199 |
sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
|
| 200 |
sft/1M3/revise_Full_remoe/logs/0625_1751_llava...a_val_llava_model_args_179bff/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 198 |
sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
|
| 199 |
sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
|
| 200 |
sft/1M3/revise_Full_remoe/logs/0625_1751_llava...a_val_llava_model_args_179bff/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
|
| 201 |
+
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/gqa.json filter=lfs diff=lfs merge=lfs -text
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/ocrbench.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank0_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 0 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank1_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 1 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank2_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 2 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank3_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 3 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/results.json
ADDED
|
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"results": {
|
| 3 |
+
"ocrbench": {
|
| 4 |
+
"ocrbench_accuracy,none": 0.152,
|
| 5 |
+
"ocrbench_accuracy_stderr,none": "N/A",
|
| 6 |
+
"alias": "ocrbench"
|
| 7 |
+
}
|
| 8 |
+
},
|
| 9 |
+
"configs": {
|
| 10 |
+
"ocrbench": {
|
| 11 |
+
"task": "ocrbench",
|
| 12 |
+
"dataset_path": "echo840/OCRBench",
|
| 13 |
+
"dataset_kwargs": {
|
| 14 |
+
"token": true
|
| 15 |
+
},
|
| 16 |
+
"test_split": "test",
|
| 17 |
+
"doc_to_visual": "<function ocrbench_doc_to_visual at 0x7f748f49dc10>",
|
| 18 |
+
"doc_to_text": "<function ocrbench_doc_to_text at 0x7f748ed7e1f0>",
|
| 19 |
+
"doc_to_target": "answer",
|
| 20 |
+
"process_results": "<function ocrbench_process_results at 0x7f748ed7e550>",
|
| 21 |
+
"description": "",
|
| 22 |
+
"target_delimiter": " ",
|
| 23 |
+
"fewshot_delimiter": "\n\n",
|
| 24 |
+
"metric_list": [
|
| 25 |
+
{
|
| 26 |
+
"metric": "ocrbench_accuracy",
|
| 27 |
+
"aggregation": "<function ocrbench_aggregate_accuracy at 0x7f748ed7e8b0>",
|
| 28 |
+
"higher_is_better": true
|
| 29 |
+
}
|
| 30 |
+
],
|
| 31 |
+
"output_type": "generate_until",
|
| 32 |
+
"generation_kwargs": {
|
| 33 |
+
"max_new_tokens": 128,
|
| 34 |
+
"temperature": 0.0,
|
| 35 |
+
"top_p": 1.0,
|
| 36 |
+
"num_beams": 1,
|
| 37 |
+
"do_sample": false,
|
| 38 |
+
"until": [
|
| 39 |
+
"\n\n"
|
| 40 |
+
]
|
| 41 |
+
},
|
| 42 |
+
"repeats": 1,
|
| 43 |
+
"should_decontaminate": false,
|
| 44 |
+
"metadata": [
|
| 45 |
+
{
|
| 46 |
+
"version": 0.0
|
| 47 |
+
}
|
| 48 |
+
]
|
| 49 |
+
}
|
| 50 |
+
},
|
| 51 |
+
"versions": {
|
| 52 |
+
"ocrbench": "Yaml"
|
| 53 |
+
},
|
| 54 |
+
"n-shot": {
|
| 55 |
+
"ocrbench": 0
|
| 56 |
+
},
|
| 57 |
+
"model_configs": {
|
| 58 |
+
"model": "llava",
|
| 59 |
+
"model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/revise_Full_remoe,conv_template=phi35",
|
| 60 |
+
"batch_size": "1",
|
| 61 |
+
"device": null,
|
| 62 |
+
"limit": null,
|
| 63 |
+
"bootstrap_iters": 100000,
|
| 64 |
+
"gen_kwargs": ""
|
| 65 |
+
},
|
| 66 |
+
"git_hash": "289c7fe5"
|
| 67 |
+
}
|
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/results/ocrbench_results.txt
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
######################### OCRBench #############################
|
| 2 |
+
Text Recognition(Total 300): 87
|
| 3 |
+
---------------- Details of Recognition Score ------------------
|
| 4 |
+
Regular Text Recognition(Total 50): 14
|
| 5 |
+
Irregular Text Recognition(Total 50): 18
|
| 6 |
+
Artistic Text Recognition(Total 50): 34
|
| 7 |
+
Handwriting Recognition(Total 50): 14
|
| 8 |
+
Digit String Recognition(Total 50): 4
|
| 9 |
+
Non-Semantic Text Recognition(Total 50): 3
|
| 10 |
+
----------------------------------------------------------------
|
| 11 |
+
Scene Text-centric VQA(Total 200): 45
|
| 12 |
+
----------------------------------------------------------------
|
| 13 |
+
Doc-oriented VQA(Total 200): 19
|
| 14 |
+
----------------------------------------------------------------
|
| 15 |
+
Key Information Extraction(Total 200): 1
|
| 16 |
+
Handwritten Mathematical Expression Recognition(Total 100): 0
|
| 17 |
+
--------------------- Final Score ------------------------------
|
| 18 |
+
Final Score(Total 1000): 152
|
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/gqa.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:32e91639bb5de8bf0769f5cc6b51f75b93551c5a1f7fe606a984a67040a07978
|
| 3 |
+
size 38394149
|
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank0_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 0 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank1_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 1 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank2_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 2 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank3_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 3 eval done
|
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/results.json
ADDED
|
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"results": {
|
| 3 |
+
"gqa": {
|
| 4 |
+
"exact_match,none": 0.39704245508029895,
|
| 5 |
+
"exact_match_stderr,none": 0.004362880425338431,
|
| 6 |
+
"alias": "gqa"
|
| 7 |
+
}
|
| 8 |
+
},
|
| 9 |
+
"configs": {
|
| 10 |
+
"gqa": {
|
| 11 |
+
"task": "gqa",
|
| 12 |
+
"dataset_path": "lmms-lab/GQA",
|
| 13 |
+
"dataset_name": "testdev_balanced_instructions",
|
| 14 |
+
"dataset_kwargs": {
|
| 15 |
+
"token": true
|
| 16 |
+
},
|
| 17 |
+
"test_split": "testdev",
|
| 18 |
+
"doc_to_visual": "<function gqa_doc_to_visual at 0x7fef7d5f2af0>",
|
| 19 |
+
"doc_to_text": "<function gqa_doc_to_text at 0x7fef7d546dc0>",
|
| 20 |
+
"doc_to_target": "answer",
|
| 21 |
+
"description": "",
|
| 22 |
+
"target_delimiter": " ",
|
| 23 |
+
"fewshot_delimiter": "\n\n",
|
| 24 |
+
"metric_list": [
|
| 25 |
+
{
|
| 26 |
+
"metric": "exact_match",
|
| 27 |
+
"aggregation": "mean",
|
| 28 |
+
"higher_is_better": true,
|
| 29 |
+
"ignore_case": true,
|
| 30 |
+
"ignore_punctuation": true
|
| 31 |
+
}
|
| 32 |
+
],
|
| 33 |
+
"output_type": "generate_until",
|
| 34 |
+
"generation_kwargs": {
|
| 35 |
+
"max_new_tokens": 16,
|
| 36 |
+
"temperature": 0.0,
|
| 37 |
+
"top_p": 1.0,
|
| 38 |
+
"num_beams": 1,
|
| 39 |
+
"do_sample": false,
|
| 40 |
+
"until": [
|
| 41 |
+
"\n\n"
|
| 42 |
+
]
|
| 43 |
+
},
|
| 44 |
+
"repeats": 1,
|
| 45 |
+
"should_decontaminate": false,
|
| 46 |
+
"metadata": [
|
| 47 |
+
{
|
| 48 |
+
"version": 0.0
|
| 49 |
+
}
|
| 50 |
+
],
|
| 51 |
+
"model_specific_prompt_kwargs": {
|
| 52 |
+
"default": {
|
| 53 |
+
"pre_prompt": "",
|
| 54 |
+
"post_prompt": "\nAnswer the question using a single word or phrase."
|
| 55 |
+
},
|
| 56 |
+
"qwen_vl": {
|
| 57 |
+
"pre_prompt": "",
|
| 58 |
+
"post_prompt": " Answer:"
|
| 59 |
+
}
|
| 60 |
+
}
|
| 61 |
+
}
|
| 62 |
+
},
|
| 63 |
+
"versions": {
|
| 64 |
+
"gqa": "Yaml"
|
| 65 |
+
},
|
| 66 |
+
"n-shot": {
|
| 67 |
+
"gqa": 0
|
| 68 |
+
},
|
| 69 |
+
"model_configs": {
|
| 70 |
+
"model": "llava",
|
| 71 |
+
"model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/revise_Full_remoe,conv_template=phi35",
|
| 72 |
+
"batch_size": "1",
|
| 73 |
+
"device": null,
|
| 74 |
+
"limit": null,
|
| 75 |
+
"bootstrap_iters": 100000,
|
| 76 |
+
"gen_kwargs": ""
|
| 77 |
+
},
|
| 78 |
+
"git_hash": "289c7fe5"
|
| 79 |
+
}
|