DavidNguyen commited on
Commit
a136fca
·
verified ·
1 Parent(s): cf4b8a4

5da582ad10fdabc083e14c315b828322bc348f47d5eaaf44d3a16a6b5d728b0e

Browse files
.gitattributes CHANGED
@@ -198,3 +198,4 @@ sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/p
198
  sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
199
  sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
200
  sft/1M3/revise_Full_remoe/logs/0625_1751_llava...a_val_llava_model_args_179bff/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
 
 
198
  sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/submissions/mmbench_en_dev_results.xlsx filter=lfs diff=lfs merge=lfs -text
199
  sft/1M3/revise_Full_remoe/logs/0625_1719_llava..._pope_llava_model_args_179bff/textvqa_val.json filter=lfs diff=lfs merge=lfs -text
200
  sft/1M3/revise_Full_remoe/logs/0625_1751_llava...a_val_llava_model_args_179bff/infovqa_val.json filter=lfs diff=lfs merge=lfs -text
201
+ sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/gqa.json filter=lfs diff=lfs merge=lfs -text
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/ocrbench.json ADDED
The diff for this file is too large to render. See raw diff
 
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/results.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ocrbench": {
4
+ "ocrbench_accuracy,none": 0.152,
5
+ "ocrbench_accuracy_stderr,none": "N/A",
6
+ "alias": "ocrbench"
7
+ }
8
+ },
9
+ "configs": {
10
+ "ocrbench": {
11
+ "task": "ocrbench",
12
+ "dataset_path": "echo840/OCRBench",
13
+ "dataset_kwargs": {
14
+ "token": true
15
+ },
16
+ "test_split": "test",
17
+ "doc_to_visual": "<function ocrbench_doc_to_visual at 0x7f748f49dc10>",
18
+ "doc_to_text": "<function ocrbench_doc_to_text at 0x7f748ed7e1f0>",
19
+ "doc_to_target": "answer",
20
+ "process_results": "<function ocrbench_process_results at 0x7f748ed7e550>",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "ocrbench_accuracy",
27
+ "aggregation": "<function ocrbench_aggregate_accuracy at 0x7f748ed7e8b0>",
28
+ "higher_is_better": true
29
+ }
30
+ ],
31
+ "output_type": "generate_until",
32
+ "generation_kwargs": {
33
+ "max_new_tokens": 128,
34
+ "temperature": 0.0,
35
+ "top_p": 1.0,
36
+ "num_beams": 1,
37
+ "do_sample": false,
38
+ "until": [
39
+ "\n\n"
40
+ ]
41
+ },
42
+ "repeats": 1,
43
+ "should_decontaminate": false,
44
+ "metadata": [
45
+ {
46
+ "version": 0.0
47
+ }
48
+ ]
49
+ }
50
+ },
51
+ "versions": {
52
+ "ocrbench": "Yaml"
53
+ },
54
+ "n-shot": {
55
+ "ocrbench": 0
56
+ },
57
+ "model_configs": {
58
+ "model": "llava",
59
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/revise_Full_remoe,conv_template=phi35",
60
+ "batch_size": "1",
61
+ "device": null,
62
+ "limit": null,
63
+ "bootstrap_iters": 100000,
64
+ "gen_kwargs": ""
65
+ },
66
+ "git_hash": "289c7fe5"
67
+ }
sft/1M3/revise_Full_remoe/logs/0626_1433_llava...bench_llava_model_args_179bff/results/ocrbench_results.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ######################### OCRBench #############################
2
+ Text Recognition(Total 300): 87
3
+ ---------------- Details of Recognition Score ------------------
4
+ Regular Text Recognition(Total 50): 14
5
+ Irregular Text Recognition(Total 50): 18
6
+ Artistic Text Recognition(Total 50): 34
7
+ Handwriting Recognition(Total 50): 14
8
+ Digit String Recognition(Total 50): 4
9
+ Non-Semantic Text Recognition(Total 50): 3
10
+ ----------------------------------------------------------------
11
+ Scene Text-centric VQA(Total 200): 45
12
+ ----------------------------------------------------------------
13
+ Doc-oriented VQA(Total 200): 19
14
+ ----------------------------------------------------------------
15
+ Key Information Extraction(Total 200): 1
16
+ Handwritten Mathematical Expression Recognition(Total 100): 0
17
+ --------------------- Final Score ------------------------------
18
+ Final Score(Total 1000): 152
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/gqa.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32e91639bb5de8bf0769f5cc6b51f75b93551c5a1f7fe606a984a67040a07978
3
+ size 38394149
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank0_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 0 eval done
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank1_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 1 eval done
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank2_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 2 eval done
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/rank3_metric_eval_done.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ rank 3 eval done
sft/1M3/revise_Full_remoe/logs/0626_1437_llava_v1.5_gqa_llava_model_args_179bff/results.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "gqa": {
4
+ "exact_match,none": 0.39704245508029895,
5
+ "exact_match_stderr,none": 0.004362880425338431,
6
+ "alias": "gqa"
7
+ }
8
+ },
9
+ "configs": {
10
+ "gqa": {
11
+ "task": "gqa",
12
+ "dataset_path": "lmms-lab/GQA",
13
+ "dataset_name": "testdev_balanced_instructions",
14
+ "dataset_kwargs": {
15
+ "token": true
16
+ },
17
+ "test_split": "testdev",
18
+ "doc_to_visual": "<function gqa_doc_to_visual at 0x7fef7d5f2af0>",
19
+ "doc_to_text": "<function gqa_doc_to_text at 0x7fef7d546dc0>",
20
+ "doc_to_target": "answer",
21
+ "description": "",
22
+ "target_delimiter": " ",
23
+ "fewshot_delimiter": "\n\n",
24
+ "metric_list": [
25
+ {
26
+ "metric": "exact_match",
27
+ "aggregation": "mean",
28
+ "higher_is_better": true,
29
+ "ignore_case": true,
30
+ "ignore_punctuation": true
31
+ }
32
+ ],
33
+ "output_type": "generate_until",
34
+ "generation_kwargs": {
35
+ "max_new_tokens": 16,
36
+ "temperature": 0.0,
37
+ "top_p": 1.0,
38
+ "num_beams": 1,
39
+ "do_sample": false,
40
+ "until": [
41
+ "\n\n"
42
+ ]
43
+ },
44
+ "repeats": 1,
45
+ "should_decontaminate": false,
46
+ "metadata": [
47
+ {
48
+ "version": 0.0
49
+ }
50
+ ],
51
+ "model_specific_prompt_kwargs": {
52
+ "default": {
53
+ "pre_prompt": "",
54
+ "post_prompt": "\nAnswer the question using a single word or phrase."
55
+ },
56
+ "qwen_vl": {
57
+ "pre_prompt": "",
58
+ "post_prompt": " Answer:"
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "versions": {
64
+ "gqa": "Yaml"
65
+ },
66
+ "n-shot": {
67
+ "gqa": 0
68
+ },
69
+ "model_configs": {
70
+ "model": "llava",
71
+ "model_args": "pretrained=/cm/archive/namnv78/checkpoints/Xphi35-siglip224/SMOE/1M3/revise_Full_remoe,conv_template=phi35",
72
+ "batch_size": "1",
73
+ "device": null,
74
+ "limit": null,
75
+ "bootstrap_iters": 100000,
76
+ "gen_kwargs": ""
77
+ },
78
+ "git_hash": "289c7fe5"
79
+ }