cc9b4cbbc93446b61caf2ef60adb62be1cbabda0051e410dada90b02fd9ece59
Browse files- sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json +0 -0
- sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank0_metric_eval_done.txt +1 -0
- sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank1_metric_eval_done.txt +1 -0
- sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank2_metric_eval_done.txt +1 -0
- sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank3_metric_eval_done.txt +1 -0
- sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/results.json +106 -0
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/ai2d.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank0_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 0 eval done
|
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank1_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 1 eval done
|
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank2_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 2 eval done
|
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/rank3_metric_eval_done.txt
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
rank 3 eval done
|
sft/665K36/revise_Full_smoe_sharev3/logs/0717_1751_llava_v1.5_ai2d_llava_model_args_5a2714/results.json
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"results": {
|
| 3 |
+
"ai2d": {
|
| 4 |
+
"exact_match,flexible-extract": 0.6473445595854922,
|
| 5 |
+
"exact_match_stderr,flexible-extract": 0.00859953028980276,
|
| 6 |
+
"alias": "ai2d"
|
| 7 |
+
}
|
| 8 |
+
},
|
| 9 |
+
"configs": {
|
| 10 |
+
"ai2d": {
|
| 11 |
+
"task": "ai2d",
|
| 12 |
+
"dataset_path": "lmms-lab/ai2d",
|
| 13 |
+
"dataset_kwargs": {
|
| 14 |
+
"token": true
|
| 15 |
+
},
|
| 16 |
+
"test_split": "test",
|
| 17 |
+
"doc_to_visual": "<function ai2d_doc_to_visual at 0x7fa461df5940>",
|
| 18 |
+
"doc_to_text": "<function ai2d_doc_to_text at 0x7fa461df5c10>",
|
| 19 |
+
"doc_to_target": "<function ai2d_doc_to_target at 0x7fa461e001f0>",
|
| 20 |
+
"description": "",
|
| 21 |
+
"target_delimiter": " ",
|
| 22 |
+
"fewshot_delimiter": "\n\n",
|
| 23 |
+
"metric_list": [
|
| 24 |
+
{
|
| 25 |
+
"metric": "exact_match",
|
| 26 |
+
"aggregation": "mean",
|
| 27 |
+
"higher_is_better": true,
|
| 28 |
+
"ignore_case": true,
|
| 29 |
+
"ignore_punctuation": true
|
| 30 |
+
}
|
| 31 |
+
],
|
| 32 |
+
"output_type": "generate_until",
|
| 33 |
+
"generation_kwargs": {
|
| 34 |
+
"max_new_tokens": 512,
|
| 35 |
+
"temperature": 0.0,
|
| 36 |
+
"do_sample": false,
|
| 37 |
+
"until": [
|
| 38 |
+
"\n\n"
|
| 39 |
+
]
|
| 40 |
+
},
|
| 41 |
+
"repeats": 1,
|
| 42 |
+
"filter_list": [
|
| 43 |
+
{
|
| 44 |
+
"name": "flexible-extract",
|
| 45 |
+
"filter": [
|
| 46 |
+
{
|
| 47 |
+
"function": "<class 'utils.MultiChoiceRegexFilter'>",
|
| 48 |
+
"group_select": 0,
|
| 49 |
+
"ignore_case": true,
|
| 50 |
+
"ignore_punctuation": true,
|
| 51 |
+
"regex_pattern": "([A-Z])\\."
|
| 52 |
+
}
|
| 53 |
+
]
|
| 54 |
+
}
|
| 55 |
+
],
|
| 56 |
+
"should_decontaminate": false,
|
| 57 |
+
"metadata": [
|
| 58 |
+
{
|
| 59 |
+
"version": 0.0
|
| 60 |
+
}
|
| 61 |
+
],
|
| 62 |
+
"model_specific_prompt_kwargs": {
|
| 63 |
+
"default": {
|
| 64 |
+
"prompt_format": "mcq",
|
| 65 |
+
"pre_prompt": "",
|
| 66 |
+
"post_prompt": "\nAnswer with the option's letter from the given choices directly."
|
| 67 |
+
},
|
| 68 |
+
"gpt4v": {
|
| 69 |
+
"prompt_format": "mcq",
|
| 70 |
+
"pre_prompt": "",
|
| 71 |
+
"post_prompt": "\nAbove choices are given in {option}. {content} format.\nPlease answer with the option letter from the given choices directly."
|
| 72 |
+
},
|
| 73 |
+
"qwen_vl": {
|
| 74 |
+
"prompt_format": "qa",
|
| 75 |
+
"pre_prompt": "",
|
| 76 |
+
"post_prompt": " Answer:"
|
| 77 |
+
},
|
| 78 |
+
"xcomposer2_4khd": {
|
| 79 |
+
"prompt_format": "mcq_xcomposer",
|
| 80 |
+
"pre_prompt": "[UNUSED_TOKEN_146]user\nQuestion: ",
|
| 81 |
+
"post_prompt": "[UNUSED_TOKEN_145]\n[UNUSED_TOKEN_146]assistant\nThe answer is"
|
| 82 |
+
}
|
| 83 |
+
},
|
| 84 |
+
"model_specific_target_kwargs": {
|
| 85 |
+
"default": "mcq",
|
| 86 |
+
"qwen_vl": "qa"
|
| 87 |
+
}
|
| 88 |
+
}
|
| 89 |
+
},
|
| 90 |
+
"versions": {
|
| 91 |
+
"ai2d": "Yaml"
|
| 92 |
+
},
|
| 93 |
+
"n-shot": {
|
| 94 |
+
"ai2d": 0
|
| 95 |
+
},
|
| 96 |
+
"model_configs": {
|
| 97 |
+
"model": "llava",
|
| 98 |
+
"model_args": "pretrained=/cm/archive/namnv78_new/revise_checkpoints/Xphi35-siglip224/SMOE/665K36/revise_Full_smoe_sharev3,conv_template=phi35",
|
| 99 |
+
"batch_size": "1",
|
| 100 |
+
"device": null,
|
| 101 |
+
"limit": null,
|
| 102 |
+
"bootstrap_iters": 100000,
|
| 103 |
+
"gen_kwargs": ""
|
| 104 |
+
},
|
| 105 |
+
"git_hash": "289c7fe5"
|
| 106 |
+
}
|