yuccaaa commited on
Commit
e48e67e
·
verified ·
1 Parent(s): 48cce71

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. BIO/ablation/temperature_stability.jsonl +0 -0
  2. BIO/ablation/test.py +157 -0
  3. BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/args.json +364 -0
  4. BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/logging.jsonl +0 -0
  5. BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug-internal.log +23 -0
  6. BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug.log +22 -0
  7. BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/config.yaml +200 -0
  8. BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/output.log +16 -0
  9. BioReason_new/wandb/run-20250812_132050-2cfwmlj6/run-2cfwmlj6.wandb +0 -0
  10. LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/up_conv_block.py +101 -0
  11. LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/weight_init.py +62 -0
  12. LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/__init__.py +4 -0
  13. LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/encoding.py +74 -0
  14. LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/wrappers.py +50 -0
  15. LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/__init__.py +4 -0
  16. LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/collect_env.py +17 -0
  17. LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/logger.py +27 -0
  18. LAVIS-main/lavis/common/vqa_tools/__init__.py +8 -0
  19. LAVIS-main/lavis/common/vqa_tools/vqa.py +211 -0
  20. LAVIS-main/lavis/common/vqa_tools/vqa_eval.py +324 -0
  21. LAVIS-main/lavis/configs/datasets/aokvqa/defaults.yaml +35 -0
  22. LAVIS-main/lavis/configs/datasets/aokvqa/defaults_instruct.yaml +52 -0
  23. LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap.yaml +49 -0
  24. LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap_instruct.yaml +52 -0
  25. LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_qa.yaml +51 -0
  26. LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap.yaml +47 -0
  27. LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap_instruct.yaml +48 -0
  28. LAVIS-main/lavis/configs/datasets/avsd/defaults_dial.yaml +24 -0
  29. LAVIS-main/lavis/configs/datasets/avsd/defaults_mm_dial_instruct.yaml +65 -0
  30. LAVIS-main/lavis/configs/datasets/blip_diffusion_datasets/defaults.yaml +14 -0
  31. LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap.yaml +30 -0
  32. LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap_instruct.yaml +34 -0
  33. LAVIS-main/lavis/configs/datasets/charade/defaults_cap.yaml +52 -0
  34. LAVIS-main/lavis/configs/datasets/charade/defaults_cap_instruct.yaml +54 -0
  35. LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap.yaml +41 -0
  36. LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap_instruct.yaml +42 -0
  37. LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_qa.yaml +44 -0
  38. LAVIS-main/lavis/configs/datasets/coco/defaults_cap.yaml +28 -0
  39. LAVIS-main/lavis/configs/datasets/coco/defaults_cap_instruct.yaml +44 -0
  40. LAVIS-main/lavis/configs/datasets/coco/defaults_ret.yaml +27 -0
  41. LAVIS-main/lavis/configs/datasets/coco/defaults_vqa.yaml +41 -0
  42. LAVIS-main/lavis/configs/datasets/coco/defaults_vqa_instruct.yaml +57 -0
  43. LAVIS-main/lavis/configs/datasets/coco/eval_vqa.yaml +27 -0
  44. LAVIS-main/lavis/configs/datasets/coin/defaults_cap.yaml +51 -0
  45. LAVIS-main/lavis/configs/datasets/coin/defaults_cap_instruct.yaml +53 -0
  46. LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m.yaml +20 -0
  47. LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m_instruct.yaml +37 -0
  48. LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m.yaml +20 -0
  49. LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m_instruct.yaml +36 -0
  50. LAVIS-main/lavis/configs/datasets/didemo/defaults_ret.yaml +25 -0
BIO/ablation/temperature_stability.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
BIO/ablation/test.py ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # #accuracy
3
+ import json
4
+ import re
5
+
6
+ def compute_accuracy_from_file(filepath):
7
+ total, correct = 0, 0
8
+ pattern = re.compile(r"<ANSWER>(.*?)</ANSWER>", re.IGNORECASE)
9
+
10
+ with open(filepath, 'r', encoding='utf-8') as f:
11
+ for line in f:
12
+ line = line.strip()
13
+ if not line:
14
+ continue
15
+ data = json.loads(line)
16
+ print(data)
17
+ tgt_match = data.get("reference_answer", "")
18
+ print(tgt_match)
19
+ pred_match = pattern.search(data.get("generated_answer", ""))
20
+ #pred_match = data.get("generated_answer", "")
21
+ print(pred_match)
22
+ if not tgt_match or not pred_match:
23
+ continue
24
+
25
+ try:
26
+ tgt_val = int(tgt_match.strip())
27
+ pred_val = int(pred_match.group(1).strip())
28
+ #.group(1).
29
+ except ValueError:
30
+ continue # 如果无法转换为int,跳过此条
31
+
32
+
33
+ total += 1
34
+ if pred_val == tgt_val:
35
+ correct += 1
36
+
37
+ if total == 0:
38
+ return 0.0, 0
39
+ return correct / total, total
40
+
41
+ if __name__ == "__main__":
42
+ filepath = "/nas/shared/kilab/wangyujia/BIO/ablation/temperature_stability.jsonl"
43
+ acc, count = compute_accuracy_from_file(filepath)
44
+ print(f"Checked {count} items. Accuracy: {acc*100:.3f}%")
45
+
46
+
47
+ # #spearman
48
+ # import json
49
+ # import re
50
+ # from scipy.stats import spearmanr
51
+
52
+ # # 文件路径(替换为你的 JSONL 文件路径)
53
+ # file_path = '/nas/shared/kilab/wangyujia/BIO/ablation/temperature_stability.jsonl'
54
+
55
+ # # 正则:提取 <answer>...</answer> 中的数值
56
+ # pattern = re.compile(r"<answer>(.*?)</answer>")
57
+
58
+ # y_true = [] # reference_answer 中的真实值
59
+ # y_pred = [] # generated_answer 预测值
60
+
61
+ # with open(file_path, 'r', encoding='utf-8') as f:
62
+ # for line in f:
63
+ # line = line.strip()
64
+ # print(line)
65
+ # if not line:
66
+ # continue
67
+ # try:
68
+ # data = json.loads(line) # 每行都是一个 JSON 对象
69
+ # reference_str = data.get("reference_answer", "").strip()
70
+ # generated_str = data.get("generated_answer", "").strip()
71
+ # print(reference_str)
72
+ # print(generated_str)
73
+ # # 提取 generated_answer 中的数值
74
+ # pred_match = pattern.search(generated_str)
75
+ # print(pred_match.group(1))
76
+ # if pred_match:
77
+ # pred_value = float(pred_match.group(1)) # 提取 <answer> 中的值
78
+ # true_value = float(reference_str) # reference_answer 本身就是数值字符串
79
+ # y_true.append(true_value)
80
+ # y_pred.append(pred_value)
81
+ # else:
82
+ # print(f"未找到 <answer> 标签,跳过:{generated_str}")
83
+ # except Exception as e:
84
+ # print(f"处理行时出错:{line}")
85
+ # print(f"错误:{e}")
86
+ # continue
87
+
88
+ # # 计算 Spearman 相关系数
89
+ # if len(y_true) > 1:
90
+ # print(len(y_true))
91
+ # print(len(y_pred))
92
+ # rho, p_value = spearmanr(y_pred,y_true)
93
+ # print(f"有效样本数:{len(y_true)}")
94
+ # print(f"Spearman correlation coefficient: {rho:.5f}, p-value: {p_value:.4e}")
95
+ # else:
96
+ # print("有效数据不足,无法计算 Spearman 相关系数。")
97
+
98
+
99
+ # # f1
100
+ # import re
101
+ # import json
102
+
103
+ # def extract_numbers_from_generated_answer(s):
104
+ # """从<answer>...</answer>中提取纯数字"""
105
+ # match = re.search(r'<answer>(.*?)</answer>', s)
106
+ # if match:
107
+ # numbers_str = match.group(1)
108
+ # numbers = re.findall(r'\d+', numbers_str)
109
+ # return set(map(int, numbers))
110
+ # else:
111
+ # return set()
112
+
113
+ # def extract_numbers_from_reference_answer(s):
114
+ # """从参考答案字符串中提取数字,忽略非数字字符"""
115
+ # numbers = re.findall(r'\d+', s)
116
+ # return set(map(int, numbers))
117
+
118
+ # def calculate_f1(pred_set, target_set):
119
+ # if not pred_set and not target_set:
120
+ # return 1.0
121
+ # if not pred_set or not target_set:
122
+ # return 0.0
123
+ # tp = len(pred_set & target_set)
124
+ # fp = len(pred_set - target_set)
125
+ # fn = len(target_set - pred_set)
126
+ # precision = tp / (tp + fp) if (tp + fp) > 0 else 0.0
127
+ # recall = tp / (tp + fn) if (tp + fn) > 0 else 0.0
128
+ # if precision + recall == 0:
129
+ # return 0.0
130
+ # return 2 * precision * recall / (precision + recall)
131
+
132
+ # # 文件路径
133
+ # file_path = '/nas/shared/kilab/wangyujia/BIO/ablation/enzyme_commission_number.jsonl'
134
+
135
+ # all_f1_scores = []
136
+
137
+ # with open(file_path, 'r', encoding='utf-8') as f:
138
+ # for line in f:
139
+ # try:
140
+ # data = json.loads(line)
141
+ # reference = data.get('reference_answer', '')
142
+ # generated = data.get('generated_answer', '')
143
+ # target_set = extract_numbers_from_reference_answer(reference)
144
+ # pred_set = extract_numbers_from_generated_answer(generated)
145
+ # f1 = calculate_f1(pred_set, target_set)
146
+ # all_f1_scores.append(f1)
147
+ # except Exception as e:
148
+ # print(f"处理行时出错:{line.strip()}")
149
+ # print(f"错误:{e}")
150
+
151
+ # # 输出结果
152
+ # if all_f1_scores:
153
+ # avg_f1 = sum(all_f1_scores) / len(all_f1_scores)
154
+ # print(f"总样本数:{len(all_f1_scores)}")
155
+ # print(f"平均 F1 分数:{avg_f1:.5f}")
156
+ # else:
157
+ # print("没有有效的F1分数可供计算")
BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/args.json ADDED
@@ -0,0 +1,364 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged",
3
+ "model_type": "qwen2_5",
4
+ "model_revision": null,
5
+ "task_type": "causal_lm",
6
+ "torch_dtype": "bfloat16",
7
+ "attn_impl": null,
8
+ "num_labels": null,
9
+ "problem_type": null,
10
+ "rope_scaling": null,
11
+ "device_map": null,
12
+ "max_memory": {},
13
+ "local_repo_path": null,
14
+ "template": "qwen2_5",
15
+ "system": null,
16
+ "max_length": 8192,
17
+ "truncation_strategy": "delete",
18
+ "max_pixels": null,
19
+ "tools_prompt": "react_en",
20
+ "norm_bbox": null,
21
+ "response_prefix": null,
22
+ "padding_side": "right",
23
+ "loss_scale": "default",
24
+ "sequence_parallel_size": 1,
25
+ "use_chat_template": true,
26
+ "template_backend": "swift",
27
+ "dataset": [
28
+ "/oss/wangyujia/ProtT3/ProtT3/data/sft/dataset/metal_ion_binding/train.jsonl"
29
+ ],
30
+ "val_dataset": [],
31
+ "split_dataset_ratio": 0.01,
32
+ "data_seed": 42,
33
+ "dataset_num_proc": 128,
34
+ "dataset_shuffle": true,
35
+ "val_dataset_shuffle": false,
36
+ "streaming": false,
37
+ "interleave_prob": null,
38
+ "stopping_strategy": "first_exhausted",
39
+ "shuffle_buffer_size": 1000,
40
+ "enable_cache": false,
41
+ "download_mode": "reuse_dataset_if_exists",
42
+ "columns": {},
43
+ "strict": false,
44
+ "remove_unused_columns": true,
45
+ "model_name": [
46
+ "qwen_bio_sft_deeplocbinary-08022035"
47
+ ],
48
+ "model_author": [
49
+ "swift"
50
+ ],
51
+ "custom_dataset_info": [],
52
+ "quant_method": null,
53
+ "quant_bits": null,
54
+ "hqq_axis": null,
55
+ "bnb_4bit_compute_dtype": "bfloat16",
56
+ "bnb_4bit_quant_type": "nf4",
57
+ "bnb_4bit_use_double_quant": true,
58
+ "bnb_4bit_quant_storage": null,
59
+ "max_new_tokens": 64,
60
+ "temperature": 0.0,
61
+ "top_k": null,
62
+ "top_p": null,
63
+ "repetition_penalty": null,
64
+ "num_beams": 1,
65
+ "stream": false,
66
+ "stop_words": [],
67
+ "logprobs": false,
68
+ "top_logprobs": null,
69
+ "ckpt_dir": "/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged",
70
+ "load_dataset_config": null,
71
+ "lora_modules": [],
72
+ "tuner_backend": "peft",
73
+ "train_type": "lora",
74
+ "adapters": [],
75
+ "external_plugins": [],
76
+ "seed": 42,
77
+ "model_kwargs": {},
78
+ "load_args": false,
79
+ "load_data_args": false,
80
+ "use_hf": false,
81
+ "hub_token": null,
82
+ "custom_register_path": [],
83
+ "ignore_args_error": false,
84
+ "use_swift_lora": false,
85
+ "output_dir": "/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822",
86
+ "overwrite_output_dir": false,
87
+ "do_train": false,
88
+ "do_eval": false,
89
+ "do_predict": false,
90
+ "eval_strategy": "steps",
91
+ "prediction_loss_only": false,
92
+ "per_device_train_batch_size": 2,
93
+ "per_device_eval_batch_size": 2,
94
+ "per_gpu_train_batch_size": null,
95
+ "per_gpu_eval_batch_size": null,
96
+ "gradient_accumulation_steps": 4,
97
+ "eval_accumulation_steps": null,
98
+ "eval_delay": 0,
99
+ "torch_empty_cache_steps": null,
100
+ "learning_rate": 1e-05,
101
+ "weight_decay": 0.1,
102
+ "adam_beta1": 0.9,
103
+ "adam_beta2": 0.95,
104
+ "adam_epsilon": 1e-08,
105
+ "max_grad_norm": 1.0,
106
+ "num_train_epochs": 3.0,
107
+ "max_steps": -1,
108
+ "lr_scheduler_type": "cosine",
109
+ "lr_scheduler_kwargs": null,
110
+ "warmup_ratio": 0.05,
111
+ "warmup_steps": 0,
112
+ "log_level": "passive",
113
+ "log_level_replica": "warning",
114
+ "log_on_each_node": true,
115
+ "logging_dir": "/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/runs",
116
+ "logging_strategy": "steps",
117
+ "logging_first_step": true,
118
+ "logging_steps": 1,
119
+ "logging_nan_inf_filter": true,
120
+ "save_strategy": "steps",
121
+ "save_steps": 5.0,
122
+ "save_total_limit": 5,
123
+ "save_safetensors": true,
124
+ "save_on_each_node": false,
125
+ "save_only_model": true,
126
+ "restore_callback_states_from_checkpoint": false,
127
+ "no_cuda": false,
128
+ "use_cpu": false,
129
+ "use_mps_device": false,
130
+ "jit_mode_eval": false,
131
+ "use_ipex": false,
132
+ "bf16": true,
133
+ "fp16": false,
134
+ "fp16_opt_level": "O1",
135
+ "half_precision_backend": "auto",
136
+ "bf16_full_eval": false,
137
+ "fp16_full_eval": false,
138
+ "tf32": null,
139
+ "local_rank": 0,
140
+ "ddp_backend": null,
141
+ "tpu_num_cores": null,
142
+ "tpu_metrics_debug": false,
143
+ "debug": null,
144
+ "dataloader_drop_last": false,
145
+ "eval_steps": 5.0,
146
+ "dataloader_num_workers": 1,
147
+ "dataloader_prefetch_factor": null,
148
+ "past_index": -1,
149
+ "run_name": "construct",
150
+ "disable_tqdm": null,
151
+ "label_names": null,
152
+ "load_best_model_at_end": false,
153
+ "metric_for_best_model": "loss",
154
+ "greater_is_better": false,
155
+ "ignore_data_skip": false,
156
+ "fsdp": "",
157
+ "fsdp_min_num_params": 0,
158
+ "fsdp_config": null,
159
+ "tp_size": 0,
160
+ "fsdp_transformer_layer_cls_to_wrap": null,
161
+ "accelerator_config": {
162
+ "dispatch_batches": false
163
+ },
164
+ "deepspeed": {
165
+ "fp16": {
166
+ "enabled": "auto",
167
+ "loss_scale": 0,
168
+ "loss_scale_window": 1000,
169
+ "initial_scale_power": 16,
170
+ "hysteresis": 2,
171
+ "min_loss_scale": 1
172
+ },
173
+ "bf16": {
174
+ "enabled": "auto"
175
+ },
176
+ "zero_optimization": {
177
+ "stage": 3,
178
+ "offload_optimizer": {
179
+ "device": "none",
180
+ "pin_memory": true
181
+ },
182
+ "offload_param": {
183
+ "device": "none",
184
+ "pin_memory": true
185
+ },
186
+ "overlap_comm": false,
187
+ "contiguous_gradients": true,
188
+ "sub_group_size": 1000000000.0,
189
+ "reduce_bucket_size": "auto",
190
+ "zero_quantized_weights": false,
191
+ "zero_quantized_gradients": false,
192
+ "stage3_prefetch_bucket_size": "auto",
193
+ "stage3_param_persistence_threshold": "auto",
194
+ "stage3_max_live_parameters": 1000000000.0,
195
+ "stage3_max_reuse_distance": 1000000000.0,
196
+ "stage3_gather_16bit_weights_on_model_save": true
197
+ },
198
+ "gradient_accumulation_steps": "auto",
199
+ "gradient_clipping": "auto",
200
+ "steps_per_print": 2000,
201
+ "train_batch_size": "auto",
202
+ "train_micro_batch_size_per_gpu": "auto",
203
+ "wall_clock_breakdown": false
204
+ },
205
+ "label_smoothing_factor": 0.0,
206
+ "optim": "adamw_torch",
207
+ "optim_args": null,
208
+ "adafactor": false,
209
+ "group_by_length": false,
210
+ "length_column_name": "length",
211
+ "report_to": [
212
+ "tensorboard"
213
+ ],
214
+ "ddp_find_unused_parameters": null,
215
+ "ddp_bucket_cap_mb": null,
216
+ "ddp_broadcast_buffers": null,
217
+ "dataloader_pin_memory": true,
218
+ "dataloader_persistent_workers": false,
219
+ "skip_memory_metrics": true,
220
+ "use_legacy_prediction_loop": false,
221
+ "push_to_hub": false,
222
+ "resume_from_checkpoint": null,
223
+ "hub_model_id": null,
224
+ "hub_strategy": "every_save",
225
+ "hub_private_repo": null,
226
+ "hub_always_push": false,
227
+ "gradient_checkpointing": true,
228
+ "gradient_checkpointing_kwargs": null,
229
+ "include_inputs_for_metrics": false,
230
+ "include_for_metrics": [],
231
+ "eval_do_concat_batches": true,
232
+ "fp16_backend": "auto",
233
+ "push_to_hub_model_id": null,
234
+ "push_to_hub_organization": null,
235
+ "push_to_hub_token": null,
236
+ "mp_parameters": "",
237
+ "auto_find_batch_size": false,
238
+ "full_determinism": false,
239
+ "torchdynamo": null,
240
+ "ray_scope": "last",
241
+ "ddp_timeout": 1800,
242
+ "torch_compile": false,
243
+ "torch_compile_backend": null,
244
+ "torch_compile_mode": null,
245
+ "include_tokens_per_second": false,
246
+ "include_num_input_tokens_seen": false,
247
+ "neftune_noise_alpha": null,
248
+ "optim_target_modules": null,
249
+ "batch_eval_metrics": false,
250
+ "eval_on_start": false,
251
+ "use_liger_kernel": false,
252
+ "eval_use_gather_object": false,
253
+ "average_tokens_across_devices": false,
254
+ "sortish_sampler": false,
255
+ "predict_with_generate": false,
256
+ "generation_max_length": null,
257
+ "generation_num_beams": null,
258
+ "generation_config": null,
259
+ "check_model": true,
260
+ "acc_strategy": "token",
261
+ "train_dataloader_shuffle": true,
262
+ "metric_warmup_step": 0,
263
+ "fsdp_num": 1,
264
+ "acc_steps": 1,
265
+ "eval_use_evalscope": false,
266
+ "eval_datasets": [],
267
+ "eval_limit": null,
268
+ "eval_datasets_args": null,
269
+ "eval_generation_config": null,
270
+ "freeze_parameters": [],
271
+ "freeze_parameters_ratio": 0.0,
272
+ "trainable_parameters": [],
273
+ "freeze_llm": false,
274
+ "freeze_vit": true,
275
+ "freeze_aligner": true,
276
+ "target_modules": [
277
+ "all-linear"
278
+ ],
279
+ "target_regex": null,
280
+ "modules_to_save": [],
281
+ "lora_rank": 8,
282
+ "lora_alpha": 32,
283
+ "lora_dropout": 0.05,
284
+ "lora_bias": "none",
285
+ "lora_dtype": null,
286
+ "lorap_lr_ratio": null,
287
+ "use_rslora": false,
288
+ "use_dora": false,
289
+ "lora_ga_batch_size": 2,
290
+ "lora_ga_iters": 2,
291
+ "lora_ga_max_length": 1024,
292
+ "lora_ga_direction": "ArB2r",
293
+ "lora_ga_scale": "stable",
294
+ "lora_ga_stable_gamma": 16,
295
+ "init_weights": true,
296
+ "fourier_n_frequency": 2000,
297
+ "fourier_scaling": 300.0,
298
+ "boft_block_size": 4,
299
+ "boft_block_num": 0,
300
+ "boft_n_butterfly_factor": 1,
301
+ "boft_dropout": 0.0,
302
+ "vera_rank": 256,
303
+ "vera_projection_prng_key": 0,
304
+ "vera_dropout": 0.0,
305
+ "vera_d_initial": 0.1,
306
+ "adapter_act": "gelu",
307
+ "adapter_length": 128,
308
+ "use_galore": false,
309
+ "galore_target_modules": null,
310
+ "galore_rank": 128,
311
+ "galore_update_proj_gap": 50,
312
+ "galore_scale": 1.0,
313
+ "galore_proj_type": "std",
314
+ "galore_optim_per_parameter": false,
315
+ "galore_with_embedding": false,
316
+ "galore_quantization": false,
317
+ "galore_proj_quant": false,
318
+ "galore_proj_bits": 4,
319
+ "galore_proj_group_size": 256,
320
+ "galore_cos_threshold": 0.4,
321
+ "galore_gamma_proj": 2,
322
+ "galore_queue_size": 5,
323
+ "adalora_target_r": 8,
324
+ "adalora_init_r": 12,
325
+ "adalora_tinit": 0,
326
+ "adalora_tfinal": 0,
327
+ "adalora_deltaT": 1,
328
+ "adalora_beta1": 0.85,
329
+ "adalora_beta2": 0.85,
330
+ "adalora_orth_reg_weight": 0.5,
331
+ "llamapro_num_new_blocks": 4,
332
+ "llamapro_num_groups": null,
333
+ "lisa_activated_layers": 0,
334
+ "lisa_step_interval": 20,
335
+ "reft_layer_key": null,
336
+ "reft_layers": null,
337
+ "reft_rank": 4,
338
+ "reft_intervention_type": "LoreftIntervention",
339
+ "reft_args": null,
340
+ "swanlab_token": null,
341
+ "swanlab_project": null,
342
+ "swanlab_workspace": null,
343
+ "swanlab_exp_name": null,
344
+ "swanlab_mode": "cloud",
345
+ "add_version": true,
346
+ "resume_only_model": false,
347
+ "create_checkpoint_symlink": false,
348
+ "packing": false,
349
+ "lazy_tokenize": false,
350
+ "loss_type": null,
351
+ "optimizer": null,
352
+ "metric": null,
353
+ "zero_hpz_partition_size": null,
354
+ "rank": 0,
355
+ "global_world_size": 8,
356
+ "local_world_size": 8,
357
+ "model_suffix": "checkpoint-50-merged",
358
+ "model_info": "ModelInfo(model_type='qwen2_5', model_dir='/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged', torch_dtype=torch.bfloat16, max_model_len=32768, quant_method=None, quant_bits=None, rope_scaling=None, config=None, task_type='causal_lm', num_labels=None)",
359
+ "model_meta": "ModelMeta(model_type='qwen2_5', model_groups=[ModelGroup(models=[Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct', hf_model_id='Qwen/Qwen2.5-3B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct', hf_model_id='Qwen/Qwen2.5-7B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct', hf_model_id='Qwen/Qwen2.5-14B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct', hf_model_id='Qwen/Qwen2.5-32B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct', hf_model_id='Qwen/Qwen2.5-72B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B', hf_model_id='Qwen/Qwen2.5-0.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B', hf_model_id='Qwen/Qwen2.5-1.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B', hf_model_id='Qwen/Qwen2.5-3B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B', hf_model_id='Qwen/Qwen2.5-7B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B', hf_model_id='Qwen/Qwen2.5-14B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B', hf_model_id='Qwen/Qwen2.5-32B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B', hf_model_id='Qwen/Qwen2.5-72B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-3B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-7B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-14B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-32B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-72B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B', hf_model_id='Qwen/Qwen2.5-Coder-0.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B', hf_model_id='Qwen/Qwen2.5-Coder-1.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B', hf_model_id='Qwen/Qwen2.5-Coder-3B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B', hf_model_id='Qwen/Qwen2.5-Coder-7B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B', hf_model_id='Qwen/Qwen2.5-Coder-14B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B', hf_model_id='Qwen/Qwen2.5-Coder-32B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=['coding'])], template='qwen2_5', get_function=<function get_model_tokenizer_with_flash_attn at 0x7fbf7497de10>, model_arch='llama', architectures=['Qwen2ForCausalLM'], additional_saved_files=[], torch_dtype=None, is_multimodal=False, is_reward=False, task_type=None, ignore_patterns=[], requires=['transformers>=4.37'], tags=[])",
360
+ "model_dir": "/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged",
361
+ "hub": "<class 'swift.hub.hub.MSHub'>",
362
+ "evaluation_strategy": "steps",
363
+ "training_args": "Seq2SeqTrainingArguments(output_dir='/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822', overwrite_output_dir=False, do_train=False, do_eval=True, do_predict=False, eval_strategy=<IntervalStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=2, per_device_eval_batch_size=2, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=4, eval_accumulation_steps=None, eval_delay=0, torch_empty_cache_steps=None, learning_rate=1e-05, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=3.0, max_steps=-1, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs=None, warmup_ratio=0.05, warmup_steps=0, log_level='passive', log_level_replica='warning', log_on_each_node=True, logging_dir='/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/runs', logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_first_step=True, logging_steps=1, logging_nan_inf_filter=True, save_strategy=<SaveStrategy.STEPS: 'steps'>, save_steps=5, save_total_limit=5, save_safetensors=True, save_on_each_node=False, save_only_model=True, restore_callback_states_from_checkpoint=False, no_cuda=False, use_cpu=False, use_mps_device=False, seed=42, data_seed=42, jit_mode_eval=False, use_ipex=False, bf16=True, fp16=False, fp16_opt_level='O1', half_precision_backend='auto', bf16_full_eval=False, fp16_full_eval=False, tf32=None, local_rank=0, ddp_backend=None, tpu_num_cores=None, tpu_metrics_debug=False, debug=[], dataloader_drop_last=False, eval_steps=5, dataloader_num_workers=1, dataloader_prefetch_factor=10, past_index=-1, run_name='/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822', disable_tqdm=False, remove_unused_columns=False, label_names=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, fsdp=[], fsdp_min_num_params=0, fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, tp_size=0, fsdp_transformer_layer_cls_to_wrap=None, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=False, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None, use_configured_state=False), deepspeed={'fp16': {'enabled': 'auto', 'loss_scale': 0, 'loss_scale_window': 1000, 'initial_scale_power': 16, 'hysteresis': 2, 'min_loss_scale': 1}, 'bf16': {'enabled': 'auto'}, 'zero_optimization': {'stage': 3, 'offload_optimizer': {'device': 'none', 'pin_memory': True}, 'offload_param': {'device': 'none', 'pin_memory': True}, 'overlap_comm': False, 'contiguous_gradients': True, 'sub_group_size': 1000000000.0, 'reduce_bucket_size': 'auto', 'zero_quantized_weights': False, 'zero_quantized_gradients': False, 'stage3_prefetch_bucket_size': 'auto', 'stage3_param_persistence_threshold': 'auto', 'stage3_max_live_parameters': 1000000000.0, 'stage3_max_reuse_distance': 1000000000.0, 'stage3_gather_16bit_weights_on_model_save': True}, 'gradient_accumulation_steps': 'auto', 'gradient_clipping': 'auto', 'steps_per_print': 2000, 'train_batch_size': 'auto', 'train_micro_batch_size_per_gpu': 'auto', 'wall_clock_breakdown': False}, label_smoothing_factor=0.0, optim=<OptimizerNames.ADAMW_TORCH: 'adamw_torch'>, optim_args=None, adafactor=False, group_by_length=False, length_column_name='length', report_to=['tensorboard'], ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, dataloader_pin_memory=True, dataloader_persistent_workers=False, skip_memory_metrics=True, use_legacy_prediction_loop=False, push_to_hub=False, resume_from_checkpoint=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_token=None, hub_private_repo=None, hub_always_push=False, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, include_inputs_for_metrics=False, include_for_metrics=[], eval_do_concat_batches=True, fp16_backend='auto', push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=None, mp_parameters='', auto_find_batch_size=False, full_determinism=False, torchdynamo=None, ray_scope='last', ddp_timeout=1800, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, include_tokens_per_second=None, include_num_input_tokens_seen=None, neftune_noise_alpha=None, optim_target_modules=None, batch_eval_metrics=False, eval_on_start=False, use_liger_kernel=False, eval_use_gather_object=False, average_tokens_across_devices=None, sortish_sampler=False, predict_with_generate=False, generation_max_length=None, generation_num_beams=None, generation_config=None, check_model=True, acc_strategy='token', train_dataloader_shuffle=True, metric_warmup_step=0, fsdp_num=1, acc_steps=1, eval_use_evalscope=False, eval_datasets=[], eval_limit=None, eval_datasets_args=None, eval_generation_config=None, train_type='lora', optimizer=None, local_repo_path=None, galore_config=None)"
364
+ }
BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/logging.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug-internal.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-08-12T13:13:01.108626404+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.1"}
2
+ {"time":"2025-08-12T13:13:02.160994324+08:00","level":"INFO","msg":"stream: created new stream","id":"zr3pdtci"}
3
+ {"time":"2025-08-12T13:13:02.161037833+08:00","level":"INFO","msg":"stream: started","id":"zr3pdtci"}
4
+ {"time":"2025-08-12T13:13:02.16105997+08:00","level":"INFO","msg":"writer: started","stream_id":"zr3pdtci"}
5
+ {"time":"2025-08-12T13:13:02.161100503+08:00","level":"INFO","msg":"sender: started","stream_id":"zr3pdtci"}
6
+ {"time":"2025-08-12T13:13:02.161076037+08:00","level":"INFO","msg":"handler: started","stream_id":"zr3pdtci"}
7
+ {"time":"2025-08-12T13:13:08.73352003+08:00","level":"INFO","msg":"stream: closing","id":"zr3pdtci"}
8
+ {"time":"2025-08-12T13:13:09.557916985+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39586->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
9
+ {"time":"2025-08-12T13:13:09.571553285+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39582->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
10
+ {"time":"2025-08-12T13:13:09.661494759+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39570->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
11
+ {"time":"2025-08-12T13:15:13.828654367+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:42648->142.250.217.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
12
+ {"time":"2025-08-12T13:15:13.836170243+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:42646->142.250.217.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
13
+ {"time":"2025-08-12T13:15:16.514141941+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39884->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
14
+ {"time":"2025-08-12T13:15:22.024313571+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:42662->142.250.217.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
15
+ {"time":"2025-08-12T13:17:24.906497736+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:55432->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
16
+ {"time":"2025-08-12T13:17:33.096841516+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:59370->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
17
+ {"time":"2025-08-12T13:17:33.124262755+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:53630->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
18
+ {"time":"2025-08-12T13:17:33.139133433+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:59376->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
19
+ {"time":"2025-08-12T13:17:37.290872326+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39894->142.250.73.155:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
20
+ {"time":"2025-08-12T13:17:47.6529587+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
21
+ {"time":"2025-08-12T13:17:48.610451999+08:00","level":"INFO","msg":"handler: closed","stream_id":"zr3pdtci"}
22
+ {"time":"2025-08-12T13:17:48.613740545+08:00","level":"INFO","msg":"sender: closed","stream_id":"zr3pdtci"}
23
+ {"time":"2025-08-12T13:17:48.613750169+08:00","level":"INFO","msg":"stream: closed","id":"zr3pdtci"}
BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug.log ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-08-12 13:13:00,887 INFO MainThread:87959 [wandb_setup.py:_flush():80] Current SDK version is 0.21.1
2
+ 2025-08-12 13:13:00,887 INFO MainThread:87959 [wandb_setup.py:_flush():80] Configure stats pid to 87959
3
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
4
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_setup.py:_flush():80] Loading settings from /nas/shared/kilab/wangyujia/BioReason_new/wandb/settings
5
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_setup.py:_flush():80] Loading settings from environment variables
6
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /nas/shared/kilab/wangyujia/BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug.log
7
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /nas/shared/kilab/wangyujia/BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug-internal.log
8
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:init():830] calling init triggers
9
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
10
+ config: {'text_model_name': '/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged', 'protein_model_name': '/nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m', 'qformer_model_name': '/nas/shared/kilab/wangyujia/ProtT3/plm_model/microsoft', 'cache_dir': '/model-weights', 'num_query_tokens': 8, 'train_dataset': '/nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/train_set.jsonl', 'valid_dataset': '/nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/valid_set.jsonl', 'eval_dataset': True, 'output_dir': './contrastive_outputs', 'num_epochs': 10, 'batch_size': 32, 'learning_rate': 0.0001, 'weight_decay': 0.01, 'warmup_steps': 1000, 'gradient_accumulation_steps': 1, 'temperature': 0.07, 'freeze_protein_model': True, 'freeze_text_model': True, 'protein_weight': 1.0, 'text_weight': 1.0, 'enable_ptm': True, 'ptm_weight': 1.0, 'max_length_protein': 1024, 'max_length_text': 512, 'num_workers': 8, 'logging_steps': 100, 'eval_steps': 500, 'save_steps': 1000, 'save_total_limit': 3, 'fp16': False, 'bf16': False, 'seed': 42, 'use_wandb': True, 'wandb_project': 'protein-llm-contrastive', 'wandb_entity': None, '_wandb': {}}
11
+ 2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:init():871] starting backend
12
+ 2025-08-12 13:13:01,096 INFO MainThread:87959 [wandb_init.py:init():874] sending inform_init request
13
+ 2025-08-12 13:13:01,103 INFO MainThread:87959 [wandb_init.py:init():882] backend started and connected
14
+ 2025-08-12 13:13:01,104 INFO MainThread:87959 [wandb_init.py:init():953] updated telemetry
15
+ 2025-08-12 13:13:01,135 INFO MainThread:87959 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
16
+ 2025-08-12 13:13:02,774 INFO MainThread:87959 [wandb_init.py:init():1029] starting run threads in backend
17
+ 2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_console_start():2494] atexit reg
18
+ 2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_redirect():2342] redirect: wrap_raw
19
+ 2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_redirect():2411] Wrapping output streams.
20
+ 2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_redirect():2434] Redirects installed.
21
+ 2025-08-12 13:13:02,886 INFO MainThread:87959 [wandb_init.py:init():1075] run started, returning control to user process
22
+ 2025-08-12 13:13:08,732 INFO MsgRouterThr:87959 [mailbox.py:close():129] [no run ID] Closing mailbox, abandoning 1 handles.
BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/config.yaml ADDED
@@ -0,0 +1,200 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.21.1
4
+ e:
5
+ crh90aob5eyx2qklfrtgkofnoya37sbu:
6
+ args:
7
+ - --text_model_name
8
+ - /oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged
9
+ - --protein_model_name
10
+ - /nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m
11
+ - --qformer_model_name
12
+ - /nas/shared/kilab/wangyujia/ProtT3/plm_model/microsoft
13
+ - --num_query_tokens
14
+ - "8"
15
+ - --train_dataset
16
+ - /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/train_set.jsonl
17
+ - --valid_dataset
18
+ - /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/valid_set.jsonl
19
+ - --output_dir
20
+ - ./contrastive_outputs
21
+ - --num_epochs
22
+ - "10"
23
+ - --batch_size
24
+ - "32"
25
+ - --learning_rate
26
+ - "1e-4"
27
+ - --temperature
28
+ - "0.07"
29
+ - --freeze_protein_model
30
+ - --freeze_text_model
31
+ - --enable_ptm
32
+ - --max_length_protein
33
+ - "1024"
34
+ - --max_length_text
35
+ - "512"
36
+ - --num_workers
37
+ - "8"
38
+ - --eval_dataset
39
+ - --use_wandb
40
+ - --wandb_project
41
+ - protein-llm-contrastive
42
+ - --logging_steps
43
+ - "100"
44
+ - --eval_steps
45
+ - "500"
46
+ - --save_steps
47
+ - "1000"
48
+ codePath: wangyujia/BioReason_new/train_contrastive.py
49
+ codePathLocal: train_contrastive.py
50
+ cpu_count: 64
51
+ cpu_count_logical: 64
52
+ cudaVersion: "12.1"
53
+ disk:
54
+ /:
55
+ total: "1623302262784"
56
+ used: "29131816960"
57
+ email: gia0603yucca@gmail.com
58
+ executable: /root/miniconda3/envs/bioreason/bin/python
59
+ git:
60
+ commit: b8caf406aa1699c788f0ca6e44a1769452c317db
61
+ remote: https://github.com/PorUna-byte/PAR.git
62
+ gpu: NVIDIA A800-SXM4-80GB
63
+ gpu_count: 8
64
+ gpu_nvidia:
65
+ - architecture: Ampere
66
+ name: NVIDIA A800-SXM4-80GB
67
+ uuid: GPU-71607f78-ad31-1ea4-19c1-908e3e31aaf1
68
+ - architecture: Ampere
69
+ name: NVIDIA A800-SXM4-80GB
70
+ uuid: GPU-92b7dbbd-7ef5-3c5f-ce1c-1d179d7fa587
71
+ - architecture: Ampere
72
+ name: NVIDIA A800-SXM4-80GB
73
+ uuid: GPU-bbc35439-ad79-578b-381b-aba6f0cc0168
74
+ - architecture: Ampere
75
+ name: NVIDIA A800-SXM4-80GB
76
+ uuid: GPU-e492e147-ca2e-76f2-85da-4e08e4deeb14
77
+ - architecture: Ampere
78
+ name: NVIDIA A800-SXM4-80GB
79
+ uuid: GPU-8c4f8e67-4b52-5107-3095-0f007e6378ac
80
+ - architecture: Ampere
81
+ name: NVIDIA A800-SXM4-80GB
82
+ uuid: GPU-7063f0b9-4ca2-6a72-522d-1262899ac5ad
83
+ - architecture: Ampere
84
+ name: NVIDIA A800-SXM4-80GB
85
+ uuid: GPU-3b6e9a37-bcf3-387c-7874-4f8de4abd115
86
+ - architecture: Ampere
87
+ name: NVIDIA A800-SXM4-80GB
88
+ uuid: GPU-92456839-e814-7be9-6817-f3e8da8aa80c
89
+ host: dsw-265304-f8bc5ff76-4mdt5
90
+ memory:
91
+ total: "549755813888"
92
+ os: Linux-5.10.134-008.18.kangaroo.al8.x86_64-x86_64-with-glibc2.35
93
+ program: /nas/shared/kilab/wangyujia/BioReason_new/train_contrastive.py
94
+ python: CPython 3.11.0
95
+ root: /nas/shared/kilab/wangyujia/BioReason_new
96
+ startedAt: "2025-08-12T05:20:50.783496Z"
97
+ writerId: crh90aob5eyx2qklfrtgkofnoya37sbu
98
+ m: []
99
+ python_version: 3.11.0
100
+ t:
101
+ "1":
102
+ - 1
103
+ - 9
104
+ - 11
105
+ - 41
106
+ - 49
107
+ - 51
108
+ - 71
109
+ - 84
110
+ - 98
111
+ - 103
112
+ "2":
113
+ - 1
114
+ - 9
115
+ - 11
116
+ - 41
117
+ - 49
118
+ - 51
119
+ - 71
120
+ - 84
121
+ - 98
122
+ - 103
123
+ "3":
124
+ - 13
125
+ - 16
126
+ "4": 3.11.0
127
+ "5": 0.21.1
128
+ "6": 4.55.0
129
+ "12": 0.21.1
130
+ "13": linux-x86_64
131
+ batch_size:
132
+ value: 32
133
+ bf16:
134
+ value: false
135
+ cache_dir:
136
+ value: /model-weights
137
+ enable_ptm:
138
+ value: true
139
+ eval_dataset:
140
+ value: true
141
+ eval_steps:
142
+ value: 500
143
+ fp16:
144
+ value: false
145
+ freeze_protein_model:
146
+ value: true
147
+ freeze_text_model:
148
+ value: true
149
+ gradient_accumulation_steps:
150
+ value: 1
151
+ learning_rate:
152
+ value: 0.0001
153
+ logging_steps:
154
+ value: 100
155
+ max_length_protein:
156
+ value: 1024
157
+ max_length_text:
158
+ value: 512
159
+ num_epochs:
160
+ value: 10
161
+ num_query_tokens:
162
+ value: 8
163
+ num_workers:
164
+ value: 8
165
+ output_dir:
166
+ value: ./contrastive_outputs
167
+ protein_model_name:
168
+ value: /nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m
169
+ protein_weight:
170
+ value: 1
171
+ ptm_weight:
172
+ value: 1
173
+ qformer_model_name:
174
+ value: /nas/shared/kilab/wangyujia/ProtT3/plm_model/microsoft
175
+ save_steps:
176
+ value: 1000
177
+ save_total_limit:
178
+ value: 3
179
+ seed:
180
+ value: 42
181
+ temperature:
182
+ value: 0.07
183
+ text_model_name:
184
+ value: /oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged
185
+ text_weight:
186
+ value: 1
187
+ train_dataset:
188
+ value: /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/train_set.jsonl
189
+ use_wandb:
190
+ value: true
191
+ valid_dataset:
192
+ value: /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/valid_set.jsonl
193
+ wandb_entity:
194
+ value: null
195
+ wandb_project:
196
+ value: protein-llm-contrastive
197
+ warmup_steps:
198
+ value: 1000
199
+ weight_decay:
200
+ value: 0.01
BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/output.log ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading model...
2
+ Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████| 4/4 [00:01<00:00, 2.57it/s]
3
+ Some weights of EsmModel were not initialized from the model checkpoint at /nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m and are newly initialized: ['pooler.dense.bias', 'pooler.dense.weight']
4
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
5
+ Loading datasets...
6
+ Training dataset size: 430595
7
+ Eval dataset size: 10000
8
+ bioreason.trainer.contrast_trainer_new
9
+ Traceback (most recent call last):
10
+ File "/nas/shared/kilab/wangyujia/BioReason_new/train_contrastive.py", line 552, in <module>
11
+ trainer = main(args)
12
+ ^^^^^^^^^^
13
+ File "/nas/shared/kilab/wangyujia/BioReason_new/train_contrastive.py", line 345, in main
14
+ training_args = ContrastiveTrainingArguments(
15
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
16
+ TypeError: ContrastiveTrainingArguments.__init__() got an unexpected keyword argument 'evaluation_strategy'
BioReason_new/wandb/run-20250812_132050-2cfwmlj6/run-2cfwmlj6.wandb ADDED
Binary file (7.45 kB). View file
 
LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/up_conv_block.py ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+ from annotator.uniformer.mmcv.cnn import ConvModule, build_upsample_layer
4
+
5
+
6
+ class UpConvBlock(nn.Module):
7
+ """Upsample convolution block in decoder for UNet.
8
+
9
+ This upsample convolution block consists of one upsample module
10
+ followed by one convolution block. The upsample module expands the
11
+ high-level low-resolution feature map and the convolution block fuses
12
+ the upsampled high-level low-resolution feature map and the low-level
13
+ high-resolution feature map from encoder.
14
+
15
+ Args:
16
+ conv_block (nn.Sequential): Sequential of convolutional layers.
17
+ in_channels (int): Number of input channels of the high-level
18
+ skip_channels (int): Number of input channels of the low-level
19
+ high-resolution feature map from encoder.
20
+ out_channels (int): Number of output channels.
21
+ num_convs (int): Number of convolutional layers in the conv_block.
22
+ Default: 2.
23
+ stride (int): Stride of convolutional layer in conv_block. Default: 1.
24
+ dilation (int): Dilation rate of convolutional layer in conv_block.
25
+ Default: 1.
26
+ with_cp (bool): Use checkpoint or not. Using checkpoint will save some
27
+ memory while slowing down the training speed. Default: False.
28
+ conv_cfg (dict | None): Config dict for convolution layer.
29
+ Default: None.
30
+ norm_cfg (dict | None): Config dict for normalization layer.
31
+ Default: dict(type='BN').
32
+ act_cfg (dict | None): Config dict for activation layer in ConvModule.
33
+ Default: dict(type='ReLU').
34
+ upsample_cfg (dict): The upsample config of the upsample module in
35
+ decoder. Default: dict(type='InterpConv'). If the size of
36
+ high-level feature map is the same as that of skip feature map
37
+ (low-level feature map from encoder), it does not need upsample the
38
+ high-level feature map and the upsample_cfg is None.
39
+ dcn (bool): Use deformable convolution in convolutional layer or not.
40
+ Default: None.
41
+ plugins (dict): plugins for convolutional layers. Default: None.
42
+ """
43
+
44
+ def __init__(self,
45
+ conv_block,
46
+ in_channels,
47
+ skip_channels,
48
+ out_channels,
49
+ num_convs=2,
50
+ stride=1,
51
+ dilation=1,
52
+ with_cp=False,
53
+ conv_cfg=None,
54
+ norm_cfg=dict(type='BN'),
55
+ act_cfg=dict(type='ReLU'),
56
+ upsample_cfg=dict(type='InterpConv'),
57
+ dcn=None,
58
+ plugins=None):
59
+ super(UpConvBlock, self).__init__()
60
+ assert dcn is None, 'Not implemented yet.'
61
+ assert plugins is None, 'Not implemented yet.'
62
+
63
+ self.conv_block = conv_block(
64
+ in_channels=2 * skip_channels,
65
+ out_channels=out_channels,
66
+ num_convs=num_convs,
67
+ stride=stride,
68
+ dilation=dilation,
69
+ with_cp=with_cp,
70
+ conv_cfg=conv_cfg,
71
+ norm_cfg=norm_cfg,
72
+ act_cfg=act_cfg,
73
+ dcn=None,
74
+ plugins=None)
75
+ if upsample_cfg is not None:
76
+ self.upsample = build_upsample_layer(
77
+ cfg=upsample_cfg,
78
+ in_channels=in_channels,
79
+ out_channels=skip_channels,
80
+ with_cp=with_cp,
81
+ norm_cfg=norm_cfg,
82
+ act_cfg=act_cfg)
83
+ else:
84
+ self.upsample = ConvModule(
85
+ in_channels,
86
+ skip_channels,
87
+ kernel_size=1,
88
+ stride=1,
89
+ padding=0,
90
+ conv_cfg=conv_cfg,
91
+ norm_cfg=norm_cfg,
92
+ act_cfg=act_cfg)
93
+
94
+ def forward(self, skip, x):
95
+ """Forward function."""
96
+
97
+ x = self.upsample(x)
98
+ out = torch.cat([skip, x], dim=1)
99
+ out = self.conv_block(out)
100
+
101
+ return out
LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/weight_init.py ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Modified from https://github.com/rwightman/pytorch-image-
2
+ models/blob/master/timm/models/layers/drop.py."""
3
+
4
+ import math
5
+ import warnings
6
+
7
+ import torch
8
+
9
+
10
+ def _no_grad_trunc_normal_(tensor, mean, std, a, b):
11
+ """Reference: https://people.sc.fsu.edu/~jburkardt/presentations
12
+ /truncated_normal.pdf"""
13
+
14
+ def norm_cdf(x):
15
+ # Computes standard normal cumulative distribution function
16
+ return (1. + math.erf(x / math.sqrt(2.))) / 2.
17
+
18
+ if (mean < a - 2 * std) or (mean > b + 2 * std):
19
+ warnings.warn(
20
+ 'mean is more than 2 std from [a, b] in nn.init.trunc_normal_. '
21
+ 'The distribution of values may be incorrect.',
22
+ stacklevel=2)
23
+
24
+ with torch.no_grad():
25
+ # Values are generated by using a truncated uniform distribution and
26
+ # then using the inverse CDF for the normal distribution.
27
+ # Get upper and lower cdf values
28
+ lower_bound = norm_cdf((a - mean) / std)
29
+ upper_bound = norm_cdf((b - mean) / std)
30
+
31
+ # Uniformly fill tensor with values from [l, u], then translate to
32
+ # [2l-1, 2u-1].
33
+ tensor.uniform_(2 * lower_bound - 1, 2 * upper_bound - 1)
34
+
35
+ # Use inverse cdf transform for normal distribution to get truncated
36
+ # standard normal
37
+ tensor.erfinv_()
38
+
39
+ # Transform to proper mean, std
40
+ tensor.mul_(std * math.sqrt(2.))
41
+ tensor.add_(mean)
42
+
43
+ # Clamp to ensure it's in the proper range
44
+ tensor.clamp_(min=a, max=b)
45
+ return tensor
46
+
47
+
48
+ def trunc_normal_(tensor, mean=0., std=1., a=-2., b=2.):
49
+ r"""Fills the input Tensor with values drawn from a truncated
50
+ normal distribution. The values are effectively drawn from the
51
+ normal distribution :math:`\mathcal{N}(\text{mean}, \text{std}^2)`
52
+ with values outside :math:`[a, b]` redrawn until they are within
53
+ the bounds. The method used for generating the random values works
54
+ best when :math:`a \leq \text{mean} \leq b`.
55
+ Args:
56
+ tensor (``torch.Tensor``): an n-dimensional `torch.Tensor`
57
+ mean (float): the mean of the normal distribution
58
+ std (float): the standard deviation of the normal distribution
59
+ a (float): the minimum cutoff value
60
+ b (float): the maximum cutoff value
61
+ """
62
+ return _no_grad_trunc_normal_(tensor, mean, std, a, b)
LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .encoding import Encoding
2
+ from .wrappers import Upsample, resize
3
+
4
+ __all__ = ['Upsample', 'resize', 'Encoding']
LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/encoding.py ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ from torch import nn
3
+ from torch.nn import functional as F
4
+
5
+
6
+ class Encoding(nn.Module):
7
+ """Encoding Layer: a learnable residual encoder.
8
+
9
+ Input is of shape (batch_size, channels, height, width).
10
+ Output is of shape (batch_size, num_codes, channels).
11
+
12
+ Args:
13
+ channels: dimension of the features or feature channels
14
+ num_codes: number of code words
15
+ """
16
+
17
+ def __init__(self, channels, num_codes):
18
+ super(Encoding, self).__init__()
19
+ # init codewords and smoothing factor
20
+ self.channels, self.num_codes = channels, num_codes
21
+ std = 1. / ((num_codes * channels)**0.5)
22
+ # [num_codes, channels]
23
+ self.codewords = nn.Parameter(
24
+ torch.empty(num_codes, channels,
25
+ dtype=torch.float).uniform_(-std, std),
26
+ requires_grad=True)
27
+ # [num_codes]
28
+ self.scale = nn.Parameter(
29
+ torch.empty(num_codes, dtype=torch.float).uniform_(-1, 0),
30
+ requires_grad=True)
31
+
32
+ @staticmethod
33
+ def scaled_l2(x, codewords, scale):
34
+ num_codes, channels = codewords.size()
35
+ batch_size = x.size(0)
36
+ reshaped_scale = scale.view((1, 1, num_codes))
37
+ expanded_x = x.unsqueeze(2).expand(
38
+ (batch_size, x.size(1), num_codes, channels))
39
+ reshaped_codewords = codewords.view((1, 1, num_codes, channels))
40
+
41
+ scaled_l2_norm = reshaped_scale * (
42
+ expanded_x - reshaped_codewords).pow(2).sum(dim=3)
43
+ return scaled_l2_norm
44
+
45
+ @staticmethod
46
+ def aggregate(assignment_weights, x, codewords):
47
+ num_codes, channels = codewords.size()
48
+ reshaped_codewords = codewords.view((1, 1, num_codes, channels))
49
+ batch_size = x.size(0)
50
+
51
+ expanded_x = x.unsqueeze(2).expand(
52
+ (batch_size, x.size(1), num_codes, channels))
53
+ encoded_feat = (assignment_weights.unsqueeze(3) *
54
+ (expanded_x - reshaped_codewords)).sum(dim=1)
55
+ return encoded_feat
56
+
57
+ def forward(self, x):
58
+ assert x.dim() == 4 and x.size(1) == self.channels
59
+ # [batch_size, channels, height, width]
60
+ batch_size = x.size(0)
61
+ # [batch_size, height x width, channels]
62
+ x = x.view(batch_size, self.channels, -1).transpose(1, 2).contiguous()
63
+ # assignment_weights: [batch_size, channels, num_codes]
64
+ assignment_weights = F.softmax(
65
+ self.scaled_l2(x, self.codewords, self.scale), dim=2)
66
+ # aggregate
67
+ encoded_feat = self.aggregate(assignment_weights, x, self.codewords)
68
+ return encoded_feat
69
+
70
+ def __repr__(self):
71
+ repr_str = self.__class__.__name__
72
+ repr_str += f'(Nx{self.channels}xHxW =>Nx{self.num_codes}' \
73
+ f'x{self.channels})'
74
+ return repr_str
LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/wrappers.py ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import warnings
2
+
3
+ import torch.nn as nn
4
+ import torch.nn.functional as F
5
+
6
+
7
+ def resize(input,
8
+ size=None,
9
+ scale_factor=None,
10
+ mode='nearest',
11
+ align_corners=None,
12
+ warning=True):
13
+ if warning:
14
+ if size is not None and align_corners:
15
+ input_h, input_w = tuple(int(x) for x in input.shape[2:])
16
+ output_h, output_w = tuple(int(x) for x in size)
17
+ if output_h > input_h or output_w > output_h:
18
+ if ((output_h > 1 and output_w > 1 and input_h > 1
19
+ and input_w > 1) and (output_h - 1) % (input_h - 1)
20
+ and (output_w - 1) % (input_w - 1)):
21
+ warnings.warn(
22
+ f'When align_corners={align_corners}, '
23
+ 'the output would more aligned if '
24
+ f'input size {(input_h, input_w)} is `x+1` and '
25
+ f'out size {(output_h, output_w)} is `nx+1`')
26
+ return F.interpolate(input, size, scale_factor, mode, align_corners)
27
+
28
+
29
+ class Upsample(nn.Module):
30
+
31
+ def __init__(self,
32
+ size=None,
33
+ scale_factor=None,
34
+ mode='nearest',
35
+ align_corners=None):
36
+ super(Upsample, self).__init__()
37
+ self.size = size
38
+ if isinstance(scale_factor, tuple):
39
+ self.scale_factor = tuple(float(factor) for factor in scale_factor)
40
+ else:
41
+ self.scale_factor = float(scale_factor) if scale_factor else None
42
+ self.mode = mode
43
+ self.align_corners = align_corners
44
+
45
+ def forward(self, x):
46
+ if not self.size:
47
+ size = [int(t * self.scale_factor) for t in x.shape[-2:]]
48
+ else:
49
+ size = self.size
50
+ return resize(x, size, None, self.mode, self.align_corners)
LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .collect_env import collect_env
2
+ from .logger import get_root_logger
3
+
4
+ __all__ = ['get_root_logger', 'collect_env']
LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/collect_env.py ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from annotator.uniformer.mmcv.utils import collect_env as collect_base_env
2
+ from annotator.uniformer.mmcv.utils import get_git_hash
3
+
4
+ import annotator.uniformer.mmseg as mmseg
5
+
6
+
7
+ def collect_env():
8
+ """Collect the information of the running environments."""
9
+ env_info = collect_base_env()
10
+ env_info['MMSegmentation'] = f'{mmseg.__version__}+{get_git_hash()[:7]}'
11
+
12
+ return env_info
13
+
14
+
15
+ if __name__ == '__main__':
16
+ for name, val in collect_env().items():
17
+ print('{}: {}'.format(name, val))
LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/logger.py ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import logging
2
+
3
+ from annotator.uniformer.mmcv.utils import get_logger
4
+
5
+
6
+ def get_root_logger(log_file=None, log_level=logging.INFO):
7
+ """Get the root logger.
8
+
9
+ The logger will be initialized if it has not been initialized. By default a
10
+ StreamHandler will be added. If `log_file` is specified, a FileHandler will
11
+ also be added. The name of the root logger is the top-level package name,
12
+ e.g., "mmseg".
13
+
14
+ Args:
15
+ log_file (str | None): The log filename. If specified, a FileHandler
16
+ will be added to the root logger.
17
+ log_level (int): The root logger level. Note that only the process of
18
+ rank 0 is affected, while other processes will set the level to
19
+ "Error" and be silent most of the time.
20
+
21
+ Returns:
22
+ logging.Logger: The root logger.
23
+ """
24
+
25
+ logger = get_logger(name='mmseg', log_file=log_file, log_level=log_level)
26
+
27
+ return logger
LAVIS-main/lavis/common/vqa_tools/__init__.py ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Copyright (c) 2022, salesforce.com, inc.
3
+ All rights reserved.
4
+ SPDX-License-Identifier: BSD-3-Clause
5
+ For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
6
+ """
7
+
8
+ __author__ = "aagrawal"
LAVIS-main/lavis/common/vqa_tools/vqa.py ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Copyright (c) 2022, salesforce.com, inc.
3
+ All rights reserved.
4
+ SPDX-License-Identifier: BSD-3-Clause
5
+ For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
6
+ """
7
+
8
+ __author__ = "aagrawal"
9
+ __version__ = "0.9"
10
+
11
+ # Interface for accessing the VQA dataset.
12
+
13
+ # This code is based on the code written by Tsung-Yi Lin for MSCOCO Python API available at the following link:
14
+ # (https://github.com/pdollar/coco/blob/master/PythonAPI/pycocotools/coco.py).
15
+
16
+ # The following functions are defined:
17
+ # VQA - VQA class that loads VQA annotation file and prepares data structures.
18
+ # getQuesIds - Get question ids that satisfy given filter conditions.
19
+ # getImgIds - Get image ids that satisfy given filter conditions.
20
+ # loadQA - Load questions and answers with the specified question ids.
21
+ # showQA - Display the specified questions and answers.
22
+ # loadRes - Load result file and create result object.
23
+
24
+ # Help on each function can be accessed by: "help(COCO.function)"
25
+
26
+ import json
27
+ import datetime
28
+ import copy
29
+
30
+
31
+ class VQA:
32
+ def __init__(self, annotation_file=None, question_file=None):
33
+ """
34
+ Constructor of VQA helper class for reading and visualizing questions and answers.
35
+ :param annotation_file (str): location of VQA annotation file
36
+ :return:
37
+ """
38
+ # load dataset
39
+ self.dataset = {}
40
+ self.questions = {}
41
+ self.qa = {}
42
+ self.qqa = {}
43
+ self.imgToQA = {}
44
+ if not annotation_file == None and not question_file == None:
45
+ print("loading VQA annotations and questions into memory...")
46
+ time_t = datetime.datetime.utcnow()
47
+ dataset = json.load(open(annotation_file, "r"))
48
+ questions = json.load(open(question_file, "r"))
49
+ self.dataset = dataset
50
+ self.questions = questions
51
+ self.createIndex()
52
+
53
+ def createIndex(self):
54
+ # create index
55
+ print("creating index...")
56
+ imgToQA = {ann["image_id"]: [] for ann in self.dataset["annotations"]}
57
+ qa = {ann["question_id"]: [] for ann in self.dataset["annotations"]}
58
+ qqa = {ann["question_id"]: [] for ann in self.dataset["annotations"]}
59
+ for ann in self.dataset["annotations"]:
60
+ imgToQA[ann["image_id"]] += [ann]
61
+ qa[ann["question_id"]] = ann
62
+ for ques in self.questions["questions"]:
63
+ qqa[ques["question_id"]] = ques
64
+ print("index created!")
65
+
66
+ # create class members
67
+ self.qa = qa
68
+ self.qqa = qqa
69
+ self.imgToQA = imgToQA
70
+
71
+ def info(self):
72
+ """
73
+ Print information about the VQA annotation file.
74
+ :return:
75
+ """
76
+ for key, value in self.datset["info"].items():
77
+ print("%s: %s" % (key, value))
78
+
79
+ def getQuesIds(self, imgIds=[], quesTypes=[], ansTypes=[]):
80
+ """
81
+ Get question ids that satisfy given filter conditions. default skips that filter
82
+ :param imgIds (int array) : get question ids for given imgs
83
+ quesTypes (str array) : get question ids for given question types
84
+ ansTypes (str array) : get question ids for given answer types
85
+ :return: ids (int array) : integer array of question ids
86
+ """
87
+ imgIds = imgIds if type(imgIds) == list else [imgIds]
88
+ quesTypes = quesTypes if type(quesTypes) == list else [quesTypes]
89
+ ansTypes = ansTypes if type(ansTypes) == list else [ansTypes]
90
+
91
+ if len(imgIds) == len(quesTypes) == len(ansTypes) == 0:
92
+ anns = self.dataset["annotations"]
93
+ else:
94
+ if not len(imgIds) == 0:
95
+ anns = sum(
96
+ [self.imgToQA[imgId] for imgId in imgIds if imgId in self.imgToQA],
97
+ [],
98
+ )
99
+ else:
100
+ anns = self.dataset["annotations"]
101
+ anns = (
102
+ anns
103
+ if len(quesTypes) == 0
104
+ else [ann for ann in anns if ann["question_type"] in quesTypes]
105
+ )
106
+ anns = (
107
+ anns
108
+ if len(ansTypes) == 0
109
+ else [ann for ann in anns if ann["answer_type"] in ansTypes]
110
+ )
111
+ ids = [ann["question_id"] for ann in anns]
112
+ return ids
113
+
114
+ def getImgIds(self, quesIds=[], quesTypes=[], ansTypes=[]):
115
+ """
116
+ Get image ids that satisfy given filter conditions. default skips that filter
117
+ :param quesIds (int array) : get image ids for given question ids
118
+ quesTypes (str array) : get image ids for given question types
119
+ ansTypes (str array) : get image ids for given answer types
120
+ :return: ids (int array) : integer array of image ids
121
+ """
122
+ quesIds = quesIds if type(quesIds) == list else [quesIds]
123
+ quesTypes = quesTypes if type(quesTypes) == list else [quesTypes]
124
+ ansTypes = ansTypes if type(ansTypes) == list else [ansTypes]
125
+
126
+ if len(quesIds) == len(quesTypes) == len(ansTypes) == 0:
127
+ anns = self.dataset["annotations"]
128
+ else:
129
+ if not len(quesIds) == 0:
130
+ anns = sum(
131
+ [self.qa[quesId] for quesId in quesIds if quesId in self.qa], []
132
+ )
133
+ else:
134
+ anns = self.dataset["annotations"]
135
+ anns = (
136
+ anns
137
+ if len(quesTypes) == 0
138
+ else [ann for ann in anns if ann["question_type"] in quesTypes]
139
+ )
140
+ anns = (
141
+ anns
142
+ if len(ansTypes) == 0
143
+ else [ann for ann in anns if ann["answer_type"] in ansTypes]
144
+ )
145
+ ids = [ann["image_id"] for ann in anns]
146
+ return ids
147
+
148
+ def loadQA(self, ids=[]):
149
+ """
150
+ Load questions and answers with the specified question ids.
151
+ :param ids (int array) : integer ids specifying question ids
152
+ :return: qa (object array) : loaded qa objects
153
+ """
154
+ if type(ids) == list:
155
+ return [self.qa[id] for id in ids]
156
+ elif type(ids) == int:
157
+ return [self.qa[ids]]
158
+
159
+ def showQA(self, anns):
160
+ """
161
+ Display the specified annotations.
162
+ :param anns (array of object): annotations to display
163
+ :return: None
164
+ """
165
+ if len(anns) == 0:
166
+ return 0
167
+ for ann in anns:
168
+ quesId = ann["question_id"]
169
+ print("Question: %s" % (self.qqa[quesId]["question"]))
170
+ for ans in ann["answers"]:
171
+ print("Answer %d: %s" % (ans["answer_id"], ans["answer"]))
172
+
173
+ def loadRes(self, resFile, quesFile):
174
+ """
175
+ Load result file and return a result object.
176
+ :param resFile (str) : file name of result file
177
+ :return: res (obj) : result api object
178
+ """
179
+ res = VQA()
180
+ res.questions = json.load(open(quesFile))
181
+ res.dataset["info"] = copy.deepcopy(self.questions["info"])
182
+ res.dataset["task_type"] = copy.deepcopy(self.questions["task_type"])
183
+ res.dataset["data_type"] = copy.deepcopy(self.questions["data_type"])
184
+ res.dataset["data_subtype"] = copy.deepcopy(self.questions["data_subtype"])
185
+ res.dataset["license"] = copy.deepcopy(self.questions["license"])
186
+
187
+ print("Loading and preparing results... ")
188
+ time_t = datetime.datetime.utcnow()
189
+ anns = json.load(open(resFile))
190
+ assert type(anns) == list, "results is not an array of objects"
191
+ annsQuesIds = [ann["question_id"] for ann in anns]
192
+ assert set(annsQuesIds) == set(
193
+ self.getQuesIds()
194
+ ), "Results do not correspond to current VQA set. Either the results do not have predictions for all question ids in annotation file or there is atleast one question id that does not belong to the question ids in the annotation file."
195
+ for ann in anns:
196
+ quesId = ann["question_id"]
197
+ if res.dataset["task_type"] == "Multiple Choice":
198
+ assert (
199
+ ann["answer"] in self.qqa[quesId]["multiple_choices"]
200
+ ), "predicted answer is not one of the multiple choices"
201
+ qaAnn = self.qa[quesId]
202
+ ann["image_id"] = qaAnn["image_id"]
203
+ ann["question_type"] = qaAnn["question_type"]
204
+ ann["answer_type"] = qaAnn["answer_type"]
205
+ print(
206
+ "DONE (t=%0.2fs)" % ((datetime.datetime.utcnow() - time_t).total_seconds())
207
+ )
208
+
209
+ res.dataset["annotations"] = anns
210
+ res.createIndex()
211
+ return res
LAVIS-main/lavis/common/vqa_tools/vqa_eval.py ADDED
@@ -0,0 +1,324 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Copyright (c) 2022, salesforce.com, inc.
3
+ All rights reserved.
4
+ SPDX-License-Identifier: BSD-3-Clause
5
+ For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
6
+ """
7
+
8
+ # coding=utf-8
9
+
10
+ __author__ = "aagrawal"
11
+
12
+ # This code is based on the code written by Tsung-Yi Lin for MSCOCO Python API available at the following link:
13
+ # (https://github.com/tylin/coco-caption/blob/master/pycocoevalcap/eval.py).
14
+ import sys
15
+ import re
16
+
17
+
18
+ class VQAEval:
19
+ def __init__(self, vqa=None, vqaRes=None, n=2):
20
+ self.n = n
21
+ self.accuracy = {}
22
+ self.evalQA = {}
23
+ self.evalQuesType = {}
24
+ self.evalAnsType = {}
25
+ self.vqa = vqa
26
+ self.vqaRes = vqaRes
27
+ if vqa is not None:
28
+ self.params = {"question_id": vqa.getQuesIds()}
29
+ self.contractions = {
30
+ "aint": "ain't",
31
+ "arent": "aren't",
32
+ "cant": "can't",
33
+ "couldve": "could've",
34
+ "couldnt": "couldn't",
35
+ "couldn'tve": "couldn't've",
36
+ "couldnt've": "couldn't've",
37
+ "didnt": "didn't",
38
+ "doesnt": "doesn't",
39
+ "dont": "don't",
40
+ "hadnt": "hadn't",
41
+ "hadnt've": "hadn't've",
42
+ "hadn'tve": "hadn't've",
43
+ "hasnt": "hasn't",
44
+ "havent": "haven't",
45
+ "hed": "he'd",
46
+ "hed've": "he'd've",
47
+ "he'dve": "he'd've",
48
+ "hes": "he's",
49
+ "howd": "how'd",
50
+ "howll": "how'll",
51
+ "hows": "how's",
52
+ "Id've": "I'd've",
53
+ "I'dve": "I'd've",
54
+ "Im": "I'm",
55
+ "Ive": "I've",
56
+ "isnt": "isn't",
57
+ "itd": "it'd",
58
+ "itd've": "it'd've",
59
+ "it'dve": "it'd've",
60
+ "itll": "it'll",
61
+ "let's": "let's",
62
+ "maam": "ma'am",
63
+ "mightnt": "mightn't",
64
+ "mightnt've": "mightn't've",
65
+ "mightn'tve": "mightn't've",
66
+ "mightve": "might've",
67
+ "mustnt": "mustn't",
68
+ "mustve": "must've",
69
+ "neednt": "needn't",
70
+ "notve": "not've",
71
+ "oclock": "o'clock",
72
+ "oughtnt": "oughtn't",
73
+ "ow's'at": "'ow's'at",
74
+ "'ows'at": "'ow's'at",
75
+ "'ow'sat": "'ow's'at",
76
+ "shant": "shan't",
77
+ "shed've": "she'd've",
78
+ "she'dve": "she'd've",
79
+ "she's": "she's",
80
+ "shouldve": "should've",
81
+ "shouldnt": "shouldn't",
82
+ "shouldnt've": "shouldn't've",
83
+ "shouldn'tve": "shouldn't've",
84
+ "somebody'd": "somebodyd",
85
+ "somebodyd've": "somebody'd've",
86
+ "somebody'dve": "somebody'd've",
87
+ "somebodyll": "somebody'll",
88
+ "somebodys": "somebody's",
89
+ "someoned": "someone'd",
90
+ "someoned've": "someone'd've",
91
+ "someone'dve": "someone'd've",
92
+ "someonell": "someone'll",
93
+ "someones": "someone's",
94
+ "somethingd": "something'd",
95
+ "somethingd've": "something'd've",
96
+ "something'dve": "something'd've",
97
+ "somethingll": "something'll",
98
+ "thats": "that's",
99
+ "thered": "there'd",
100
+ "thered've": "there'd've",
101
+ "there'dve": "there'd've",
102
+ "therere": "there're",
103
+ "theres": "there's",
104
+ "theyd": "they'd",
105
+ "theyd've": "they'd've",
106
+ "they'dve": "they'd've",
107
+ "theyll": "they'll",
108
+ "theyre": "they're",
109
+ "theyve": "they've",
110
+ "twas": "'twas",
111
+ "wasnt": "wasn't",
112
+ "wed've": "we'd've",
113
+ "we'dve": "we'd've",
114
+ "weve": "we've",
115
+ "werent": "weren't",
116
+ "whatll": "what'll",
117
+ "whatre": "what're",
118
+ "whats": "what's",
119
+ "whatve": "what've",
120
+ "whens": "when's",
121
+ "whered": "where'd",
122
+ "wheres": "where's",
123
+ "whereve": "where've",
124
+ "whod": "who'd",
125
+ "whod've": "who'd've",
126
+ "who'dve": "who'd've",
127
+ "wholl": "who'll",
128
+ "whos": "who's",
129
+ "whove": "who've",
130
+ "whyll": "why'll",
131
+ "whyre": "why're",
132
+ "whys": "why's",
133
+ "wont": "won't",
134
+ "wouldve": "would've",
135
+ "wouldnt": "wouldn't",
136
+ "wouldnt've": "wouldn't've",
137
+ "wouldn'tve": "wouldn't've",
138
+ "yall": "y'all",
139
+ "yall'll": "y'all'll",
140
+ "y'allll": "y'all'll",
141
+ "yall'd've": "y'all'd've",
142
+ "y'alld've": "y'all'd've",
143
+ "y'all'dve": "y'all'd've",
144
+ "youd": "you'd",
145
+ "youd've": "you'd've",
146
+ "you'dve": "you'd've",
147
+ "youll": "you'll",
148
+ "youre": "you're",
149
+ "youve": "you've",
150
+ }
151
+ self.manualMap = {
152
+ "none": "0",
153
+ "zero": "0",
154
+ "one": "1",
155
+ "two": "2",
156
+ "three": "3",
157
+ "four": "4",
158
+ "five": "5",
159
+ "six": "6",
160
+ "seven": "7",
161
+ "eight": "8",
162
+ "nine": "9",
163
+ "ten": "10",
164
+ }
165
+ self.articles = ["a", "an", "the"]
166
+
167
+ self.periodStrip = re.compile("(?!<=\d)(\.)(?!\d)")
168
+ self.commaStrip = re.compile("(\d)(,)(\d)")
169
+ self.punct = [
170
+ ";",
171
+ r"/",
172
+ "[",
173
+ "]",
174
+ '"',
175
+ "{",
176
+ "}",
177
+ "(",
178
+ ")",
179
+ "=",
180
+ "+",
181
+ "\\",
182
+ "_",
183
+ "-",
184
+ ">",
185
+ "<",
186
+ "@",
187
+ "`",
188
+ ",",
189
+ "?",
190
+ "!",
191
+ ]
192
+
193
+ def evaluate(self, quesIds=None):
194
+ if quesIds == None:
195
+ quesIds = [quesId for quesId in self.params["question_id"]]
196
+ gts = {}
197
+ res = {}
198
+ for quesId in quesIds:
199
+ gts[quesId] = self.vqa.qa[quesId]
200
+ res[quesId] = self.vqaRes.qa[quesId]
201
+
202
+ # =================================================
203
+ # Compute accuracy
204
+ # =================================================
205
+ accQA = []
206
+ accQuesType = {}
207
+ accAnsType = {}
208
+ print("computing accuracy")
209
+ step = 0
210
+ for quesId in quesIds:
211
+ resAns = res[quesId]["answer"]
212
+ resAns = resAns.replace("\n", " ")
213
+ resAns = resAns.replace("\t", " ")
214
+ resAns = resAns.strip()
215
+ resAns = self.processPunctuation(resAns)
216
+ resAns = self.processDigitArticle(resAns)
217
+ gtAcc = []
218
+ gtAnswers = [ans["answer"] for ans in gts[quesId]["answers"]]
219
+ if len(set(gtAnswers)) > 1:
220
+ for ansDic in gts[quesId]["answers"]:
221
+ ansDic["answer"] = self.processPunctuation(ansDic["answer"])
222
+ for gtAnsDatum in gts[quesId]["answers"]:
223
+ otherGTAns = [
224
+ item for item in gts[quesId]["answers"] if item != gtAnsDatum
225
+ ]
226
+ matchingAns = [item for item in otherGTAns if item["answer"] == resAns]
227
+ acc = min(1, float(len(matchingAns)) / 3)
228
+ gtAcc.append(acc)
229
+ quesType = gts[quesId]["question_type"]
230
+ ansType = gts[quesId]["answer_type"]
231
+ avgGTAcc = float(sum(gtAcc)) / len(gtAcc)
232
+ accQA.append(avgGTAcc)
233
+ if quesType not in accQuesType:
234
+ accQuesType[quesType] = []
235
+ accQuesType[quesType].append(avgGTAcc)
236
+ if ansType not in accAnsType:
237
+ accAnsType[ansType] = []
238
+ accAnsType[ansType].append(avgGTAcc)
239
+ self.setEvalQA(quesId, avgGTAcc)
240
+ self.setEvalQuesType(quesId, quesType, avgGTAcc)
241
+ self.setEvalAnsType(quesId, ansType, avgGTAcc)
242
+ if step % 100 == 0:
243
+ self.updateProgress(step / float(len(quesIds)))
244
+ step = step + 1
245
+
246
+ self.setAccuracy(accQA, accQuesType, accAnsType)
247
+ print("Done computing accuracy")
248
+
249
+ def processPunctuation(self, inText):
250
+ outText = inText
251
+ for p in self.punct:
252
+ if (p + " " in inText or " " + p in inText) or (
253
+ re.search(self.commaStrip, inText) != None
254
+ ):
255
+ outText = outText.replace(p, "")
256
+ else:
257
+ outText = outText.replace(p, " ")
258
+ outText = self.periodStrip.sub("", outText, re.UNICODE)
259
+ return outText
260
+
261
+ def processDigitArticle(self, inText):
262
+ outText = []
263
+ tempText = inText.lower().split()
264
+ for word in tempText:
265
+ word = self.manualMap.setdefault(word, word)
266
+ if word not in self.articles:
267
+ outText.append(word)
268
+ else:
269
+ pass
270
+ for wordId, word in enumerate(outText):
271
+ if word in self.contractions:
272
+ outText[wordId] = self.contractions[word]
273
+ outText = " ".join(outText)
274
+ return outText
275
+
276
+ def setAccuracy(self, accQA, accQuesType, accAnsType):
277
+ self.accuracy["overall"] = round(100 * float(sum(accQA)) / len(accQA), self.n)
278
+ self.accuracy["perQuestionType"] = {
279
+ quesType: round(
280
+ 100 * float(sum(accQuesType[quesType])) / len(accQuesType[quesType]),
281
+ self.n,
282
+ )
283
+ for quesType in accQuesType
284
+ }
285
+ self.accuracy["perAnswerType"] = {
286
+ ansType: round(
287
+ 100 * float(sum(accAnsType[ansType])) / len(accAnsType[ansType]), self.n
288
+ )
289
+ for ansType in accAnsType
290
+ }
291
+
292
+ def setEvalQA(self, quesId, acc):
293
+ self.evalQA[quesId] = round(100 * acc, self.n)
294
+
295
+ def setEvalQuesType(self, quesId, quesType, acc):
296
+ if quesType not in self.evalQuesType:
297
+ self.evalQuesType[quesType] = {}
298
+ self.evalQuesType[quesType][quesId] = round(100 * acc, self.n)
299
+
300
+ def setEvalAnsType(self, quesId, ansType, acc):
301
+ if ansType not in self.evalAnsType:
302
+ self.evalAnsType[ansType] = {}
303
+ self.evalAnsType[ansType][quesId] = round(100 * acc, self.n)
304
+
305
+ def updateProgress(self, progress):
306
+ barLength = 20
307
+ status = ""
308
+ if isinstance(progress, int):
309
+ progress = float(progress)
310
+ if not isinstance(progress, float):
311
+ progress = 0
312
+ status = "error: progress var must be float\r\n"
313
+ if progress < 0:
314
+ progress = 0
315
+ status = "Halt...\r\n"
316
+ if progress >= 1:
317
+ progress = 1
318
+ status = "Done...\r\n"
319
+ block = int(round(barLength * progress))
320
+ text = "\rFinshed Percent: [{0}] {1}% {2}".format(
321
+ "#" * block + "-" * (barLength - block), int(progress * 100), status
322
+ )
323
+ sys.stdout.write(text)
324
+ sys.stdout.flush()
LAVIS-main/lavis/configs/datasets/aokvqa/defaults.yaml ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ aok_vqa:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url:
16
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_train.json
17
+ storage:
18
+ - aokvqa/annotations/aokvqa_v1p0_train.json
19
+ val:
20
+ url:
21
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_val.json
22
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
23
+ storage:
24
+ - aokvqa/annotations/aokvqa_v1p0_val.json
25
+ - aokvqa/annotations/specialized_vocab_train_lavis.json
26
+ # - aokvqa/annotations/large_vocab_train_lavis.json
27
+ test:
28
+ url:
29
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_test.json
30
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
31
+ storage:
32
+ - aokvqa/annotations/aokvqa_v1p0_test.json
33
+ - aokvqa/annotations/specialized_vocab_train_lavis.json
34
+ images:
35
+ storage: coco/images/
LAVIS-main/lavis/configs/datasets/aokvqa/defaults_instruct.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ aok_vqa_instruct:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: "clip_image_train"
14
+ image_size: 224
15
+ eval:
16
+ name: "clip_image_eval"
17
+ image_size: 224
18
+
19
+ text_processor:
20
+ train:
21
+ name: blip_instruction
22
+ modality: image
23
+ task: qa
24
+ eval:
25
+ name: blip_question
26
+
27
+ build_info:
28
+ # Be careful not to append minus sign (-) before split to avoid itemizing
29
+ annotations:
30
+ train:
31
+ url:
32
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_train.json
33
+ storage:
34
+ - aokvqa/annotations/aokvqa_v1p0_train.json
35
+ # val:
36
+ # url:
37
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_val.json
38
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
39
+ # storage:
40
+ # - aokvqa/annotations/aokvqa_v1p0_val.json
41
+ # - aokvqa/annotations/specialized_vocab_train_lavis.json
42
+ # # - aokvqa/annotations/large_vocab_train_lavis.json
43
+ # test:
44
+ # url:
45
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_test.json
46
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
47
+ # storage:
48
+ # - aokvqa/annotations/aokvqa_v1p0_test.json
49
+ # - aokvqa/annotations/specialized_vocab_train_lavis.json
50
+ images:
51
+ # storage: /coco/images
52
+ storage: /export/share/datasets/vision/coco/images
LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap.yaml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ audiocaps_mm_caption: # name of the dataset builder
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ sampling_rate: 16000
12
+ eval:
13
+ name: beats_audio
14
+ sampling_rate: 16000
15
+
16
+ text_processor:
17
+ train:
18
+ name: "blip_instruction"
19
+ modality: audio
20
+ task: caption
21
+ eval:
22
+ name: "blip_caption"
23
+
24
+ data_type: [audio]
25
+
26
+ build_info:
27
+ kwargs:
28
+ missing_ids: [2sh7ZkazyO8, 966jA2-z0mQ, 52RlolYyjAE, HVAc9hm4jjk, 8lPjqvYWNyM, eXgPnnE3TuQ]
29
+ annotations:
30
+ train:
31
+ url:
32
+ - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/train.csv
33
+ storage:
34
+ - audiocaps/annotations/train.csv
35
+
36
+ val:
37
+ url:
38
+ - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/val.csv
39
+ storage:
40
+ - audiocaps/annotations/val.csv
41
+
42
+ test:
43
+ url:
44
+ - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/test.csv
45
+ storage:
46
+ - audiocaps/annotations/test.csv
47
+
48
+ audio:
49
+ storage: /export/einstein-vision/audio_datasets/audiocaps/AUDIOCAPS_32000Hz/audio
LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap_instruct.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ audiocaps_mm_caption_instruct: # name of the dataset builder
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ sampling_rate: 16000
12
+ eval:
13
+ name: beats_audio
14
+ sampling_rate: 16000
15
+
16
+ text_processor:
17
+ train:
18
+ name: "blip_instruction"
19
+ modality: audio
20
+ task: caption
21
+ eval:
22
+ name: "blip_caption"
23
+
24
+ data_type: [audio]
25
+
26
+ missing_ids: [2sh7ZkazyO8, 966jA2-z0mQ, 52RlolYyjAE, HVAc9hm4jjk, 8lPjqvYWNyM, eXgPnnE3TuQ]
27
+
28
+ build_info:
29
+ kwargs:
30
+ cached: False
31
+ cached_dir: /export/einstein-vision/audio_datasets/audiocaps/beats_features
32
+ annotations:
33
+ train:
34
+ url:
35
+ - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/train.csv
36
+ storage:
37
+ - audiocaps/annotations/train.csv
38
+
39
+ # val:
40
+ # url:
41
+ # - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/val.csv
42
+ # storage:
43
+ # - audiocaps/annotation/val.csv
44
+
45
+ # test:
46
+ # url:
47
+ # - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/test.csv
48
+ # storage:
49
+ # - /export/einstein-vision/audio_datasets/audiocaps/dataset/test.csv
50
+
51
+ audio:
52
+ storage: /export/einstein-vision/audio_datasets/audiocaps/AUDIOCAPS_32000Hz/audio
LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_qa.yaml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ audiocaps_mm_qa: # name of the dataset builder
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ sampling_rate: 16000
12
+ eval:
13
+ name: beats_audio
14
+ sampling_rate: 16000
15
+ is_eval: True
16
+
17
+ text_processor:
18
+ train:
19
+ name: "blip_instruction"
20
+ modality: audio
21
+ task: qa
22
+ eval:
23
+ name: "blip_question"
24
+
25
+ data_type: [audio]
26
+
27
+ build_info:
28
+ kwargs:
29
+ cached: False
30
+ # add_binary: True
31
+ cached_dir: /export/einstein-vision/audio_datasets/audiocaps/beats_features
32
+ missing_ids: [2sh7ZkazyO8, 966jA2-z0mQ, 52RlolYyjAE, HVAc9hm4jjk, 8lPjqvYWNyM, eXgPnnE3TuQ]
33
+ annotations:
34
+ train:
35
+ url:
36
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/audiocaps/audio_qa_final_train.csv
37
+ # - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_train.csv
38
+ storage:
39
+ - audiocaps_qa/annotations/train.csv
40
+ # - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_train.csv
41
+
42
+ # val:
43
+ # url:
44
+ # # - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/audiocaps/audio_qa_final_val.csv
45
+ # - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_val.csv
46
+ # storage:
47
+ # # - audiocaps_qa/annotations/val.csv
48
+ # - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_val.csv
49
+
50
+ audio:
51
+ storage: /export/einstein-vision/audio_datasets/audiocaps/AUDIOCAPS_32000Hz/audio
LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap.yaml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ audioset_mm_caption: # 14141
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ sampling_rate: 16000
12
+ eval:
13
+ name: beats_audio
14
+ sampling_rate: 16000
15
+ is_eval: False
16
+
17
+ text_processor:
18
+ train:
19
+ name: blip_instruction
20
+ modality: audio
21
+ task: classification
22
+ eval:
23
+ name: blip_caption
24
+
25
+ data_type: [audio]
26
+
27
+ build_info:
28
+ annotations:
29
+ train:
30
+ url:
31
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data//audioset/balanced_train_clean.csv
32
+ # - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
33
+ - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
34
+ storage:
35
+ - audioset/balanced_train_clean.csv
36
+ # - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
37
+ - audioset/annotations/class_labels_indices.csv
38
+
39
+ # val:
40
+ # url:
41
+ # - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/eval_segments.csv
42
+ # - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
43
+ # storage:
44
+ # - audioset/annotations/eval_segments.csv
45
+ # - audioset/annotations/class_labels_indices.csv
46
+ audio:
47
+ storage: /export/einstein-vision/audio_datasets/AudioSet/all_audio
LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap_instruct.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ audioset_mm_caption_instruct: # 14141
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ sampling_rate: 16000
12
+ eval:
13
+ name: beats_audio
14
+ sampling_rate: 16000
15
+ is_eval: False
16
+
17
+ text_processor:
18
+ train:
19
+ name: blip_instruction
20
+ modality: audio
21
+ task: classification
22
+ eval:
23
+ name: blip_caption
24
+
25
+ data_type: [audio]
26
+
27
+ build_info:
28
+ annotations:
29
+ train:
30
+ url:
31
+ # - https://storage.googleapis.com/sfr-xinstructblip-data-research/data//audioset/balanced_train_clean.csv
32
+ - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
33
+ - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
34
+ storage:
35
+ - audioset/annotations/balanced_train_clean.csv
36
+ # - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
37
+ - audioset/annotations/class_labels_indices.csv
38
+
39
+ # val:
40
+ # url:
41
+ # - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/eval_segments.csv
42
+ # - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
43
+ # storage:
44
+ # - audioset/annotations/eval_segments.csv
45
+ # - audioset/annotations/class_labels_indices.csv
46
+
47
+ audio:
48
+ storage: /export/einstein-vision/audio_datasets/AudioSet/all_audio
LAVIS-main/lavis/configs/datasets/avsd/defaults_dial.yaml ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ avsd_dialogue: # name of the dataset builder
8
+ dataset_card: dataset_card/avsd_dialogue.md
9
+ data_type: features #extracted features of videos (I3D, VGGish) # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_train.json
16
+ storage: avsd/annotations/train.json
17
+ val:
18
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_val.json
19
+ storage: avsd/annotations/val.json
20
+ test:
21
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_test.json
22
+ storage: avsd/annotations/test.json
23
+ features:
24
+ storage: avsd/features/
LAVIS-main/lavis/configs/datasets/avsd/defaults_mm_dial_instruct.yaml ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ avsd_mm_dialogue_instruct: # name of the dataset builder
8
+ data_type: [video, audio]
9
+
10
+ audio_processor:
11
+ train:
12
+ name: beats_audio
13
+ sampling_rate: 16000
14
+ eval:
15
+ name: beats_audio
16
+ sampling_rate: 16000
17
+
18
+ video_processor:
19
+ train:
20
+ name: alpro_video_train
21
+ n_frms: 4
22
+ image_size: 224
23
+ min_scale: 0.9
24
+ max_scale: 1.0
25
+ full_video: True
26
+ eval:
27
+ name: alpro_video_eval
28
+ n_frms: 4
29
+ image_size: 224
30
+ min_scale: 0.9
31
+ max_scale: 1.0
32
+ full_video: True
33
+
34
+ text_processor:
35
+ train:
36
+ name: "blip_caption"
37
+ eval:
38
+ name: "blip_caption"
39
+
40
+ build_info:
41
+ # Be careful not to append minus sign (-) before split to avoid itemizing
42
+ annotations:
43
+ train:
44
+ url:
45
+ - https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_train.json
46
+ storage:
47
+ - avsd/annotations/train.json
48
+ val:
49
+ url:
50
+ - https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_val.json
51
+ storage:
52
+ - avsd/annotations/val.json
53
+ test:
54
+ url:
55
+ - https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_test.json
56
+ storage:
57
+ - avsd/annotations/test.json
58
+ templates: null
59
+
60
+ audio:
61
+ storage: /export/video-language-dataset/data/charade/videos
62
+
63
+ video:
64
+ storage: /export/video-language-dataset/data/charade/videos
65
+
LAVIS-main/lavis/configs/datasets/blip_diffusion_datasets/defaults.yaml ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ blip_diffusion_finetune: # name of the dataset builder
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ images:
14
+ storage: ""
LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap.yaml ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ capfilt14m: # 13873136
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: "clip_image_train"
14
+ image_size: 224
15
+ text_processor:
16
+ train:
17
+ name: blip_caption
18
+
19
+ build_info:
20
+ # Be careful not to append minus sign (-) before split to avoid itemizing
21
+ annotations:
22
+ train:
23
+ url:
24
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/capfilt14m/annotation.json
25
+ # - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
26
+ storage:
27
+ - capfilt14m/annotations/annotation.json
28
+ # - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
29
+ images:
30
+ storage: /export/share/datasets/vision/coco/images
LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap_instruct.yaml ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ capfilt14m_instruct: # 13873136
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: "clip_image_train"
14
+ image_size: 224
15
+
16
+ text_processor:
17
+ train:
18
+ name: blip_instruction
19
+ modality: image
20
+ task: caption
21
+
22
+ build_info:
23
+ # Be careful not to append minus sign (-) before split to avoid itemizing
24
+ annotations:
25
+ train:
26
+ url:
27
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/capfilt14m/annotation.json
28
+ # - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
29
+ storage:
30
+ - capfilt14m/annotations/annotation.json
31
+ # - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
32
+
33
+ images:
34
+ storage: /export/share/datasets/vision/coco/images
LAVIS-main/lavis/configs/datasets/charade/defaults_cap.yaml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ charade_caption: # name of the dataset builder
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: videos # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: alpro_video_train
14
+ n_frms: 4
15
+ image_size: 224
16
+ min_scale: 0.9
17
+ max_scale: 1.0
18
+ full_video: False
19
+ eval:
20
+ name: alpro_video_eval
21
+ n_frms: 4
22
+ image_size: 224
23
+ min_scale: 0.9
24
+ max_scale: 1.0
25
+ full_video: False
26
+
27
+ text_processor:
28
+ train:
29
+ name: blip_caption
30
+ eval:
31
+ name: blip_caption
32
+
33
+
34
+ build_info:
35
+ # Be careful not to append minus sign (-) before split to avoid itemizing
36
+ annotations:
37
+ train:
38
+ url:
39
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/train_lavis.json
40
+ # - /export/video-language-dataset/data/charade/train_lavis.json
41
+ storage:
42
+ - charade/annotations/train.json
43
+ # - /export/video-language-dataset/data/charade/train_lavis.json
44
+ val:
45
+ url:
46
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/val_lavis.json
47
+ # - /export/video-language-dataset/data/charade/val_lavis.json
48
+ storage:
49
+ - charade/annotations/val.json
50
+ # - /export/video-language-dataset/data/charade/val_lavis.json
51
+ videos:
52
+ storage: /export/video-language-dataset/data/charade/videos
LAVIS-main/lavis/configs/datasets/charade/defaults_cap_instruct.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ charade_caption_instruct: # name of the dataset builder
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: videos # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: alpro_video_train
14
+ n_frms: 4
15
+ image_size: 224
16
+ min_scale: 0.9
17
+ max_scale: 1.0
18
+ full_video: False
19
+ eval:
20
+ name: alpro_video_eval
21
+ n_frms: 4
22
+ image_size: 224
23
+ min_scale: 0.9
24
+ max_scale: 1.0
25
+ full_video: False
26
+
27
+
28
+ text_processor:
29
+ train:
30
+ name: blip_instruction
31
+ modality: video
32
+ task: caption
33
+ eval:
34
+ name: blip_caption
35
+
36
+ build_info:
37
+ # Be careful not to append minus sign (-) before split to avoid itemizing
38
+ annotations:
39
+ train:
40
+ url:
41
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/train_lavis.json
42
+ # - /export/video-language-dataset/data/charade/train_lavis.json
43
+ storage:
44
+ - charade/annotations/train.json
45
+ # - /export/video-language-dataset/data/charade/train_lavis.json
46
+ val:
47
+ url:
48
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/val_lavis.json
49
+ # - /export/video-language-dataset/data/charade/val_lavis.json
50
+ storage:
51
+ - charade/annotations/val.json
52
+ # - /export/video-language-dataset/data/charade/val_lavis.json
53
+ videos:
54
+ storage: /export/video-language-dataset/data/charade/videos
LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap.yaml ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ clothov2: # name of the dataset builder
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ eval:
12
+ name: beats_audio
13
+
14
+ text_processor:
15
+ train:
16
+ name: "blip_caption"
17
+ eval:
18
+ name: "blip_caption"
19
+
20
+
21
+ data_type: [audio]
22
+
23
+ build_info:
24
+ kwargs:
25
+ clotho_root: /export/einstein-vision/audio_datasets/clothov2/
26
+ split: eval
27
+
28
+ annotations:
29
+ train:
30
+ url:
31
+ - https://zenodo.org/record/4783391/files/clotho_captions_development.csv
32
+ storage:
33
+ - clothov2/annotations/clotho_captions_development.csv
34
+ val:
35
+ url:
36
+ - https://zenodo.org/record/4783391/files/clotho_captions_evaluation.csv
37
+ storage:
38
+ - clothov2/annotations/clotho_captions_evaluation.csv
39
+ audio:
40
+ storage: /export/einstein-vision/audio_datasets/clothov2/CLOTHO_v2.1/clotho_audio_files/
41
+
LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap_instruct.yaml ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ clothov2_instruct: # name of the dataset builder
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ eval:
12
+ name: beats_audio
13
+
14
+ text_processor:
15
+ train:
16
+ name: "blip_instruction"
17
+ modality: audio
18
+ task: caption
19
+ eval:
20
+ name: "blip_caption"
21
+
22
+ data_type: [audio]
23
+
24
+ build_info:
25
+ kwargs:
26
+ clotho_root: /export/einstein-vision/audio_datasets/clothov2/
27
+ split: eval
28
+
29
+ annotations:
30
+ train:
31
+ url:
32
+ - https://zenodo.org/record/4783391/files/clotho_captions_development.csv
33
+ storage:
34
+ - clothov2/annotations/clotho_captions_development.csv
35
+ val:
36
+ url:
37
+ - https://zenodo.org/record/4783391/files/clotho_captions_evaluation.csv
38
+ storage:
39
+ - clothov2/annotations/clotho_captions_evaluation.csv
40
+ audio:
41
+ storage: /export/einstein-vision/audio_datasets/clothov2/CLOTHO_v2.1/clotho_audio_files/
42
+
LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_qa.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ clotho_qa: # name of the dataset builder
8
+ audio_processor:
9
+ train:
10
+ name: beats_audio
11
+ eval:
12
+ name: beats_audio
13
+
14
+ text_processor:
15
+ train:
16
+ name: "blip_caption"
17
+ eval:
18
+ name: "blip_caption"
19
+
20
+
21
+ data_type: [audio]
22
+
23
+ build_info:
24
+
25
+ annotations:
26
+ train:
27
+ url:
28
+ - https://zenodo.org/records/6473207/files/clotho_aqa_train.csv
29
+ storage:
30
+ - clotho_Qa/annotations/clotho_aqa_train.csv
31
+ val:
32
+ url:
33
+ - https://zenodo.org/records/6473207/files/clotho_aqa_val.csv
34
+ storage:
35
+ - clotho_qa/annotations/clotho_aqa_val.csv
36
+
37
+ test:
38
+ url:
39
+ - https://zenodo.org/records/6473207/files/clotho_aqa_test.csv
40
+ storage:
41
+ - clotho_qa/annotations/clotho_aqa_test.csv
42
+ audio:
43
+ storage: /export/einstein-vision/audio_datasets/clotho-aqa/audio_files
44
+
LAVIS-main/lavis/configs/datasets/coco/defaults_cap.yaml ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coco_caption: # name of the dataset builder
8
+ dataset_card: dataset_card/coco_caption.md
9
+ # data_dir: ${env.data_dir}/datasets
10
+ data_type: images # [images|videos|features]
11
+
12
+ build_info:
13
+ # Be careful not to append minus sign (-) before split to avoid itemizing
14
+ annotations:
15
+ train:
16
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_train.json
17
+ md5: aa31ac474cf6250ebb81d18348a07ed8
18
+ storage: coco/annotations/coco_karpathy_train.json
19
+ val:
20
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_val.json
21
+ md5: b273847456ef5580e33713b1f7de52a0
22
+ storage: coco/annotations/coco_karpathy_val.json
23
+ test:
24
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_test.json
25
+ md5: 3ff34b0ef2db02d01c37399f6a2a6cd1
26
+ storage: coco/annotations/coco_karpathy_test.json
27
+ images:
28
+ storage: coco/images/
LAVIS-main/lavis/configs/datasets/coco/defaults_cap_instruct.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coco_caption_instruct: # name of the dataset builder
8
+ dataset_card: dataset_card/coco_caption.md
9
+ # data_dir: ${env.data_dir}/datasets
10
+ data_type: images # [images|videos|features]
11
+
12
+ vis_processor:
13
+ train:
14
+ name: "clip_image_train"
15
+ image_size: 224
16
+ eval:
17
+ name: "clip_image_eval"
18
+ image_size: 224
19
+
20
+ text_processor:
21
+ train:
22
+ name: blip_instruction
23
+ modality: image
24
+ task: caption
25
+ eval:
26
+ name: blip_caption
27
+
28
+ build_info:
29
+ # Be careful not to append minus sign (-) before split to avoid itemizing
30
+ annotations:
31
+ train:
32
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_train.json
33
+ md5: aa31ac474cf6250ebb81d18348a07ed8
34
+ storage: coco/annotations/coco_karpathy_train.json
35
+ # val:
36
+ # url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_val.json
37
+ # md5: b273847456ef5580e33713b1f7de52a0
38
+ # storage: coco/annotations/coco_karpathy_val.json
39
+ # test:
40
+ # url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_test.json
41
+ # md5: 3ff34b0ef2db02d01c37399f6a2a6cd1
42
+ # storage: coco/annotations/coco_karpathy_test.json
43
+ images:
44
+ storage: /export/share/datasets/vision/coco/images
LAVIS-main/lavis/configs/datasets/coco/defaults_ret.yaml ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coco_retrieval:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_train.json
16
+ md5: aa31ac474cf6250ebb81d18348a07ed8
17
+ storage: coco/annotations/coco_karpathy_train.json
18
+ val:
19
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_val.json
20
+ md5: b273847456ef5580e33713b1f7de52a0
21
+ storage: coco/annotations/coco_karpathy_val.json
22
+ test:
23
+ url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_test.json
24
+ md5: 3ff34b0ef2db02d01c37399f6a2a6cd1
25
+ storage: coco/annotations/coco_karpathy_test.json
26
+ images:
27
+ storage: coco/images/
LAVIS-main/lavis/configs/datasets/coco/defaults_vqa.yaml ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coco_vqa:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url:
16
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_train.json
17
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val.json
18
+ storage:
19
+ - coco/annotations/vqa_train.json
20
+ - coco/annotations/vqa_val.json
21
+ val:
22
+ url:
23
+ # TODO make this order insensitive
24
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val_eval.json
25
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
26
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_OpenEnded_mscoco_val2014_questions.json
27
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_mscoco_val2014_annotations.json
28
+ storage:
29
+ - coco/annotations/vqa_val_eval.json
30
+ - coco/annotations/answer_list.json
31
+ - coco/annotations/v2_OpenEnded_mscoco_val2014_questions.json
32
+ - coco/annotations/v2_mscoco_val2014_annotations.json
33
+ test:
34
+ url:
35
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_test.json
36
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
37
+ storage:
38
+ - coco/annotations/vqa_test.json
39
+ - coco/annotations/answer_list.json
40
+ images:
41
+ storage: coco/images/
LAVIS-main/lavis/configs/datasets/coco/defaults_vqa_instruct.yaml ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coco_vqa_instruct:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: "clip_image_train"
14
+ image_size: 224
15
+ eval:
16
+ name: "clip_image_eval"
17
+ image_size: 224
18
+
19
+ text_processor:
20
+ train:
21
+ name: blip_instruction
22
+ modality: image
23
+ task: qa
24
+ eval:
25
+ name: blip_caption
26
+
27
+ build_info:
28
+ # Be careful not to append minus sign (-) before split to avoid itemizing
29
+ annotations:
30
+ train:
31
+ url:
32
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_train.json
33
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val.json
34
+ storage:
35
+ - coco/annotations/vqa_train.json
36
+ - coco/annotations/vqa_val.json
37
+ # val:
38
+ # url:
39
+ # # TODO make this order insensitive
40
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val_eval.json
41
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
42
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_OpenEnded_mscoco_val2014_questions.json
43
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_mscoco_val2014_annotations.json
44
+ # storage:
45
+ # - coco/annotations/vqa_val_eval.json
46
+ # - coco/annotations/answer_list.json
47
+ # - coco/annotations/v2_OpenEnded_mscoco_val2014_questions.json
48
+ # - coco/annotations/v2_mscoco_val2014_annotations.json
49
+ # test:
50
+ # url:
51
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_test.json
52
+ # - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
53
+ # storage:
54
+ # - coco/annotations/vqa_test.json
55
+ # - coco/annotations/answer_list.json
56
+ images:
57
+ storage: /export/share/datasets/vision/coco/images
LAVIS-main/lavis/configs/datasets/coco/eval_vqa.yaml ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coco_vqa:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ val:
15
+ url:
16
+ # TODO make this order insensitive
17
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val_eval.json
18
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
19
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_OpenEnded_mscoco_val2014_questions.json
20
+ - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_mscoco_val2014_annotations.json
21
+ storage:
22
+ - coco/annotations/vqa_val_eval.json
23
+ - coco/annotations/answer_list.json
24
+ - coco/annotations/v2_OpenEnded_mscoco_val2014_questions.json
25
+ - coco/annotations/v2_mscoco_val2014_annotations.json
26
+ images:
27
+ storage: coco/images/
LAVIS-main/lavis/configs/datasets/coin/defaults_cap.yaml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coin_caption: # name of the dataset builder
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: videos # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: alpro_video_train
14
+ n_frms: 4
15
+ image_size: 224
16
+ min_scale: 0.9
17
+ max_scale: 1.0
18
+ full_video: False
19
+ eval:
20
+ name: alpro_video_eval
21
+ n_frms: 4
22
+ image_size: 224
23
+ min_scale: 0.9
24
+ max_scale: 1.0
25
+ full_video: False
26
+
27
+ text_processor:
28
+ train:
29
+ name: blip_caption
30
+ eval:
31
+ name: blip_caption
32
+
33
+ build_info:
34
+ # Be careful not to append minus sign (-) before split to avoid itemizing
35
+ annotations:
36
+ train:
37
+ url:
38
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/train.json
39
+ # - /export/video-language-dataset/data/coin/annotations/train_lavis.json
40
+ storage:
41
+ - coin/annotations/train.json
42
+ # - /export/video-language-dataset/data/coin/annotations/train_lavis.json
43
+ val:
44
+ url:
45
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/val.json
46
+ # - /export/video-language-dataset/data/coin/annotations/val_lavis.json
47
+ storage:
48
+ - coin/annotations/val.json
49
+ # - /export/video-language-dataset/data/coin/annotations/val_lavis.json
50
+ videos:
51
+ storage: /export/video-language-dataset/data/coin/annotations/videos/
LAVIS-main/lavis/configs/datasets/coin/defaults_cap_instruct.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ coin_caption_instruct: # name of the dataset builder
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: videos # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: alpro_video_train
14
+ n_frms: 4
15
+ image_size: 224
16
+ min_scale: 0.9
17
+ max_scale: 1.0
18
+ full_video: False
19
+ eval:
20
+ name: alpro_video_eval
21
+ n_frms: 4
22
+ image_size: 224
23
+ min_scale: 0.9
24
+ max_scale: 1.0
25
+ full_video: False
26
+
27
+ text_processor:
28
+ train:
29
+ name: blip_instruction
30
+ task: caption
31
+ modality: image
32
+ eval:
33
+ name: blip_caption
34
+
35
+ build_info:
36
+ # Be careful not to append minus sign (-) before split to avoid itemizing
37
+ annotations:
38
+ train:
39
+ url:
40
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/train.json
41
+ # - /export/video-language-dataset/data/coin/annotations/train_lavis.json
42
+ storage:
43
+ - coin/annotations/train.json
44
+ # - /export/video-language-dataset/data/coin/annotations/train_lavis.json
45
+ val:
46
+ url:
47
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/val.json
48
+ # - /export/video-language-dataset/data/coin/annotations/val_lavis.json
49
+ storage:
50
+ - coin/annotations/val.json
51
+ # - /export/video-language-dataset/data/coin/annotations/val_lavis.json
52
+ videos:
53
+ storage: /export/video-language-dataset/data/coin/annotations/videos/
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m.yaml ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ conceptual_caption_12m:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url:
16
+ - /export/home/workspace/datasets/cc12m.json
17
+ storage:
18
+ - conceptual_caption/annotations/cc12m.json
19
+ images:
20
+ storage: conceptual_caption/images_12m
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m_instruct.yaml ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ conceptual_caption_12m_instruct:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: "clip_image_train"
14
+ image_size: 224
15
+ eval:
16
+ name: "clip_image_eval"
17
+ image_size: 224
18
+
19
+ text_processor:
20
+ train:
21
+ name: blip_instruction
22
+ task: caption
23
+ modality: image
24
+ eval:
25
+ name: blip_caption
26
+
27
+ build_info:
28
+ # Be careful not to append minus sign (-) before split to avoid itemizing
29
+ annotations:
30
+ train:
31
+ url:
32
+ - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/cc12m/x_instructblip_clean.json
33
+ # - /export/home/workspace/datasets/cc12m.json
34
+ storage:
35
+ - conceptual_caption/annotations/cc12m.json
36
+ images:
37
+ storage: conceptual_caption/images_12m
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m.yaml ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ conceptual_caption_3m:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url:
16
+ - /export/home/workspace/datasets/cc3m.json
17
+ storage:
18
+ - conceptual_caption/annotations/cc3m.json
19
+ images:
20
+ storage: conceptual_caption/images
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m_instruct.yaml ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2023, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ conceptual_caption_3m_instruct:
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: images # [images|videos|features]
10
+
11
+ vis_processor:
12
+ train:
13
+ name: "clip_image_train"
14
+ image_size: 224
15
+ eval:
16
+ name: "clip_image_eval"
17
+ image_size: 224
18
+
19
+ text_processor:
20
+ train:
21
+ name: blip_instruction
22
+ task: caption
23
+ modality: image
24
+ eval:
25
+ name: blip_caption
26
+
27
+ build_info:
28
+ # Be careful not to append minus sign (-) before split to avoid itemizing
29
+ annotations:
30
+ train:
31
+ url:
32
+ - /export/home/workspace/datasets/cc3m.json
33
+ storage:
34
+ - conceptual_caption/annotations/cc3m.json
35
+ images:
36
+ storage: conceptual_caption/images
LAVIS-main/lavis/configs/datasets/didemo/defaults_ret.yaml ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) 2022, salesforce.com, inc.
2
+ # All rights reserved.
3
+ # SPDX-License-Identifier: BSD-3-Clause
4
+ # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
5
+
6
+ datasets:
7
+ didemo_retrieval: # name of the dataset builder
8
+ # data_dir: ${env.data_dir}/datasets
9
+ data_type: videos # [images|videos|features]
10
+
11
+ build_info:
12
+ # Be careful not to append minus sign (-) before split to avoid itemizing
13
+ annotations:
14
+ train:
15
+ url: https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/didemo/retrieval_train.json
16
+ storage: didemo/annotations/retrieval_train.json
17
+ val:
18
+ url: https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/didemo/retrieval_val.json
19
+ storage: didemo/annotations/retrieval_val.json
20
+ test:
21
+ url: https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/didemo/retrieval_test.json
22
+ storage: didemo/annotations/retrieval_test.json
23
+ videos:
24
+ storage: didemo/videos
25
+ # storage: /export/share/dongxuli/data/didemo_retrieval/videos