Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- BIO/ablation/temperature_stability.jsonl +0 -0
- BIO/ablation/test.py +157 -0
- BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/args.json +364 -0
- BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/logging.jsonl +0 -0
- BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug-internal.log +23 -0
- BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug.log +22 -0
- BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/config.yaml +200 -0
- BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/output.log +16 -0
- BioReason_new/wandb/run-20250812_132050-2cfwmlj6/run-2cfwmlj6.wandb +0 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/up_conv_block.py +101 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/weight_init.py +62 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/__init__.py +4 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/encoding.py +74 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/wrappers.py +50 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/__init__.py +4 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/collect_env.py +17 -0
- LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/logger.py +27 -0
- LAVIS-main/lavis/common/vqa_tools/__init__.py +8 -0
- LAVIS-main/lavis/common/vqa_tools/vqa.py +211 -0
- LAVIS-main/lavis/common/vqa_tools/vqa_eval.py +324 -0
- LAVIS-main/lavis/configs/datasets/aokvqa/defaults.yaml +35 -0
- LAVIS-main/lavis/configs/datasets/aokvqa/defaults_instruct.yaml +52 -0
- LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap.yaml +49 -0
- LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap_instruct.yaml +52 -0
- LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_qa.yaml +51 -0
- LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap.yaml +47 -0
- LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap_instruct.yaml +48 -0
- LAVIS-main/lavis/configs/datasets/avsd/defaults_dial.yaml +24 -0
- LAVIS-main/lavis/configs/datasets/avsd/defaults_mm_dial_instruct.yaml +65 -0
- LAVIS-main/lavis/configs/datasets/blip_diffusion_datasets/defaults.yaml +14 -0
- LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap.yaml +30 -0
- LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap_instruct.yaml +34 -0
- LAVIS-main/lavis/configs/datasets/charade/defaults_cap.yaml +52 -0
- LAVIS-main/lavis/configs/datasets/charade/defaults_cap_instruct.yaml +54 -0
- LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap.yaml +41 -0
- LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap_instruct.yaml +42 -0
- LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_qa.yaml +44 -0
- LAVIS-main/lavis/configs/datasets/coco/defaults_cap.yaml +28 -0
- LAVIS-main/lavis/configs/datasets/coco/defaults_cap_instruct.yaml +44 -0
- LAVIS-main/lavis/configs/datasets/coco/defaults_ret.yaml +27 -0
- LAVIS-main/lavis/configs/datasets/coco/defaults_vqa.yaml +41 -0
- LAVIS-main/lavis/configs/datasets/coco/defaults_vqa_instruct.yaml +57 -0
- LAVIS-main/lavis/configs/datasets/coco/eval_vqa.yaml +27 -0
- LAVIS-main/lavis/configs/datasets/coin/defaults_cap.yaml +51 -0
- LAVIS-main/lavis/configs/datasets/coin/defaults_cap_instruct.yaml +53 -0
- LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m.yaml +20 -0
- LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m_instruct.yaml +37 -0
- LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m.yaml +20 -0
- LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m_instruct.yaml +36 -0
- LAVIS-main/lavis/configs/datasets/didemo/defaults_ret.yaml +25 -0
BIO/ablation/temperature_stability.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
BIO/ablation/test.py
ADDED
|
@@ -0,0 +1,157 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
|
| 2 |
+
# #accuracy
|
| 3 |
+
import json
|
| 4 |
+
import re
|
| 5 |
+
|
| 6 |
+
def compute_accuracy_from_file(filepath):
|
| 7 |
+
total, correct = 0, 0
|
| 8 |
+
pattern = re.compile(r"<ANSWER>(.*?)</ANSWER>", re.IGNORECASE)
|
| 9 |
+
|
| 10 |
+
with open(filepath, 'r', encoding='utf-8') as f:
|
| 11 |
+
for line in f:
|
| 12 |
+
line = line.strip()
|
| 13 |
+
if not line:
|
| 14 |
+
continue
|
| 15 |
+
data = json.loads(line)
|
| 16 |
+
print(data)
|
| 17 |
+
tgt_match = data.get("reference_answer", "")
|
| 18 |
+
print(tgt_match)
|
| 19 |
+
pred_match = pattern.search(data.get("generated_answer", ""))
|
| 20 |
+
#pred_match = data.get("generated_answer", "")
|
| 21 |
+
print(pred_match)
|
| 22 |
+
if not tgt_match or not pred_match:
|
| 23 |
+
continue
|
| 24 |
+
|
| 25 |
+
try:
|
| 26 |
+
tgt_val = int(tgt_match.strip())
|
| 27 |
+
pred_val = int(pred_match.group(1).strip())
|
| 28 |
+
#.group(1).
|
| 29 |
+
except ValueError:
|
| 30 |
+
continue # 如果无法转换为int,跳过此条
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
total += 1
|
| 34 |
+
if pred_val == tgt_val:
|
| 35 |
+
correct += 1
|
| 36 |
+
|
| 37 |
+
if total == 0:
|
| 38 |
+
return 0.0, 0
|
| 39 |
+
return correct / total, total
|
| 40 |
+
|
| 41 |
+
if __name__ == "__main__":
|
| 42 |
+
filepath = "/nas/shared/kilab/wangyujia/BIO/ablation/temperature_stability.jsonl"
|
| 43 |
+
acc, count = compute_accuracy_from_file(filepath)
|
| 44 |
+
print(f"Checked {count} items. Accuracy: {acc*100:.3f}%")
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
# #spearman
|
| 48 |
+
# import json
|
| 49 |
+
# import re
|
| 50 |
+
# from scipy.stats import spearmanr
|
| 51 |
+
|
| 52 |
+
# # 文件路径(替换为你的 JSONL 文件路径)
|
| 53 |
+
# file_path = '/nas/shared/kilab/wangyujia/BIO/ablation/temperature_stability.jsonl'
|
| 54 |
+
|
| 55 |
+
# # 正则:提取 <answer>...</answer> 中的数值
|
| 56 |
+
# pattern = re.compile(r"<answer>(.*?)</answer>")
|
| 57 |
+
|
| 58 |
+
# y_true = [] # reference_answer 中的真实值
|
| 59 |
+
# y_pred = [] # generated_answer 预测值
|
| 60 |
+
|
| 61 |
+
# with open(file_path, 'r', encoding='utf-8') as f:
|
| 62 |
+
# for line in f:
|
| 63 |
+
# line = line.strip()
|
| 64 |
+
# print(line)
|
| 65 |
+
# if not line:
|
| 66 |
+
# continue
|
| 67 |
+
# try:
|
| 68 |
+
# data = json.loads(line) # 每行都是一个 JSON 对象
|
| 69 |
+
# reference_str = data.get("reference_answer", "").strip()
|
| 70 |
+
# generated_str = data.get("generated_answer", "").strip()
|
| 71 |
+
# print(reference_str)
|
| 72 |
+
# print(generated_str)
|
| 73 |
+
# # 提取 generated_answer 中的数值
|
| 74 |
+
# pred_match = pattern.search(generated_str)
|
| 75 |
+
# print(pred_match.group(1))
|
| 76 |
+
# if pred_match:
|
| 77 |
+
# pred_value = float(pred_match.group(1)) # 提取 <answer> 中的值
|
| 78 |
+
# true_value = float(reference_str) # reference_answer 本身就是数值字符串
|
| 79 |
+
# y_true.append(true_value)
|
| 80 |
+
# y_pred.append(pred_value)
|
| 81 |
+
# else:
|
| 82 |
+
# print(f"未找到 <answer> 标签,跳过:{generated_str}")
|
| 83 |
+
# except Exception as e:
|
| 84 |
+
# print(f"处理行时出错:{line}")
|
| 85 |
+
# print(f"错误:{e}")
|
| 86 |
+
# continue
|
| 87 |
+
|
| 88 |
+
# # 计算 Spearman 相关系数
|
| 89 |
+
# if len(y_true) > 1:
|
| 90 |
+
# print(len(y_true))
|
| 91 |
+
# print(len(y_pred))
|
| 92 |
+
# rho, p_value = spearmanr(y_pred,y_true)
|
| 93 |
+
# print(f"有效样本数:{len(y_true)}")
|
| 94 |
+
# print(f"Spearman correlation coefficient: {rho:.5f}, p-value: {p_value:.4e}")
|
| 95 |
+
# else:
|
| 96 |
+
# print("有效数据不足,无法计算 Spearman 相关系数。")
|
| 97 |
+
|
| 98 |
+
|
| 99 |
+
# # f1
|
| 100 |
+
# import re
|
| 101 |
+
# import json
|
| 102 |
+
|
| 103 |
+
# def extract_numbers_from_generated_answer(s):
|
| 104 |
+
# """从<answer>...</answer>中提取纯数字"""
|
| 105 |
+
# match = re.search(r'<answer>(.*?)</answer>', s)
|
| 106 |
+
# if match:
|
| 107 |
+
# numbers_str = match.group(1)
|
| 108 |
+
# numbers = re.findall(r'\d+', numbers_str)
|
| 109 |
+
# return set(map(int, numbers))
|
| 110 |
+
# else:
|
| 111 |
+
# return set()
|
| 112 |
+
|
| 113 |
+
# def extract_numbers_from_reference_answer(s):
|
| 114 |
+
# """从参考答案字符串中提取数字,忽略非数字字符"""
|
| 115 |
+
# numbers = re.findall(r'\d+', s)
|
| 116 |
+
# return set(map(int, numbers))
|
| 117 |
+
|
| 118 |
+
# def calculate_f1(pred_set, target_set):
|
| 119 |
+
# if not pred_set and not target_set:
|
| 120 |
+
# return 1.0
|
| 121 |
+
# if not pred_set or not target_set:
|
| 122 |
+
# return 0.0
|
| 123 |
+
# tp = len(pred_set & target_set)
|
| 124 |
+
# fp = len(pred_set - target_set)
|
| 125 |
+
# fn = len(target_set - pred_set)
|
| 126 |
+
# precision = tp / (tp + fp) if (tp + fp) > 0 else 0.0
|
| 127 |
+
# recall = tp / (tp + fn) if (tp + fn) > 0 else 0.0
|
| 128 |
+
# if precision + recall == 0:
|
| 129 |
+
# return 0.0
|
| 130 |
+
# return 2 * precision * recall / (precision + recall)
|
| 131 |
+
|
| 132 |
+
# # 文件路径
|
| 133 |
+
# file_path = '/nas/shared/kilab/wangyujia/BIO/ablation/enzyme_commission_number.jsonl'
|
| 134 |
+
|
| 135 |
+
# all_f1_scores = []
|
| 136 |
+
|
| 137 |
+
# with open(file_path, 'r', encoding='utf-8') as f:
|
| 138 |
+
# for line in f:
|
| 139 |
+
# try:
|
| 140 |
+
# data = json.loads(line)
|
| 141 |
+
# reference = data.get('reference_answer', '')
|
| 142 |
+
# generated = data.get('generated_answer', '')
|
| 143 |
+
# target_set = extract_numbers_from_reference_answer(reference)
|
| 144 |
+
# pred_set = extract_numbers_from_generated_answer(generated)
|
| 145 |
+
# f1 = calculate_f1(pred_set, target_set)
|
| 146 |
+
# all_f1_scores.append(f1)
|
| 147 |
+
# except Exception as e:
|
| 148 |
+
# print(f"处理行时出错:{line.strip()}")
|
| 149 |
+
# print(f"错误:{e}")
|
| 150 |
+
|
| 151 |
+
# # 输出结果
|
| 152 |
+
# if all_f1_scores:
|
| 153 |
+
# avg_f1 = sum(all_f1_scores) / len(all_f1_scores)
|
| 154 |
+
# print(f"总样本数:{len(all_f1_scores)}")
|
| 155 |
+
# print(f"平均 F1 分数:{avg_f1:.5f}")
|
| 156 |
+
# else:
|
| 157 |
+
# print("没有有效的F1分数可供计算")
|
BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/args.json
ADDED
|
@@ -0,0 +1,364 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model": "/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged",
|
| 3 |
+
"model_type": "qwen2_5",
|
| 4 |
+
"model_revision": null,
|
| 5 |
+
"task_type": "causal_lm",
|
| 6 |
+
"torch_dtype": "bfloat16",
|
| 7 |
+
"attn_impl": null,
|
| 8 |
+
"num_labels": null,
|
| 9 |
+
"problem_type": null,
|
| 10 |
+
"rope_scaling": null,
|
| 11 |
+
"device_map": null,
|
| 12 |
+
"max_memory": {},
|
| 13 |
+
"local_repo_path": null,
|
| 14 |
+
"template": "qwen2_5",
|
| 15 |
+
"system": null,
|
| 16 |
+
"max_length": 8192,
|
| 17 |
+
"truncation_strategy": "delete",
|
| 18 |
+
"max_pixels": null,
|
| 19 |
+
"tools_prompt": "react_en",
|
| 20 |
+
"norm_bbox": null,
|
| 21 |
+
"response_prefix": null,
|
| 22 |
+
"padding_side": "right",
|
| 23 |
+
"loss_scale": "default",
|
| 24 |
+
"sequence_parallel_size": 1,
|
| 25 |
+
"use_chat_template": true,
|
| 26 |
+
"template_backend": "swift",
|
| 27 |
+
"dataset": [
|
| 28 |
+
"/oss/wangyujia/ProtT3/ProtT3/data/sft/dataset/metal_ion_binding/train.jsonl"
|
| 29 |
+
],
|
| 30 |
+
"val_dataset": [],
|
| 31 |
+
"split_dataset_ratio": 0.01,
|
| 32 |
+
"data_seed": 42,
|
| 33 |
+
"dataset_num_proc": 128,
|
| 34 |
+
"dataset_shuffle": true,
|
| 35 |
+
"val_dataset_shuffle": false,
|
| 36 |
+
"streaming": false,
|
| 37 |
+
"interleave_prob": null,
|
| 38 |
+
"stopping_strategy": "first_exhausted",
|
| 39 |
+
"shuffle_buffer_size": 1000,
|
| 40 |
+
"enable_cache": false,
|
| 41 |
+
"download_mode": "reuse_dataset_if_exists",
|
| 42 |
+
"columns": {},
|
| 43 |
+
"strict": false,
|
| 44 |
+
"remove_unused_columns": true,
|
| 45 |
+
"model_name": [
|
| 46 |
+
"qwen_bio_sft_deeplocbinary-08022035"
|
| 47 |
+
],
|
| 48 |
+
"model_author": [
|
| 49 |
+
"swift"
|
| 50 |
+
],
|
| 51 |
+
"custom_dataset_info": [],
|
| 52 |
+
"quant_method": null,
|
| 53 |
+
"quant_bits": null,
|
| 54 |
+
"hqq_axis": null,
|
| 55 |
+
"bnb_4bit_compute_dtype": "bfloat16",
|
| 56 |
+
"bnb_4bit_quant_type": "nf4",
|
| 57 |
+
"bnb_4bit_use_double_quant": true,
|
| 58 |
+
"bnb_4bit_quant_storage": null,
|
| 59 |
+
"max_new_tokens": 64,
|
| 60 |
+
"temperature": 0.0,
|
| 61 |
+
"top_k": null,
|
| 62 |
+
"top_p": null,
|
| 63 |
+
"repetition_penalty": null,
|
| 64 |
+
"num_beams": 1,
|
| 65 |
+
"stream": false,
|
| 66 |
+
"stop_words": [],
|
| 67 |
+
"logprobs": false,
|
| 68 |
+
"top_logprobs": null,
|
| 69 |
+
"ckpt_dir": "/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged",
|
| 70 |
+
"load_dataset_config": null,
|
| 71 |
+
"lora_modules": [],
|
| 72 |
+
"tuner_backend": "peft",
|
| 73 |
+
"train_type": "lora",
|
| 74 |
+
"adapters": [],
|
| 75 |
+
"external_plugins": [],
|
| 76 |
+
"seed": 42,
|
| 77 |
+
"model_kwargs": {},
|
| 78 |
+
"load_args": false,
|
| 79 |
+
"load_data_args": false,
|
| 80 |
+
"use_hf": false,
|
| 81 |
+
"hub_token": null,
|
| 82 |
+
"custom_register_path": [],
|
| 83 |
+
"ignore_args_error": false,
|
| 84 |
+
"use_swift_lora": false,
|
| 85 |
+
"output_dir": "/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822",
|
| 86 |
+
"overwrite_output_dir": false,
|
| 87 |
+
"do_train": false,
|
| 88 |
+
"do_eval": false,
|
| 89 |
+
"do_predict": false,
|
| 90 |
+
"eval_strategy": "steps",
|
| 91 |
+
"prediction_loss_only": false,
|
| 92 |
+
"per_device_train_batch_size": 2,
|
| 93 |
+
"per_device_eval_batch_size": 2,
|
| 94 |
+
"per_gpu_train_batch_size": null,
|
| 95 |
+
"per_gpu_eval_batch_size": null,
|
| 96 |
+
"gradient_accumulation_steps": 4,
|
| 97 |
+
"eval_accumulation_steps": null,
|
| 98 |
+
"eval_delay": 0,
|
| 99 |
+
"torch_empty_cache_steps": null,
|
| 100 |
+
"learning_rate": 1e-05,
|
| 101 |
+
"weight_decay": 0.1,
|
| 102 |
+
"adam_beta1": 0.9,
|
| 103 |
+
"adam_beta2": 0.95,
|
| 104 |
+
"adam_epsilon": 1e-08,
|
| 105 |
+
"max_grad_norm": 1.0,
|
| 106 |
+
"num_train_epochs": 3.0,
|
| 107 |
+
"max_steps": -1,
|
| 108 |
+
"lr_scheduler_type": "cosine",
|
| 109 |
+
"lr_scheduler_kwargs": null,
|
| 110 |
+
"warmup_ratio": 0.05,
|
| 111 |
+
"warmup_steps": 0,
|
| 112 |
+
"log_level": "passive",
|
| 113 |
+
"log_level_replica": "warning",
|
| 114 |
+
"log_on_each_node": true,
|
| 115 |
+
"logging_dir": "/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/runs",
|
| 116 |
+
"logging_strategy": "steps",
|
| 117 |
+
"logging_first_step": true,
|
| 118 |
+
"logging_steps": 1,
|
| 119 |
+
"logging_nan_inf_filter": true,
|
| 120 |
+
"save_strategy": "steps",
|
| 121 |
+
"save_steps": 5.0,
|
| 122 |
+
"save_total_limit": 5,
|
| 123 |
+
"save_safetensors": true,
|
| 124 |
+
"save_on_each_node": false,
|
| 125 |
+
"save_only_model": true,
|
| 126 |
+
"restore_callback_states_from_checkpoint": false,
|
| 127 |
+
"no_cuda": false,
|
| 128 |
+
"use_cpu": false,
|
| 129 |
+
"use_mps_device": false,
|
| 130 |
+
"jit_mode_eval": false,
|
| 131 |
+
"use_ipex": false,
|
| 132 |
+
"bf16": true,
|
| 133 |
+
"fp16": false,
|
| 134 |
+
"fp16_opt_level": "O1",
|
| 135 |
+
"half_precision_backend": "auto",
|
| 136 |
+
"bf16_full_eval": false,
|
| 137 |
+
"fp16_full_eval": false,
|
| 138 |
+
"tf32": null,
|
| 139 |
+
"local_rank": 0,
|
| 140 |
+
"ddp_backend": null,
|
| 141 |
+
"tpu_num_cores": null,
|
| 142 |
+
"tpu_metrics_debug": false,
|
| 143 |
+
"debug": null,
|
| 144 |
+
"dataloader_drop_last": false,
|
| 145 |
+
"eval_steps": 5.0,
|
| 146 |
+
"dataloader_num_workers": 1,
|
| 147 |
+
"dataloader_prefetch_factor": null,
|
| 148 |
+
"past_index": -1,
|
| 149 |
+
"run_name": "construct",
|
| 150 |
+
"disable_tqdm": null,
|
| 151 |
+
"label_names": null,
|
| 152 |
+
"load_best_model_at_end": false,
|
| 153 |
+
"metric_for_best_model": "loss",
|
| 154 |
+
"greater_is_better": false,
|
| 155 |
+
"ignore_data_skip": false,
|
| 156 |
+
"fsdp": "",
|
| 157 |
+
"fsdp_min_num_params": 0,
|
| 158 |
+
"fsdp_config": null,
|
| 159 |
+
"tp_size": 0,
|
| 160 |
+
"fsdp_transformer_layer_cls_to_wrap": null,
|
| 161 |
+
"accelerator_config": {
|
| 162 |
+
"dispatch_batches": false
|
| 163 |
+
},
|
| 164 |
+
"deepspeed": {
|
| 165 |
+
"fp16": {
|
| 166 |
+
"enabled": "auto",
|
| 167 |
+
"loss_scale": 0,
|
| 168 |
+
"loss_scale_window": 1000,
|
| 169 |
+
"initial_scale_power": 16,
|
| 170 |
+
"hysteresis": 2,
|
| 171 |
+
"min_loss_scale": 1
|
| 172 |
+
},
|
| 173 |
+
"bf16": {
|
| 174 |
+
"enabled": "auto"
|
| 175 |
+
},
|
| 176 |
+
"zero_optimization": {
|
| 177 |
+
"stage": 3,
|
| 178 |
+
"offload_optimizer": {
|
| 179 |
+
"device": "none",
|
| 180 |
+
"pin_memory": true
|
| 181 |
+
},
|
| 182 |
+
"offload_param": {
|
| 183 |
+
"device": "none",
|
| 184 |
+
"pin_memory": true
|
| 185 |
+
},
|
| 186 |
+
"overlap_comm": false,
|
| 187 |
+
"contiguous_gradients": true,
|
| 188 |
+
"sub_group_size": 1000000000.0,
|
| 189 |
+
"reduce_bucket_size": "auto",
|
| 190 |
+
"zero_quantized_weights": false,
|
| 191 |
+
"zero_quantized_gradients": false,
|
| 192 |
+
"stage3_prefetch_bucket_size": "auto",
|
| 193 |
+
"stage3_param_persistence_threshold": "auto",
|
| 194 |
+
"stage3_max_live_parameters": 1000000000.0,
|
| 195 |
+
"stage3_max_reuse_distance": 1000000000.0,
|
| 196 |
+
"stage3_gather_16bit_weights_on_model_save": true
|
| 197 |
+
},
|
| 198 |
+
"gradient_accumulation_steps": "auto",
|
| 199 |
+
"gradient_clipping": "auto",
|
| 200 |
+
"steps_per_print": 2000,
|
| 201 |
+
"train_batch_size": "auto",
|
| 202 |
+
"train_micro_batch_size_per_gpu": "auto",
|
| 203 |
+
"wall_clock_breakdown": false
|
| 204 |
+
},
|
| 205 |
+
"label_smoothing_factor": 0.0,
|
| 206 |
+
"optim": "adamw_torch",
|
| 207 |
+
"optim_args": null,
|
| 208 |
+
"adafactor": false,
|
| 209 |
+
"group_by_length": false,
|
| 210 |
+
"length_column_name": "length",
|
| 211 |
+
"report_to": [
|
| 212 |
+
"tensorboard"
|
| 213 |
+
],
|
| 214 |
+
"ddp_find_unused_parameters": null,
|
| 215 |
+
"ddp_bucket_cap_mb": null,
|
| 216 |
+
"ddp_broadcast_buffers": null,
|
| 217 |
+
"dataloader_pin_memory": true,
|
| 218 |
+
"dataloader_persistent_workers": false,
|
| 219 |
+
"skip_memory_metrics": true,
|
| 220 |
+
"use_legacy_prediction_loop": false,
|
| 221 |
+
"push_to_hub": false,
|
| 222 |
+
"resume_from_checkpoint": null,
|
| 223 |
+
"hub_model_id": null,
|
| 224 |
+
"hub_strategy": "every_save",
|
| 225 |
+
"hub_private_repo": null,
|
| 226 |
+
"hub_always_push": false,
|
| 227 |
+
"gradient_checkpointing": true,
|
| 228 |
+
"gradient_checkpointing_kwargs": null,
|
| 229 |
+
"include_inputs_for_metrics": false,
|
| 230 |
+
"include_for_metrics": [],
|
| 231 |
+
"eval_do_concat_batches": true,
|
| 232 |
+
"fp16_backend": "auto",
|
| 233 |
+
"push_to_hub_model_id": null,
|
| 234 |
+
"push_to_hub_organization": null,
|
| 235 |
+
"push_to_hub_token": null,
|
| 236 |
+
"mp_parameters": "",
|
| 237 |
+
"auto_find_batch_size": false,
|
| 238 |
+
"full_determinism": false,
|
| 239 |
+
"torchdynamo": null,
|
| 240 |
+
"ray_scope": "last",
|
| 241 |
+
"ddp_timeout": 1800,
|
| 242 |
+
"torch_compile": false,
|
| 243 |
+
"torch_compile_backend": null,
|
| 244 |
+
"torch_compile_mode": null,
|
| 245 |
+
"include_tokens_per_second": false,
|
| 246 |
+
"include_num_input_tokens_seen": false,
|
| 247 |
+
"neftune_noise_alpha": null,
|
| 248 |
+
"optim_target_modules": null,
|
| 249 |
+
"batch_eval_metrics": false,
|
| 250 |
+
"eval_on_start": false,
|
| 251 |
+
"use_liger_kernel": false,
|
| 252 |
+
"eval_use_gather_object": false,
|
| 253 |
+
"average_tokens_across_devices": false,
|
| 254 |
+
"sortish_sampler": false,
|
| 255 |
+
"predict_with_generate": false,
|
| 256 |
+
"generation_max_length": null,
|
| 257 |
+
"generation_num_beams": null,
|
| 258 |
+
"generation_config": null,
|
| 259 |
+
"check_model": true,
|
| 260 |
+
"acc_strategy": "token",
|
| 261 |
+
"train_dataloader_shuffle": true,
|
| 262 |
+
"metric_warmup_step": 0,
|
| 263 |
+
"fsdp_num": 1,
|
| 264 |
+
"acc_steps": 1,
|
| 265 |
+
"eval_use_evalscope": false,
|
| 266 |
+
"eval_datasets": [],
|
| 267 |
+
"eval_limit": null,
|
| 268 |
+
"eval_datasets_args": null,
|
| 269 |
+
"eval_generation_config": null,
|
| 270 |
+
"freeze_parameters": [],
|
| 271 |
+
"freeze_parameters_ratio": 0.0,
|
| 272 |
+
"trainable_parameters": [],
|
| 273 |
+
"freeze_llm": false,
|
| 274 |
+
"freeze_vit": true,
|
| 275 |
+
"freeze_aligner": true,
|
| 276 |
+
"target_modules": [
|
| 277 |
+
"all-linear"
|
| 278 |
+
],
|
| 279 |
+
"target_regex": null,
|
| 280 |
+
"modules_to_save": [],
|
| 281 |
+
"lora_rank": 8,
|
| 282 |
+
"lora_alpha": 32,
|
| 283 |
+
"lora_dropout": 0.05,
|
| 284 |
+
"lora_bias": "none",
|
| 285 |
+
"lora_dtype": null,
|
| 286 |
+
"lorap_lr_ratio": null,
|
| 287 |
+
"use_rslora": false,
|
| 288 |
+
"use_dora": false,
|
| 289 |
+
"lora_ga_batch_size": 2,
|
| 290 |
+
"lora_ga_iters": 2,
|
| 291 |
+
"lora_ga_max_length": 1024,
|
| 292 |
+
"lora_ga_direction": "ArB2r",
|
| 293 |
+
"lora_ga_scale": "stable",
|
| 294 |
+
"lora_ga_stable_gamma": 16,
|
| 295 |
+
"init_weights": true,
|
| 296 |
+
"fourier_n_frequency": 2000,
|
| 297 |
+
"fourier_scaling": 300.0,
|
| 298 |
+
"boft_block_size": 4,
|
| 299 |
+
"boft_block_num": 0,
|
| 300 |
+
"boft_n_butterfly_factor": 1,
|
| 301 |
+
"boft_dropout": 0.0,
|
| 302 |
+
"vera_rank": 256,
|
| 303 |
+
"vera_projection_prng_key": 0,
|
| 304 |
+
"vera_dropout": 0.0,
|
| 305 |
+
"vera_d_initial": 0.1,
|
| 306 |
+
"adapter_act": "gelu",
|
| 307 |
+
"adapter_length": 128,
|
| 308 |
+
"use_galore": false,
|
| 309 |
+
"galore_target_modules": null,
|
| 310 |
+
"galore_rank": 128,
|
| 311 |
+
"galore_update_proj_gap": 50,
|
| 312 |
+
"galore_scale": 1.0,
|
| 313 |
+
"galore_proj_type": "std",
|
| 314 |
+
"galore_optim_per_parameter": false,
|
| 315 |
+
"galore_with_embedding": false,
|
| 316 |
+
"galore_quantization": false,
|
| 317 |
+
"galore_proj_quant": false,
|
| 318 |
+
"galore_proj_bits": 4,
|
| 319 |
+
"galore_proj_group_size": 256,
|
| 320 |
+
"galore_cos_threshold": 0.4,
|
| 321 |
+
"galore_gamma_proj": 2,
|
| 322 |
+
"galore_queue_size": 5,
|
| 323 |
+
"adalora_target_r": 8,
|
| 324 |
+
"adalora_init_r": 12,
|
| 325 |
+
"adalora_tinit": 0,
|
| 326 |
+
"adalora_tfinal": 0,
|
| 327 |
+
"adalora_deltaT": 1,
|
| 328 |
+
"adalora_beta1": 0.85,
|
| 329 |
+
"adalora_beta2": 0.85,
|
| 330 |
+
"adalora_orth_reg_weight": 0.5,
|
| 331 |
+
"llamapro_num_new_blocks": 4,
|
| 332 |
+
"llamapro_num_groups": null,
|
| 333 |
+
"lisa_activated_layers": 0,
|
| 334 |
+
"lisa_step_interval": 20,
|
| 335 |
+
"reft_layer_key": null,
|
| 336 |
+
"reft_layers": null,
|
| 337 |
+
"reft_rank": 4,
|
| 338 |
+
"reft_intervention_type": "LoreftIntervention",
|
| 339 |
+
"reft_args": null,
|
| 340 |
+
"swanlab_token": null,
|
| 341 |
+
"swanlab_project": null,
|
| 342 |
+
"swanlab_workspace": null,
|
| 343 |
+
"swanlab_exp_name": null,
|
| 344 |
+
"swanlab_mode": "cloud",
|
| 345 |
+
"add_version": true,
|
| 346 |
+
"resume_only_model": false,
|
| 347 |
+
"create_checkpoint_symlink": false,
|
| 348 |
+
"packing": false,
|
| 349 |
+
"lazy_tokenize": false,
|
| 350 |
+
"loss_type": null,
|
| 351 |
+
"optimizer": null,
|
| 352 |
+
"metric": null,
|
| 353 |
+
"zero_hpz_partition_size": null,
|
| 354 |
+
"rank": 0,
|
| 355 |
+
"global_world_size": 8,
|
| 356 |
+
"local_world_size": 8,
|
| 357 |
+
"model_suffix": "checkpoint-50-merged",
|
| 358 |
+
"model_info": "ModelInfo(model_type='qwen2_5', model_dir='/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged', torch_dtype=torch.bfloat16, max_model_len=32768, quant_method=None, quant_bits=None, rope_scaling=None, config=None, task_type='causal_lm', num_labels=None)",
|
| 359 |
+
"model_meta": "ModelMeta(model_type='qwen2_5', model_groups=[ModelGroup(models=[Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct', hf_model_id='Qwen/Qwen2.5-3B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct', hf_model_id='Qwen/Qwen2.5-7B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct', hf_model_id='Qwen/Qwen2.5-14B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct', hf_model_id='Qwen/Qwen2.5-32B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct', hf_model_id='Qwen/Qwen2.5-72B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B', hf_model_id='Qwen/Qwen2.5-0.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B', hf_model_id='Qwen/Qwen2.5-1.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B', hf_model_id='Qwen/Qwen2.5-3B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B', hf_model_id='Qwen/Qwen2.5-7B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B', hf_model_id='Qwen/Qwen2.5-14B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B', hf_model_id='Qwen/Qwen2.5-32B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B', hf_model_id='Qwen/Qwen2.5-72B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-3B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-32B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-72B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-0.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-0.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-1.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-1.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-3B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-3B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-7B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-7B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-14B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-14B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-32B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-32B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-72B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-72B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=[]), ModelGroup(models=[Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B', hf_model_id='Qwen/Qwen2.5-Coder-0.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B', hf_model_id='Qwen/Qwen2.5-Coder-1.5B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B', hf_model_id='Qwen/Qwen2.5-Coder-3B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B', hf_model_id='Qwen/Qwen2.5-Coder-7B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B', hf_model_id='Qwen/Qwen2.5-Coder-14B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B', hf_model_id='Qwen/Qwen2.5-Coder-32B', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-AWQ', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-AWQ', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-0.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-1.5B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-3B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int4', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int4', model_path=None, ms_revision=None, hf_revision=None), Model(ms_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int8', hf_model_id='Qwen/Qwen2.5-Coder-32B-Instruct-GPTQ-Int8', model_path=None, ms_revision=None, hf_revision=None)], ignore_patterns=None, requires=None, tags=['coding'])], template='qwen2_5', get_function=<function get_model_tokenizer_with_flash_attn at 0x7fbf7497de10>, model_arch='llama', architectures=['Qwen2ForCausalLM'], additional_saved_files=[], torch_dtype=None, is_multimodal=False, is_reward=False, task_type=None, ignore_patterns=[], requires=['transformers>=4.37'], tags=[])",
|
| 360 |
+
"model_dir": "/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged",
|
| 361 |
+
"hub": "<class 'swift.hub.hub.MSHub'>",
|
| 362 |
+
"evaluation_strategy": "steps",
|
| 363 |
+
"training_args": "Seq2SeqTrainingArguments(output_dir='/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822', overwrite_output_dir=False, do_train=False, do_eval=True, do_predict=False, eval_strategy=<IntervalStrategy.STEPS: 'steps'>, prediction_loss_only=False, per_device_train_batch_size=2, per_device_eval_batch_size=2, per_gpu_train_batch_size=None, per_gpu_eval_batch_size=None, gradient_accumulation_steps=4, eval_accumulation_steps=None, eval_delay=0, torch_empty_cache_steps=None, learning_rate=1e-05, weight_decay=0.1, adam_beta1=0.9, adam_beta2=0.95, adam_epsilon=1e-08, max_grad_norm=1.0, num_train_epochs=3.0, max_steps=-1, lr_scheduler_type=<SchedulerType.COSINE: 'cosine'>, lr_scheduler_kwargs=None, warmup_ratio=0.05, warmup_steps=0, log_level='passive', log_level_replica='warning', log_on_each_node=True, logging_dir='/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/runs', logging_strategy=<IntervalStrategy.STEPS: 'steps'>, logging_first_step=True, logging_steps=1, logging_nan_inf_filter=True, save_strategy=<SaveStrategy.STEPS: 'steps'>, save_steps=5, save_total_limit=5, save_safetensors=True, save_on_each_node=False, save_only_model=True, restore_callback_states_from_checkpoint=False, no_cuda=False, use_cpu=False, use_mps_device=False, seed=42, data_seed=42, jit_mode_eval=False, use_ipex=False, bf16=True, fp16=False, fp16_opt_level='O1', half_precision_backend='auto', bf16_full_eval=False, fp16_full_eval=False, tf32=None, local_rank=0, ddp_backend=None, tpu_num_cores=None, tpu_metrics_debug=False, debug=[], dataloader_drop_last=False, eval_steps=5, dataloader_num_workers=1, dataloader_prefetch_factor=10, past_index=-1, run_name='/nas/shared/kilab/wangyujia/BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822', disable_tqdm=False, remove_unused_columns=False, label_names=None, load_best_model_at_end=False, metric_for_best_model='loss', greater_is_better=False, ignore_data_skip=False, fsdp=[], fsdp_min_num_params=0, fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, tp_size=0, fsdp_transformer_layer_cls_to_wrap=None, accelerator_config=AcceleratorConfig(split_batches=False, dispatch_batches=False, even_batches=True, use_seedable_sampler=True, non_blocking=False, gradient_accumulation_kwargs=None, use_configured_state=False), deepspeed={'fp16': {'enabled': 'auto', 'loss_scale': 0, 'loss_scale_window': 1000, 'initial_scale_power': 16, 'hysteresis': 2, 'min_loss_scale': 1}, 'bf16': {'enabled': 'auto'}, 'zero_optimization': {'stage': 3, 'offload_optimizer': {'device': 'none', 'pin_memory': True}, 'offload_param': {'device': 'none', 'pin_memory': True}, 'overlap_comm': False, 'contiguous_gradients': True, 'sub_group_size': 1000000000.0, 'reduce_bucket_size': 'auto', 'zero_quantized_weights': False, 'zero_quantized_gradients': False, 'stage3_prefetch_bucket_size': 'auto', 'stage3_param_persistence_threshold': 'auto', 'stage3_max_live_parameters': 1000000000.0, 'stage3_max_reuse_distance': 1000000000.0, 'stage3_gather_16bit_weights_on_model_save': True}, 'gradient_accumulation_steps': 'auto', 'gradient_clipping': 'auto', 'steps_per_print': 2000, 'train_batch_size': 'auto', 'train_micro_batch_size_per_gpu': 'auto', 'wall_clock_breakdown': False}, label_smoothing_factor=0.0, optim=<OptimizerNames.ADAMW_TORCH: 'adamw_torch'>, optim_args=None, adafactor=False, group_by_length=False, length_column_name='length', report_to=['tensorboard'], ddp_find_unused_parameters=None, ddp_bucket_cap_mb=None, ddp_broadcast_buffers=None, dataloader_pin_memory=True, dataloader_persistent_workers=False, skip_memory_metrics=True, use_legacy_prediction_loop=False, push_to_hub=False, resume_from_checkpoint=None, hub_model_id=None, hub_strategy=<HubStrategy.EVERY_SAVE: 'every_save'>, hub_token=None, hub_private_repo=None, hub_always_push=False, gradient_checkpointing=True, gradient_checkpointing_kwargs=None, include_inputs_for_metrics=False, include_for_metrics=[], eval_do_concat_batches=True, fp16_backend='auto', push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=None, mp_parameters='', auto_find_batch_size=False, full_determinism=False, torchdynamo=None, ray_scope='last', ddp_timeout=1800, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, include_tokens_per_second=None, include_num_input_tokens_seen=None, neftune_noise_alpha=None, optim_target_modules=None, batch_eval_metrics=False, eval_on_start=False, use_liger_kernel=False, eval_use_gather_object=False, average_tokens_across_devices=None, sortish_sampler=False, predict_with_generate=False, generation_max_length=None, generation_num_beams=None, generation_config=None, check_model=True, acc_strategy='token', train_dataloader_shuffle=True, metric_warmup_step=0, fsdp_num=1, acc_steps=1, eval_use_evalscope=False, eval_datasets=[], eval_limit=None, eval_datasets_args=None, eval_generation_config=None, train_type='lora', optimizer=None, local_repo_path=None, galore_config=None)"
|
| 364 |
+
}
|
BIO/sft/qwen-metal_ion_binding-08022141/v0-20250802-215822/logging.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-08-12T13:13:01.108626404+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.1"}
|
| 2 |
+
{"time":"2025-08-12T13:13:02.160994324+08:00","level":"INFO","msg":"stream: created new stream","id":"zr3pdtci"}
|
| 3 |
+
{"time":"2025-08-12T13:13:02.161037833+08:00","level":"INFO","msg":"stream: started","id":"zr3pdtci"}
|
| 4 |
+
{"time":"2025-08-12T13:13:02.16105997+08:00","level":"INFO","msg":"writer: started","stream_id":"zr3pdtci"}
|
| 5 |
+
{"time":"2025-08-12T13:13:02.161100503+08:00","level":"INFO","msg":"sender: started","stream_id":"zr3pdtci"}
|
| 6 |
+
{"time":"2025-08-12T13:13:02.161076037+08:00","level":"INFO","msg":"handler: started","stream_id":"zr3pdtci"}
|
| 7 |
+
{"time":"2025-08-12T13:13:08.73352003+08:00","level":"INFO","msg":"stream: closing","id":"zr3pdtci"}
|
| 8 |
+
{"time":"2025-08-12T13:13:09.557916985+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39586->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 9 |
+
{"time":"2025-08-12T13:13:09.571553285+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39582->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 10 |
+
{"time":"2025-08-12T13:13:09.661494759+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39570->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 11 |
+
{"time":"2025-08-12T13:15:13.828654367+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:42648->142.250.217.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 12 |
+
{"time":"2025-08-12T13:15:13.836170243+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:42646->142.250.217.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 13 |
+
{"time":"2025-08-12T13:15:16.514141941+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39884->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 14 |
+
{"time":"2025-08-12T13:15:22.024313571+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:42662->142.250.217.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 15 |
+
{"time":"2025-08-12T13:17:24.906497736+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:55432->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-metadata.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=036a136a48cc430b2e124c6a943bd3365d0148bc43f506ac3f83240e62fddef675010a7570e490507404976772f3c801d25c150805086db53f8d97e556c9a7d808333fb6fb906571829baac68a31803efb6406d05de6c960dfb4b34bc7f4a211fdcd35bbb67b78b70cd99028fa098777ec10b296a9152931092b7f023604c9ac1bbe017f8f5d1f2cfd40e62862a63667e27832daa864ce133d6fa234ee8bc848b4eda73609490a4a1131c125fef8a276fa267d984774633ac0be08bad83765243743aada64d391d87a24fc4c2919668044acd67af20bac5bbda64742f0e6d3f0b6f43e0482ffba9aee3beb99ffb8bdb2f6a0587bba67046a76eec719aa3c139b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 16 |
+
{"time":"2025-08-12T13:17:33.096841516+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:59370->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/config.yaml?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=138c39feefe589f0780f5c3a2264b19164b3810353e82168fca5f72c77a0ff8ab6814d953abfad7556522255795b5088904fa68d9a4d44be2894692dda633e81dc549c16cca0ead7036ed628003be06ebe81cf3bbaf6f4603fa2b5788863b914e3fdd5fcece8aa1f69a1e301eba60c33d14778c9d55f374f1e63b33428e1bfb9a65cf747a8e4e7ff3c5951ba44402b91691b364dc5c9cd74c9b94b729124a72c3a31f170c9fe00bc2ed1881d46c6f184fbfbe7904a5e3bfd8654a9b8c122d9de9e23dce34c7b9f33015b95e28e2a9830bca505dc80c4dd32b203fbe486a67120ba82245c6231187385d7ba86c1e17dc3b0aabd6c7b4c82017dc9208ef261033c&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 17 |
+
{"time":"2025-08-12T13:17:33.124262755+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:53630->142.250.73.91:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/requirements.txt?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051303Z&X-Goog-Expires=86399&X-Goog-Signature=61075e983922f3c8156d622bd1220d9e33613bdf5a3a15e70dc02495dd21ed4a27b4878630ad40fc98bbc5e9d97f7cbd6628671dc561fe7b8a04dadd9d5b4d3fd7a90f52e7c5b528e286503ac876ad3c8d84c9517ece6ba2cb4eba83423f4c2c0934da96dfc1631a3db950c65785d04461db81841e71103eea269269dbb2d38c28649f379358827eeb3365b14dfed462f0152614c6c243df9d4bee02ab6c1903138248a07186d58542e829f8bb9f1ae9294b7c1e4138f21066f14f69726847ec0f4b18898a6b8430ee42482beab7654668dc3b39e4ff45b5be12c4f768fcad2b8597c9c78b41c6c666db898849b9c687af7aec0e9f7e554804466aa1d24d953b&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 18 |
+
{"time":"2025-08-12T13:17:33.139133433+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:59376->142.250.73.123:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/wandb-summary.json?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=ba3376ee29fe30c376fb9382ae1e910d7358057b038a08cf2dbbeb15d1b7959f78cde6d4088e7c59a007a66da465739db944f3898a2a6a8056b44b13665594e3c14405e85d0feeb7e2dd1e1c297d3739ead411570cb8f8242e30aa7ac4cc754c91b063ed9f8113701d729238667886da423c5a52c40e49606aede11302d3b922a364dd3873c5565ffd71ef97a53e79bb1ad7d333cce435e1dbb9fdbe8ef87543b5f34348f3c0f50084bba7df83b57ca5e37d45189191f1f86d4b95d89683f89eab18e2f9841aac3a4b9b9763154753c6f889c3f0c7e4aecb36220a6707d7dfbee9b3b240f415e8c83cc6dc23ecc23e3104fb2a2f07a4a075331295550cb42238&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 19 |
+
{"time":"2025-08-12T13:17:37.290872326+08:00","level":"ERROR","msg":"request failed","error":"Put \"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca\": read tcp 10.1.8.118:39894->142.250.73.155:443: read: connection reset by peer","method":"PUT","url":"https://storage.googleapis.com/wandb-production.appspot.com/gia0603yucca/protein-llm-contrastive/zr3pdtci/output.log?X-Goog-Algorithm=GOOG4-RSA-SHA256&X-Goog-Credential=gorilla-files-url-signer-man%40wandb-production.iam.gserviceaccount.com%2F20250812%2Fauto%2Fstorage%2Fgoog4_request&X-Goog-Date=20250812T051309Z&X-Goog-Expires=86399&X-Goog-Signature=65d4e0d34ebdaa8bac5b4cd6c135204f00f5907278622ab191037fc2a5cb99a426481eadac62db007021a0d27d982d1f49ccacbe82920d53ef716bba62e271239f454c342305deb0bc880bfd60bb8a55870e869a445ced72496005481e34fa97512fd8a5f00c9faaac098d661c8246bb5ab8cd5537ed25b3f3305dfa297f69b8dcbd61c8cf07e137fbc5b25d7e03f7427d3cbf6ed8c9132b72e3c450a9c1969ca102016ba31a04ac0bcfcb4ee1ccb1af1366bf88a04dd4bdec32c8ca42f09276420822be59548ec2f693b71fd42958b610c6f96d3020c12e1c655ee953073b439b2ace4a590abc4490235ce17ca37a5cd8caa440ed31b894f9b5446737da527f&X-Goog-SignedHeaders=host&X-User=gia0603yucca"}
|
| 20 |
+
{"time":"2025-08-12T13:17:47.6529587+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 21 |
+
{"time":"2025-08-12T13:17:48.610451999+08:00","level":"INFO","msg":"handler: closed","stream_id":"zr3pdtci"}
|
| 22 |
+
{"time":"2025-08-12T13:17:48.613740545+08:00","level":"INFO","msg":"sender: closed","stream_id":"zr3pdtci"}
|
| 23 |
+
{"time":"2025-08-12T13:17:48.613750169+08:00","level":"INFO","msg":"stream: closed","id":"zr3pdtci"}
|
BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug.log
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-08-12 13:13:00,887 INFO MainThread:87959 [wandb_setup.py:_flush():80] Current SDK version is 0.21.1
|
| 2 |
+
2025-08-12 13:13:00,887 INFO MainThread:87959 [wandb_setup.py:_flush():80] Configure stats pid to 87959
|
| 3 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
|
| 4 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_setup.py:_flush():80] Loading settings from /nas/shared/kilab/wangyujia/BioReason_new/wandb/settings
|
| 5 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_setup.py:_flush():80] Loading settings from environment variables
|
| 6 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /nas/shared/kilab/wangyujia/BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug.log
|
| 7 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /nas/shared/kilab/wangyujia/BioReason_new/wandb/run-20250812_131300-zr3pdtci/logs/debug-internal.log
|
| 8 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:init():830] calling init triggers
|
| 9 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:init():835] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'text_model_name': '/oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged', 'protein_model_name': '/nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m', 'qformer_model_name': '/nas/shared/kilab/wangyujia/ProtT3/plm_model/microsoft', 'cache_dir': '/model-weights', 'num_query_tokens': 8, 'train_dataset': '/nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/train_set.jsonl', 'valid_dataset': '/nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/valid_set.jsonl', 'eval_dataset': True, 'output_dir': './contrastive_outputs', 'num_epochs': 10, 'batch_size': 32, 'learning_rate': 0.0001, 'weight_decay': 0.01, 'warmup_steps': 1000, 'gradient_accumulation_steps': 1, 'temperature': 0.07, 'freeze_protein_model': True, 'freeze_text_model': True, 'protein_weight': 1.0, 'text_weight': 1.0, 'enable_ptm': True, 'ptm_weight': 1.0, 'max_length_protein': 1024, 'max_length_text': 512, 'num_workers': 8, 'logging_steps': 100, 'eval_steps': 500, 'save_steps': 1000, 'save_total_limit': 3, 'fp16': False, 'bf16': False, 'seed': 42, 'use_wandb': True, 'wandb_project': 'protein-llm-contrastive', 'wandb_entity': None, '_wandb': {}}
|
| 11 |
+
2025-08-12 13:13:00,888 INFO MainThread:87959 [wandb_init.py:init():871] starting backend
|
| 12 |
+
2025-08-12 13:13:01,096 INFO MainThread:87959 [wandb_init.py:init():874] sending inform_init request
|
| 13 |
+
2025-08-12 13:13:01,103 INFO MainThread:87959 [wandb_init.py:init():882] backend started and connected
|
| 14 |
+
2025-08-12 13:13:01,104 INFO MainThread:87959 [wandb_init.py:init():953] updated telemetry
|
| 15 |
+
2025-08-12 13:13:01,135 INFO MainThread:87959 [wandb_init.py:init():977] communicating run to backend with 90.0 second timeout
|
| 16 |
+
2025-08-12 13:13:02,774 INFO MainThread:87959 [wandb_init.py:init():1029] starting run threads in backend
|
| 17 |
+
2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_console_start():2494] atexit reg
|
| 18 |
+
2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_redirect():2342] redirect: wrap_raw
|
| 19 |
+
2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_redirect():2411] Wrapping output streams.
|
| 20 |
+
2025-08-12 13:13:02,883 INFO MainThread:87959 [wandb_run.py:_redirect():2434] Redirects installed.
|
| 21 |
+
2025-08-12 13:13:02,886 INFO MainThread:87959 [wandb_init.py:init():1075] run started, returning control to user process
|
| 22 |
+
2025-08-12 13:13:08,732 INFO MsgRouterThr:87959 [mailbox.py:close():129] [no run ID] Closing mailbox, abandoning 1 handles.
|
BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/config.yaml
ADDED
|
@@ -0,0 +1,200 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.21.1
|
| 4 |
+
e:
|
| 5 |
+
crh90aob5eyx2qklfrtgkofnoya37sbu:
|
| 6 |
+
args:
|
| 7 |
+
- --text_model_name
|
| 8 |
+
- /oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged
|
| 9 |
+
- --protein_model_name
|
| 10 |
+
- /nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m
|
| 11 |
+
- --qformer_model_name
|
| 12 |
+
- /nas/shared/kilab/wangyujia/ProtT3/plm_model/microsoft
|
| 13 |
+
- --num_query_tokens
|
| 14 |
+
- "8"
|
| 15 |
+
- --train_dataset
|
| 16 |
+
- /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/train_set.jsonl
|
| 17 |
+
- --valid_dataset
|
| 18 |
+
- /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/valid_set.jsonl
|
| 19 |
+
- --output_dir
|
| 20 |
+
- ./contrastive_outputs
|
| 21 |
+
- --num_epochs
|
| 22 |
+
- "10"
|
| 23 |
+
- --batch_size
|
| 24 |
+
- "32"
|
| 25 |
+
- --learning_rate
|
| 26 |
+
- "1e-4"
|
| 27 |
+
- --temperature
|
| 28 |
+
- "0.07"
|
| 29 |
+
- --freeze_protein_model
|
| 30 |
+
- --freeze_text_model
|
| 31 |
+
- --enable_ptm
|
| 32 |
+
- --max_length_protein
|
| 33 |
+
- "1024"
|
| 34 |
+
- --max_length_text
|
| 35 |
+
- "512"
|
| 36 |
+
- --num_workers
|
| 37 |
+
- "8"
|
| 38 |
+
- --eval_dataset
|
| 39 |
+
- --use_wandb
|
| 40 |
+
- --wandb_project
|
| 41 |
+
- protein-llm-contrastive
|
| 42 |
+
- --logging_steps
|
| 43 |
+
- "100"
|
| 44 |
+
- --eval_steps
|
| 45 |
+
- "500"
|
| 46 |
+
- --save_steps
|
| 47 |
+
- "1000"
|
| 48 |
+
codePath: wangyujia/BioReason_new/train_contrastive.py
|
| 49 |
+
codePathLocal: train_contrastive.py
|
| 50 |
+
cpu_count: 64
|
| 51 |
+
cpu_count_logical: 64
|
| 52 |
+
cudaVersion: "12.1"
|
| 53 |
+
disk:
|
| 54 |
+
/:
|
| 55 |
+
total: "1623302262784"
|
| 56 |
+
used: "29131816960"
|
| 57 |
+
email: gia0603yucca@gmail.com
|
| 58 |
+
executable: /root/miniconda3/envs/bioreason/bin/python
|
| 59 |
+
git:
|
| 60 |
+
commit: b8caf406aa1699c788f0ca6e44a1769452c317db
|
| 61 |
+
remote: https://github.com/PorUna-byte/PAR.git
|
| 62 |
+
gpu: NVIDIA A800-SXM4-80GB
|
| 63 |
+
gpu_count: 8
|
| 64 |
+
gpu_nvidia:
|
| 65 |
+
- architecture: Ampere
|
| 66 |
+
name: NVIDIA A800-SXM4-80GB
|
| 67 |
+
uuid: GPU-71607f78-ad31-1ea4-19c1-908e3e31aaf1
|
| 68 |
+
- architecture: Ampere
|
| 69 |
+
name: NVIDIA A800-SXM4-80GB
|
| 70 |
+
uuid: GPU-92b7dbbd-7ef5-3c5f-ce1c-1d179d7fa587
|
| 71 |
+
- architecture: Ampere
|
| 72 |
+
name: NVIDIA A800-SXM4-80GB
|
| 73 |
+
uuid: GPU-bbc35439-ad79-578b-381b-aba6f0cc0168
|
| 74 |
+
- architecture: Ampere
|
| 75 |
+
name: NVIDIA A800-SXM4-80GB
|
| 76 |
+
uuid: GPU-e492e147-ca2e-76f2-85da-4e08e4deeb14
|
| 77 |
+
- architecture: Ampere
|
| 78 |
+
name: NVIDIA A800-SXM4-80GB
|
| 79 |
+
uuid: GPU-8c4f8e67-4b52-5107-3095-0f007e6378ac
|
| 80 |
+
- architecture: Ampere
|
| 81 |
+
name: NVIDIA A800-SXM4-80GB
|
| 82 |
+
uuid: GPU-7063f0b9-4ca2-6a72-522d-1262899ac5ad
|
| 83 |
+
- architecture: Ampere
|
| 84 |
+
name: NVIDIA A800-SXM4-80GB
|
| 85 |
+
uuid: GPU-3b6e9a37-bcf3-387c-7874-4f8de4abd115
|
| 86 |
+
- architecture: Ampere
|
| 87 |
+
name: NVIDIA A800-SXM4-80GB
|
| 88 |
+
uuid: GPU-92456839-e814-7be9-6817-f3e8da8aa80c
|
| 89 |
+
host: dsw-265304-f8bc5ff76-4mdt5
|
| 90 |
+
memory:
|
| 91 |
+
total: "549755813888"
|
| 92 |
+
os: Linux-5.10.134-008.18.kangaroo.al8.x86_64-x86_64-with-glibc2.35
|
| 93 |
+
program: /nas/shared/kilab/wangyujia/BioReason_new/train_contrastive.py
|
| 94 |
+
python: CPython 3.11.0
|
| 95 |
+
root: /nas/shared/kilab/wangyujia/BioReason_new
|
| 96 |
+
startedAt: "2025-08-12T05:20:50.783496Z"
|
| 97 |
+
writerId: crh90aob5eyx2qklfrtgkofnoya37sbu
|
| 98 |
+
m: []
|
| 99 |
+
python_version: 3.11.0
|
| 100 |
+
t:
|
| 101 |
+
"1":
|
| 102 |
+
- 1
|
| 103 |
+
- 9
|
| 104 |
+
- 11
|
| 105 |
+
- 41
|
| 106 |
+
- 49
|
| 107 |
+
- 51
|
| 108 |
+
- 71
|
| 109 |
+
- 84
|
| 110 |
+
- 98
|
| 111 |
+
- 103
|
| 112 |
+
"2":
|
| 113 |
+
- 1
|
| 114 |
+
- 9
|
| 115 |
+
- 11
|
| 116 |
+
- 41
|
| 117 |
+
- 49
|
| 118 |
+
- 51
|
| 119 |
+
- 71
|
| 120 |
+
- 84
|
| 121 |
+
- 98
|
| 122 |
+
- 103
|
| 123 |
+
"3":
|
| 124 |
+
- 13
|
| 125 |
+
- 16
|
| 126 |
+
"4": 3.11.0
|
| 127 |
+
"5": 0.21.1
|
| 128 |
+
"6": 4.55.0
|
| 129 |
+
"12": 0.21.1
|
| 130 |
+
"13": linux-x86_64
|
| 131 |
+
batch_size:
|
| 132 |
+
value: 32
|
| 133 |
+
bf16:
|
| 134 |
+
value: false
|
| 135 |
+
cache_dir:
|
| 136 |
+
value: /model-weights
|
| 137 |
+
enable_ptm:
|
| 138 |
+
value: true
|
| 139 |
+
eval_dataset:
|
| 140 |
+
value: true
|
| 141 |
+
eval_steps:
|
| 142 |
+
value: 500
|
| 143 |
+
fp16:
|
| 144 |
+
value: false
|
| 145 |
+
freeze_protein_model:
|
| 146 |
+
value: true
|
| 147 |
+
freeze_text_model:
|
| 148 |
+
value: true
|
| 149 |
+
gradient_accumulation_steps:
|
| 150 |
+
value: 1
|
| 151 |
+
learning_rate:
|
| 152 |
+
value: 0.0001
|
| 153 |
+
logging_steps:
|
| 154 |
+
value: 100
|
| 155 |
+
max_length_protein:
|
| 156 |
+
value: 1024
|
| 157 |
+
max_length_text:
|
| 158 |
+
value: 512
|
| 159 |
+
num_epochs:
|
| 160 |
+
value: 10
|
| 161 |
+
num_query_tokens:
|
| 162 |
+
value: 8
|
| 163 |
+
num_workers:
|
| 164 |
+
value: 8
|
| 165 |
+
output_dir:
|
| 166 |
+
value: ./contrastive_outputs
|
| 167 |
+
protein_model_name:
|
| 168 |
+
value: /nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m
|
| 169 |
+
protein_weight:
|
| 170 |
+
value: 1
|
| 171 |
+
ptm_weight:
|
| 172 |
+
value: 1
|
| 173 |
+
qformer_model_name:
|
| 174 |
+
value: /nas/shared/kilab/wangyujia/ProtT3/plm_model/microsoft
|
| 175 |
+
save_steps:
|
| 176 |
+
value: 1000
|
| 177 |
+
save_total_limit:
|
| 178 |
+
value: 3
|
| 179 |
+
seed:
|
| 180 |
+
value: 42
|
| 181 |
+
temperature:
|
| 182 |
+
value: 0.07
|
| 183 |
+
text_model_name:
|
| 184 |
+
value: /oss/wangyujia/BIO/construction_finetuning/alpaca/v1-20250609-141541/checkpoint-50-merged
|
| 185 |
+
text_weight:
|
| 186 |
+
value: 1
|
| 187 |
+
train_dataset:
|
| 188 |
+
value: /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/train_set.jsonl
|
| 189 |
+
use_wandb:
|
| 190 |
+
value: true
|
| 191 |
+
valid_dataset:
|
| 192 |
+
value: /nas/shared/kilab/wangyujia/ProtT3/data/SwissProtV3/valid_set.jsonl
|
| 193 |
+
wandb_entity:
|
| 194 |
+
value: null
|
| 195 |
+
wandb_project:
|
| 196 |
+
value: protein-llm-contrastive
|
| 197 |
+
warmup_steps:
|
| 198 |
+
value: 1000
|
| 199 |
+
weight_decay:
|
| 200 |
+
value: 0.01
|
BioReason_new/wandb/run-20250812_132050-2cfwmlj6/files/output.log
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Loading model...
|
| 2 |
+
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████| 4/4 [00:01<00:00, 2.57it/s]
|
| 3 |
+
Some weights of EsmModel were not initialized from the model checkpoint at /nas/shared/kilab/wangyujia/ProtT3/plm_model/esm2-150m and are newly initialized: ['pooler.dense.bias', 'pooler.dense.weight']
|
| 4 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
| 5 |
+
Loading datasets...
|
| 6 |
+
Training dataset size: 430595
|
| 7 |
+
Eval dataset size: 10000
|
| 8 |
+
bioreason.trainer.contrast_trainer_new
|
| 9 |
+
Traceback (most recent call last):
|
| 10 |
+
File "/nas/shared/kilab/wangyujia/BioReason_new/train_contrastive.py", line 552, in <module>
|
| 11 |
+
trainer = main(args)
|
| 12 |
+
^^^^^^^^^^
|
| 13 |
+
File "/nas/shared/kilab/wangyujia/BioReason_new/train_contrastive.py", line 345, in main
|
| 14 |
+
training_args = ContrastiveTrainingArguments(
|
| 15 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 16 |
+
TypeError: ContrastiveTrainingArguments.__init__() got an unexpected keyword argument 'evaluation_strategy'
|
BioReason_new/wandb/run-20250812_132050-2cfwmlj6/run-2cfwmlj6.wandb
ADDED
|
Binary file (7.45 kB). View file
|
|
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/up_conv_block.py
ADDED
|
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
from annotator.uniformer.mmcv.cnn import ConvModule, build_upsample_layer
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
class UpConvBlock(nn.Module):
|
| 7 |
+
"""Upsample convolution block in decoder for UNet.
|
| 8 |
+
|
| 9 |
+
This upsample convolution block consists of one upsample module
|
| 10 |
+
followed by one convolution block. The upsample module expands the
|
| 11 |
+
high-level low-resolution feature map and the convolution block fuses
|
| 12 |
+
the upsampled high-level low-resolution feature map and the low-level
|
| 13 |
+
high-resolution feature map from encoder.
|
| 14 |
+
|
| 15 |
+
Args:
|
| 16 |
+
conv_block (nn.Sequential): Sequential of convolutional layers.
|
| 17 |
+
in_channels (int): Number of input channels of the high-level
|
| 18 |
+
skip_channels (int): Number of input channels of the low-level
|
| 19 |
+
high-resolution feature map from encoder.
|
| 20 |
+
out_channels (int): Number of output channels.
|
| 21 |
+
num_convs (int): Number of convolutional layers in the conv_block.
|
| 22 |
+
Default: 2.
|
| 23 |
+
stride (int): Stride of convolutional layer in conv_block. Default: 1.
|
| 24 |
+
dilation (int): Dilation rate of convolutional layer in conv_block.
|
| 25 |
+
Default: 1.
|
| 26 |
+
with_cp (bool): Use checkpoint or not. Using checkpoint will save some
|
| 27 |
+
memory while slowing down the training speed. Default: False.
|
| 28 |
+
conv_cfg (dict | None): Config dict for convolution layer.
|
| 29 |
+
Default: None.
|
| 30 |
+
norm_cfg (dict | None): Config dict for normalization layer.
|
| 31 |
+
Default: dict(type='BN').
|
| 32 |
+
act_cfg (dict | None): Config dict for activation layer in ConvModule.
|
| 33 |
+
Default: dict(type='ReLU').
|
| 34 |
+
upsample_cfg (dict): The upsample config of the upsample module in
|
| 35 |
+
decoder. Default: dict(type='InterpConv'). If the size of
|
| 36 |
+
high-level feature map is the same as that of skip feature map
|
| 37 |
+
(low-level feature map from encoder), it does not need upsample the
|
| 38 |
+
high-level feature map and the upsample_cfg is None.
|
| 39 |
+
dcn (bool): Use deformable convolution in convolutional layer or not.
|
| 40 |
+
Default: None.
|
| 41 |
+
plugins (dict): plugins for convolutional layers. Default: None.
|
| 42 |
+
"""
|
| 43 |
+
|
| 44 |
+
def __init__(self,
|
| 45 |
+
conv_block,
|
| 46 |
+
in_channels,
|
| 47 |
+
skip_channels,
|
| 48 |
+
out_channels,
|
| 49 |
+
num_convs=2,
|
| 50 |
+
stride=1,
|
| 51 |
+
dilation=1,
|
| 52 |
+
with_cp=False,
|
| 53 |
+
conv_cfg=None,
|
| 54 |
+
norm_cfg=dict(type='BN'),
|
| 55 |
+
act_cfg=dict(type='ReLU'),
|
| 56 |
+
upsample_cfg=dict(type='InterpConv'),
|
| 57 |
+
dcn=None,
|
| 58 |
+
plugins=None):
|
| 59 |
+
super(UpConvBlock, self).__init__()
|
| 60 |
+
assert dcn is None, 'Not implemented yet.'
|
| 61 |
+
assert plugins is None, 'Not implemented yet.'
|
| 62 |
+
|
| 63 |
+
self.conv_block = conv_block(
|
| 64 |
+
in_channels=2 * skip_channels,
|
| 65 |
+
out_channels=out_channels,
|
| 66 |
+
num_convs=num_convs,
|
| 67 |
+
stride=stride,
|
| 68 |
+
dilation=dilation,
|
| 69 |
+
with_cp=with_cp,
|
| 70 |
+
conv_cfg=conv_cfg,
|
| 71 |
+
norm_cfg=norm_cfg,
|
| 72 |
+
act_cfg=act_cfg,
|
| 73 |
+
dcn=None,
|
| 74 |
+
plugins=None)
|
| 75 |
+
if upsample_cfg is not None:
|
| 76 |
+
self.upsample = build_upsample_layer(
|
| 77 |
+
cfg=upsample_cfg,
|
| 78 |
+
in_channels=in_channels,
|
| 79 |
+
out_channels=skip_channels,
|
| 80 |
+
with_cp=with_cp,
|
| 81 |
+
norm_cfg=norm_cfg,
|
| 82 |
+
act_cfg=act_cfg)
|
| 83 |
+
else:
|
| 84 |
+
self.upsample = ConvModule(
|
| 85 |
+
in_channels,
|
| 86 |
+
skip_channels,
|
| 87 |
+
kernel_size=1,
|
| 88 |
+
stride=1,
|
| 89 |
+
padding=0,
|
| 90 |
+
conv_cfg=conv_cfg,
|
| 91 |
+
norm_cfg=norm_cfg,
|
| 92 |
+
act_cfg=act_cfg)
|
| 93 |
+
|
| 94 |
+
def forward(self, skip, x):
|
| 95 |
+
"""Forward function."""
|
| 96 |
+
|
| 97 |
+
x = self.upsample(x)
|
| 98 |
+
out = torch.cat([skip, x], dim=1)
|
| 99 |
+
out = self.conv_block(out)
|
| 100 |
+
|
| 101 |
+
return out
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/models/utils/weight_init.py
ADDED
|
@@ -0,0 +1,62 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""Modified from https://github.com/rwightman/pytorch-image-
|
| 2 |
+
models/blob/master/timm/models/layers/drop.py."""
|
| 3 |
+
|
| 4 |
+
import math
|
| 5 |
+
import warnings
|
| 6 |
+
|
| 7 |
+
import torch
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
def _no_grad_trunc_normal_(tensor, mean, std, a, b):
|
| 11 |
+
"""Reference: https://people.sc.fsu.edu/~jburkardt/presentations
|
| 12 |
+
/truncated_normal.pdf"""
|
| 13 |
+
|
| 14 |
+
def norm_cdf(x):
|
| 15 |
+
# Computes standard normal cumulative distribution function
|
| 16 |
+
return (1. + math.erf(x / math.sqrt(2.))) / 2.
|
| 17 |
+
|
| 18 |
+
if (mean < a - 2 * std) or (mean > b + 2 * std):
|
| 19 |
+
warnings.warn(
|
| 20 |
+
'mean is more than 2 std from [a, b] in nn.init.trunc_normal_. '
|
| 21 |
+
'The distribution of values may be incorrect.',
|
| 22 |
+
stacklevel=2)
|
| 23 |
+
|
| 24 |
+
with torch.no_grad():
|
| 25 |
+
# Values are generated by using a truncated uniform distribution and
|
| 26 |
+
# then using the inverse CDF for the normal distribution.
|
| 27 |
+
# Get upper and lower cdf values
|
| 28 |
+
lower_bound = norm_cdf((a - mean) / std)
|
| 29 |
+
upper_bound = norm_cdf((b - mean) / std)
|
| 30 |
+
|
| 31 |
+
# Uniformly fill tensor with values from [l, u], then translate to
|
| 32 |
+
# [2l-1, 2u-1].
|
| 33 |
+
tensor.uniform_(2 * lower_bound - 1, 2 * upper_bound - 1)
|
| 34 |
+
|
| 35 |
+
# Use inverse cdf transform for normal distribution to get truncated
|
| 36 |
+
# standard normal
|
| 37 |
+
tensor.erfinv_()
|
| 38 |
+
|
| 39 |
+
# Transform to proper mean, std
|
| 40 |
+
tensor.mul_(std * math.sqrt(2.))
|
| 41 |
+
tensor.add_(mean)
|
| 42 |
+
|
| 43 |
+
# Clamp to ensure it's in the proper range
|
| 44 |
+
tensor.clamp_(min=a, max=b)
|
| 45 |
+
return tensor
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
def trunc_normal_(tensor, mean=0., std=1., a=-2., b=2.):
|
| 49 |
+
r"""Fills the input Tensor with values drawn from a truncated
|
| 50 |
+
normal distribution. The values are effectively drawn from the
|
| 51 |
+
normal distribution :math:`\mathcal{N}(\text{mean}, \text{std}^2)`
|
| 52 |
+
with values outside :math:`[a, b]` redrawn until they are within
|
| 53 |
+
the bounds. The method used for generating the random values works
|
| 54 |
+
best when :math:`a \leq \text{mean} \leq b`.
|
| 55 |
+
Args:
|
| 56 |
+
tensor (``torch.Tensor``): an n-dimensional `torch.Tensor`
|
| 57 |
+
mean (float): the mean of the normal distribution
|
| 58 |
+
std (float): the standard deviation of the normal distribution
|
| 59 |
+
a (float): the minimum cutoff value
|
| 60 |
+
b (float): the maximum cutoff value
|
| 61 |
+
"""
|
| 62 |
+
return _no_grad_trunc_normal_(tensor, mean, std, a, b)
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/__init__.py
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from .encoding import Encoding
|
| 2 |
+
from .wrappers import Upsample, resize
|
| 3 |
+
|
| 4 |
+
__all__ = ['Upsample', 'resize', 'Encoding']
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/encoding.py
ADDED
|
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
from torch import nn
|
| 3 |
+
from torch.nn import functional as F
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
class Encoding(nn.Module):
|
| 7 |
+
"""Encoding Layer: a learnable residual encoder.
|
| 8 |
+
|
| 9 |
+
Input is of shape (batch_size, channels, height, width).
|
| 10 |
+
Output is of shape (batch_size, num_codes, channels).
|
| 11 |
+
|
| 12 |
+
Args:
|
| 13 |
+
channels: dimension of the features or feature channels
|
| 14 |
+
num_codes: number of code words
|
| 15 |
+
"""
|
| 16 |
+
|
| 17 |
+
def __init__(self, channels, num_codes):
|
| 18 |
+
super(Encoding, self).__init__()
|
| 19 |
+
# init codewords and smoothing factor
|
| 20 |
+
self.channels, self.num_codes = channels, num_codes
|
| 21 |
+
std = 1. / ((num_codes * channels)**0.5)
|
| 22 |
+
# [num_codes, channels]
|
| 23 |
+
self.codewords = nn.Parameter(
|
| 24 |
+
torch.empty(num_codes, channels,
|
| 25 |
+
dtype=torch.float).uniform_(-std, std),
|
| 26 |
+
requires_grad=True)
|
| 27 |
+
# [num_codes]
|
| 28 |
+
self.scale = nn.Parameter(
|
| 29 |
+
torch.empty(num_codes, dtype=torch.float).uniform_(-1, 0),
|
| 30 |
+
requires_grad=True)
|
| 31 |
+
|
| 32 |
+
@staticmethod
|
| 33 |
+
def scaled_l2(x, codewords, scale):
|
| 34 |
+
num_codes, channels = codewords.size()
|
| 35 |
+
batch_size = x.size(0)
|
| 36 |
+
reshaped_scale = scale.view((1, 1, num_codes))
|
| 37 |
+
expanded_x = x.unsqueeze(2).expand(
|
| 38 |
+
(batch_size, x.size(1), num_codes, channels))
|
| 39 |
+
reshaped_codewords = codewords.view((1, 1, num_codes, channels))
|
| 40 |
+
|
| 41 |
+
scaled_l2_norm = reshaped_scale * (
|
| 42 |
+
expanded_x - reshaped_codewords).pow(2).sum(dim=3)
|
| 43 |
+
return scaled_l2_norm
|
| 44 |
+
|
| 45 |
+
@staticmethod
|
| 46 |
+
def aggregate(assignment_weights, x, codewords):
|
| 47 |
+
num_codes, channels = codewords.size()
|
| 48 |
+
reshaped_codewords = codewords.view((1, 1, num_codes, channels))
|
| 49 |
+
batch_size = x.size(0)
|
| 50 |
+
|
| 51 |
+
expanded_x = x.unsqueeze(2).expand(
|
| 52 |
+
(batch_size, x.size(1), num_codes, channels))
|
| 53 |
+
encoded_feat = (assignment_weights.unsqueeze(3) *
|
| 54 |
+
(expanded_x - reshaped_codewords)).sum(dim=1)
|
| 55 |
+
return encoded_feat
|
| 56 |
+
|
| 57 |
+
def forward(self, x):
|
| 58 |
+
assert x.dim() == 4 and x.size(1) == self.channels
|
| 59 |
+
# [batch_size, channels, height, width]
|
| 60 |
+
batch_size = x.size(0)
|
| 61 |
+
# [batch_size, height x width, channels]
|
| 62 |
+
x = x.view(batch_size, self.channels, -1).transpose(1, 2).contiguous()
|
| 63 |
+
# assignment_weights: [batch_size, channels, num_codes]
|
| 64 |
+
assignment_weights = F.softmax(
|
| 65 |
+
self.scaled_l2(x, self.codewords, self.scale), dim=2)
|
| 66 |
+
# aggregate
|
| 67 |
+
encoded_feat = self.aggregate(assignment_weights, x, self.codewords)
|
| 68 |
+
return encoded_feat
|
| 69 |
+
|
| 70 |
+
def __repr__(self):
|
| 71 |
+
repr_str = self.__class__.__name__
|
| 72 |
+
repr_str += f'(Nx{self.channels}xHxW =>Nx{self.num_codes}' \
|
| 73 |
+
f'x{self.channels})'
|
| 74 |
+
return repr_str
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/ops/wrappers.py
ADDED
|
@@ -0,0 +1,50 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import warnings
|
| 2 |
+
|
| 3 |
+
import torch.nn as nn
|
| 4 |
+
import torch.nn.functional as F
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
def resize(input,
|
| 8 |
+
size=None,
|
| 9 |
+
scale_factor=None,
|
| 10 |
+
mode='nearest',
|
| 11 |
+
align_corners=None,
|
| 12 |
+
warning=True):
|
| 13 |
+
if warning:
|
| 14 |
+
if size is not None and align_corners:
|
| 15 |
+
input_h, input_w = tuple(int(x) for x in input.shape[2:])
|
| 16 |
+
output_h, output_w = tuple(int(x) for x in size)
|
| 17 |
+
if output_h > input_h or output_w > output_h:
|
| 18 |
+
if ((output_h > 1 and output_w > 1 and input_h > 1
|
| 19 |
+
and input_w > 1) and (output_h - 1) % (input_h - 1)
|
| 20 |
+
and (output_w - 1) % (input_w - 1)):
|
| 21 |
+
warnings.warn(
|
| 22 |
+
f'When align_corners={align_corners}, '
|
| 23 |
+
'the output would more aligned if '
|
| 24 |
+
f'input size {(input_h, input_w)} is `x+1` and '
|
| 25 |
+
f'out size {(output_h, output_w)} is `nx+1`')
|
| 26 |
+
return F.interpolate(input, size, scale_factor, mode, align_corners)
|
| 27 |
+
|
| 28 |
+
|
| 29 |
+
class Upsample(nn.Module):
|
| 30 |
+
|
| 31 |
+
def __init__(self,
|
| 32 |
+
size=None,
|
| 33 |
+
scale_factor=None,
|
| 34 |
+
mode='nearest',
|
| 35 |
+
align_corners=None):
|
| 36 |
+
super(Upsample, self).__init__()
|
| 37 |
+
self.size = size
|
| 38 |
+
if isinstance(scale_factor, tuple):
|
| 39 |
+
self.scale_factor = tuple(float(factor) for factor in scale_factor)
|
| 40 |
+
else:
|
| 41 |
+
self.scale_factor = float(scale_factor) if scale_factor else None
|
| 42 |
+
self.mode = mode
|
| 43 |
+
self.align_corners = align_corners
|
| 44 |
+
|
| 45 |
+
def forward(self, x):
|
| 46 |
+
if not self.size:
|
| 47 |
+
size = [int(t * self.scale_factor) for t in x.shape[-2:]]
|
| 48 |
+
else:
|
| 49 |
+
size = self.size
|
| 50 |
+
return resize(x, size, None, self.mode, self.align_corners)
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/__init__.py
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from .collect_env import collect_env
|
| 2 |
+
from .logger import get_root_logger
|
| 3 |
+
|
| 4 |
+
__all__ = ['get_root_logger', 'collect_env']
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/collect_env.py
ADDED
|
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from annotator.uniformer.mmcv.utils import collect_env as collect_base_env
|
| 2 |
+
from annotator.uniformer.mmcv.utils import get_git_hash
|
| 3 |
+
|
| 4 |
+
import annotator.uniformer.mmseg as mmseg
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
def collect_env():
|
| 8 |
+
"""Collect the information of the running environments."""
|
| 9 |
+
env_info = collect_base_env()
|
| 10 |
+
env_info['MMSegmentation'] = f'{mmseg.__version__}+{get_git_hash()[:7]}'
|
| 11 |
+
|
| 12 |
+
return env_info
|
| 13 |
+
|
| 14 |
+
|
| 15 |
+
if __name__ == '__main__':
|
| 16 |
+
for name, val in collect_env().items():
|
| 17 |
+
print('{}: {}'.format(name, val))
|
LAVIS-main/lavis/common/annotator/uniformer/mmseg/utils/logger.py
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import logging
|
| 2 |
+
|
| 3 |
+
from annotator.uniformer.mmcv.utils import get_logger
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
def get_root_logger(log_file=None, log_level=logging.INFO):
|
| 7 |
+
"""Get the root logger.
|
| 8 |
+
|
| 9 |
+
The logger will be initialized if it has not been initialized. By default a
|
| 10 |
+
StreamHandler will be added. If `log_file` is specified, a FileHandler will
|
| 11 |
+
also be added. The name of the root logger is the top-level package name,
|
| 12 |
+
e.g., "mmseg".
|
| 13 |
+
|
| 14 |
+
Args:
|
| 15 |
+
log_file (str | None): The log filename. If specified, a FileHandler
|
| 16 |
+
will be added to the root logger.
|
| 17 |
+
log_level (int): The root logger level. Note that only the process of
|
| 18 |
+
rank 0 is affected, while other processes will set the level to
|
| 19 |
+
"Error" and be silent most of the time.
|
| 20 |
+
|
| 21 |
+
Returns:
|
| 22 |
+
logging.Logger: The root logger.
|
| 23 |
+
"""
|
| 24 |
+
|
| 25 |
+
logger = get_logger(name='mmseg', log_file=log_file, log_level=log_level)
|
| 26 |
+
|
| 27 |
+
return logger
|
LAVIS-main/lavis/common/vqa_tools/__init__.py
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Copyright (c) 2022, salesforce.com, inc.
|
| 3 |
+
All rights reserved.
|
| 4 |
+
SPDX-License-Identifier: BSD-3-Clause
|
| 5 |
+
For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 6 |
+
"""
|
| 7 |
+
|
| 8 |
+
__author__ = "aagrawal"
|
LAVIS-main/lavis/common/vqa_tools/vqa.py
ADDED
|
@@ -0,0 +1,211 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Copyright (c) 2022, salesforce.com, inc.
|
| 3 |
+
All rights reserved.
|
| 4 |
+
SPDX-License-Identifier: BSD-3-Clause
|
| 5 |
+
For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 6 |
+
"""
|
| 7 |
+
|
| 8 |
+
__author__ = "aagrawal"
|
| 9 |
+
__version__ = "0.9"
|
| 10 |
+
|
| 11 |
+
# Interface for accessing the VQA dataset.
|
| 12 |
+
|
| 13 |
+
# This code is based on the code written by Tsung-Yi Lin for MSCOCO Python API available at the following link:
|
| 14 |
+
# (https://github.com/pdollar/coco/blob/master/PythonAPI/pycocotools/coco.py).
|
| 15 |
+
|
| 16 |
+
# The following functions are defined:
|
| 17 |
+
# VQA - VQA class that loads VQA annotation file and prepares data structures.
|
| 18 |
+
# getQuesIds - Get question ids that satisfy given filter conditions.
|
| 19 |
+
# getImgIds - Get image ids that satisfy given filter conditions.
|
| 20 |
+
# loadQA - Load questions and answers with the specified question ids.
|
| 21 |
+
# showQA - Display the specified questions and answers.
|
| 22 |
+
# loadRes - Load result file and create result object.
|
| 23 |
+
|
| 24 |
+
# Help on each function can be accessed by: "help(COCO.function)"
|
| 25 |
+
|
| 26 |
+
import json
|
| 27 |
+
import datetime
|
| 28 |
+
import copy
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
class VQA:
|
| 32 |
+
def __init__(self, annotation_file=None, question_file=None):
|
| 33 |
+
"""
|
| 34 |
+
Constructor of VQA helper class for reading and visualizing questions and answers.
|
| 35 |
+
:param annotation_file (str): location of VQA annotation file
|
| 36 |
+
:return:
|
| 37 |
+
"""
|
| 38 |
+
# load dataset
|
| 39 |
+
self.dataset = {}
|
| 40 |
+
self.questions = {}
|
| 41 |
+
self.qa = {}
|
| 42 |
+
self.qqa = {}
|
| 43 |
+
self.imgToQA = {}
|
| 44 |
+
if not annotation_file == None and not question_file == None:
|
| 45 |
+
print("loading VQA annotations and questions into memory...")
|
| 46 |
+
time_t = datetime.datetime.utcnow()
|
| 47 |
+
dataset = json.load(open(annotation_file, "r"))
|
| 48 |
+
questions = json.load(open(question_file, "r"))
|
| 49 |
+
self.dataset = dataset
|
| 50 |
+
self.questions = questions
|
| 51 |
+
self.createIndex()
|
| 52 |
+
|
| 53 |
+
def createIndex(self):
|
| 54 |
+
# create index
|
| 55 |
+
print("creating index...")
|
| 56 |
+
imgToQA = {ann["image_id"]: [] for ann in self.dataset["annotations"]}
|
| 57 |
+
qa = {ann["question_id"]: [] for ann in self.dataset["annotations"]}
|
| 58 |
+
qqa = {ann["question_id"]: [] for ann in self.dataset["annotations"]}
|
| 59 |
+
for ann in self.dataset["annotations"]:
|
| 60 |
+
imgToQA[ann["image_id"]] += [ann]
|
| 61 |
+
qa[ann["question_id"]] = ann
|
| 62 |
+
for ques in self.questions["questions"]:
|
| 63 |
+
qqa[ques["question_id"]] = ques
|
| 64 |
+
print("index created!")
|
| 65 |
+
|
| 66 |
+
# create class members
|
| 67 |
+
self.qa = qa
|
| 68 |
+
self.qqa = qqa
|
| 69 |
+
self.imgToQA = imgToQA
|
| 70 |
+
|
| 71 |
+
def info(self):
|
| 72 |
+
"""
|
| 73 |
+
Print information about the VQA annotation file.
|
| 74 |
+
:return:
|
| 75 |
+
"""
|
| 76 |
+
for key, value in self.datset["info"].items():
|
| 77 |
+
print("%s: %s" % (key, value))
|
| 78 |
+
|
| 79 |
+
def getQuesIds(self, imgIds=[], quesTypes=[], ansTypes=[]):
|
| 80 |
+
"""
|
| 81 |
+
Get question ids that satisfy given filter conditions. default skips that filter
|
| 82 |
+
:param imgIds (int array) : get question ids for given imgs
|
| 83 |
+
quesTypes (str array) : get question ids for given question types
|
| 84 |
+
ansTypes (str array) : get question ids for given answer types
|
| 85 |
+
:return: ids (int array) : integer array of question ids
|
| 86 |
+
"""
|
| 87 |
+
imgIds = imgIds if type(imgIds) == list else [imgIds]
|
| 88 |
+
quesTypes = quesTypes if type(quesTypes) == list else [quesTypes]
|
| 89 |
+
ansTypes = ansTypes if type(ansTypes) == list else [ansTypes]
|
| 90 |
+
|
| 91 |
+
if len(imgIds) == len(quesTypes) == len(ansTypes) == 0:
|
| 92 |
+
anns = self.dataset["annotations"]
|
| 93 |
+
else:
|
| 94 |
+
if not len(imgIds) == 0:
|
| 95 |
+
anns = sum(
|
| 96 |
+
[self.imgToQA[imgId] for imgId in imgIds if imgId in self.imgToQA],
|
| 97 |
+
[],
|
| 98 |
+
)
|
| 99 |
+
else:
|
| 100 |
+
anns = self.dataset["annotations"]
|
| 101 |
+
anns = (
|
| 102 |
+
anns
|
| 103 |
+
if len(quesTypes) == 0
|
| 104 |
+
else [ann for ann in anns if ann["question_type"] in quesTypes]
|
| 105 |
+
)
|
| 106 |
+
anns = (
|
| 107 |
+
anns
|
| 108 |
+
if len(ansTypes) == 0
|
| 109 |
+
else [ann for ann in anns if ann["answer_type"] in ansTypes]
|
| 110 |
+
)
|
| 111 |
+
ids = [ann["question_id"] for ann in anns]
|
| 112 |
+
return ids
|
| 113 |
+
|
| 114 |
+
def getImgIds(self, quesIds=[], quesTypes=[], ansTypes=[]):
|
| 115 |
+
"""
|
| 116 |
+
Get image ids that satisfy given filter conditions. default skips that filter
|
| 117 |
+
:param quesIds (int array) : get image ids for given question ids
|
| 118 |
+
quesTypes (str array) : get image ids for given question types
|
| 119 |
+
ansTypes (str array) : get image ids for given answer types
|
| 120 |
+
:return: ids (int array) : integer array of image ids
|
| 121 |
+
"""
|
| 122 |
+
quesIds = quesIds if type(quesIds) == list else [quesIds]
|
| 123 |
+
quesTypes = quesTypes if type(quesTypes) == list else [quesTypes]
|
| 124 |
+
ansTypes = ansTypes if type(ansTypes) == list else [ansTypes]
|
| 125 |
+
|
| 126 |
+
if len(quesIds) == len(quesTypes) == len(ansTypes) == 0:
|
| 127 |
+
anns = self.dataset["annotations"]
|
| 128 |
+
else:
|
| 129 |
+
if not len(quesIds) == 0:
|
| 130 |
+
anns = sum(
|
| 131 |
+
[self.qa[quesId] for quesId in quesIds if quesId in self.qa], []
|
| 132 |
+
)
|
| 133 |
+
else:
|
| 134 |
+
anns = self.dataset["annotations"]
|
| 135 |
+
anns = (
|
| 136 |
+
anns
|
| 137 |
+
if len(quesTypes) == 0
|
| 138 |
+
else [ann for ann in anns if ann["question_type"] in quesTypes]
|
| 139 |
+
)
|
| 140 |
+
anns = (
|
| 141 |
+
anns
|
| 142 |
+
if len(ansTypes) == 0
|
| 143 |
+
else [ann for ann in anns if ann["answer_type"] in ansTypes]
|
| 144 |
+
)
|
| 145 |
+
ids = [ann["image_id"] for ann in anns]
|
| 146 |
+
return ids
|
| 147 |
+
|
| 148 |
+
def loadQA(self, ids=[]):
|
| 149 |
+
"""
|
| 150 |
+
Load questions and answers with the specified question ids.
|
| 151 |
+
:param ids (int array) : integer ids specifying question ids
|
| 152 |
+
:return: qa (object array) : loaded qa objects
|
| 153 |
+
"""
|
| 154 |
+
if type(ids) == list:
|
| 155 |
+
return [self.qa[id] for id in ids]
|
| 156 |
+
elif type(ids) == int:
|
| 157 |
+
return [self.qa[ids]]
|
| 158 |
+
|
| 159 |
+
def showQA(self, anns):
|
| 160 |
+
"""
|
| 161 |
+
Display the specified annotations.
|
| 162 |
+
:param anns (array of object): annotations to display
|
| 163 |
+
:return: None
|
| 164 |
+
"""
|
| 165 |
+
if len(anns) == 0:
|
| 166 |
+
return 0
|
| 167 |
+
for ann in anns:
|
| 168 |
+
quesId = ann["question_id"]
|
| 169 |
+
print("Question: %s" % (self.qqa[quesId]["question"]))
|
| 170 |
+
for ans in ann["answers"]:
|
| 171 |
+
print("Answer %d: %s" % (ans["answer_id"], ans["answer"]))
|
| 172 |
+
|
| 173 |
+
def loadRes(self, resFile, quesFile):
|
| 174 |
+
"""
|
| 175 |
+
Load result file and return a result object.
|
| 176 |
+
:param resFile (str) : file name of result file
|
| 177 |
+
:return: res (obj) : result api object
|
| 178 |
+
"""
|
| 179 |
+
res = VQA()
|
| 180 |
+
res.questions = json.load(open(quesFile))
|
| 181 |
+
res.dataset["info"] = copy.deepcopy(self.questions["info"])
|
| 182 |
+
res.dataset["task_type"] = copy.deepcopy(self.questions["task_type"])
|
| 183 |
+
res.dataset["data_type"] = copy.deepcopy(self.questions["data_type"])
|
| 184 |
+
res.dataset["data_subtype"] = copy.deepcopy(self.questions["data_subtype"])
|
| 185 |
+
res.dataset["license"] = copy.deepcopy(self.questions["license"])
|
| 186 |
+
|
| 187 |
+
print("Loading and preparing results... ")
|
| 188 |
+
time_t = datetime.datetime.utcnow()
|
| 189 |
+
anns = json.load(open(resFile))
|
| 190 |
+
assert type(anns) == list, "results is not an array of objects"
|
| 191 |
+
annsQuesIds = [ann["question_id"] for ann in anns]
|
| 192 |
+
assert set(annsQuesIds) == set(
|
| 193 |
+
self.getQuesIds()
|
| 194 |
+
), "Results do not correspond to current VQA set. Either the results do not have predictions for all question ids in annotation file or there is atleast one question id that does not belong to the question ids in the annotation file."
|
| 195 |
+
for ann in anns:
|
| 196 |
+
quesId = ann["question_id"]
|
| 197 |
+
if res.dataset["task_type"] == "Multiple Choice":
|
| 198 |
+
assert (
|
| 199 |
+
ann["answer"] in self.qqa[quesId]["multiple_choices"]
|
| 200 |
+
), "predicted answer is not one of the multiple choices"
|
| 201 |
+
qaAnn = self.qa[quesId]
|
| 202 |
+
ann["image_id"] = qaAnn["image_id"]
|
| 203 |
+
ann["question_type"] = qaAnn["question_type"]
|
| 204 |
+
ann["answer_type"] = qaAnn["answer_type"]
|
| 205 |
+
print(
|
| 206 |
+
"DONE (t=%0.2fs)" % ((datetime.datetime.utcnow() - time_t).total_seconds())
|
| 207 |
+
)
|
| 208 |
+
|
| 209 |
+
res.dataset["annotations"] = anns
|
| 210 |
+
res.createIndex()
|
| 211 |
+
return res
|
LAVIS-main/lavis/common/vqa_tools/vqa_eval.py
ADDED
|
@@ -0,0 +1,324 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Copyright (c) 2022, salesforce.com, inc.
|
| 3 |
+
All rights reserved.
|
| 4 |
+
SPDX-License-Identifier: BSD-3-Clause
|
| 5 |
+
For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 6 |
+
"""
|
| 7 |
+
|
| 8 |
+
# coding=utf-8
|
| 9 |
+
|
| 10 |
+
__author__ = "aagrawal"
|
| 11 |
+
|
| 12 |
+
# This code is based on the code written by Tsung-Yi Lin for MSCOCO Python API available at the following link:
|
| 13 |
+
# (https://github.com/tylin/coco-caption/blob/master/pycocoevalcap/eval.py).
|
| 14 |
+
import sys
|
| 15 |
+
import re
|
| 16 |
+
|
| 17 |
+
|
| 18 |
+
class VQAEval:
|
| 19 |
+
def __init__(self, vqa=None, vqaRes=None, n=2):
|
| 20 |
+
self.n = n
|
| 21 |
+
self.accuracy = {}
|
| 22 |
+
self.evalQA = {}
|
| 23 |
+
self.evalQuesType = {}
|
| 24 |
+
self.evalAnsType = {}
|
| 25 |
+
self.vqa = vqa
|
| 26 |
+
self.vqaRes = vqaRes
|
| 27 |
+
if vqa is not None:
|
| 28 |
+
self.params = {"question_id": vqa.getQuesIds()}
|
| 29 |
+
self.contractions = {
|
| 30 |
+
"aint": "ain't",
|
| 31 |
+
"arent": "aren't",
|
| 32 |
+
"cant": "can't",
|
| 33 |
+
"couldve": "could've",
|
| 34 |
+
"couldnt": "couldn't",
|
| 35 |
+
"couldn'tve": "couldn't've",
|
| 36 |
+
"couldnt've": "couldn't've",
|
| 37 |
+
"didnt": "didn't",
|
| 38 |
+
"doesnt": "doesn't",
|
| 39 |
+
"dont": "don't",
|
| 40 |
+
"hadnt": "hadn't",
|
| 41 |
+
"hadnt've": "hadn't've",
|
| 42 |
+
"hadn'tve": "hadn't've",
|
| 43 |
+
"hasnt": "hasn't",
|
| 44 |
+
"havent": "haven't",
|
| 45 |
+
"hed": "he'd",
|
| 46 |
+
"hed've": "he'd've",
|
| 47 |
+
"he'dve": "he'd've",
|
| 48 |
+
"hes": "he's",
|
| 49 |
+
"howd": "how'd",
|
| 50 |
+
"howll": "how'll",
|
| 51 |
+
"hows": "how's",
|
| 52 |
+
"Id've": "I'd've",
|
| 53 |
+
"I'dve": "I'd've",
|
| 54 |
+
"Im": "I'm",
|
| 55 |
+
"Ive": "I've",
|
| 56 |
+
"isnt": "isn't",
|
| 57 |
+
"itd": "it'd",
|
| 58 |
+
"itd've": "it'd've",
|
| 59 |
+
"it'dve": "it'd've",
|
| 60 |
+
"itll": "it'll",
|
| 61 |
+
"let's": "let's",
|
| 62 |
+
"maam": "ma'am",
|
| 63 |
+
"mightnt": "mightn't",
|
| 64 |
+
"mightnt've": "mightn't've",
|
| 65 |
+
"mightn'tve": "mightn't've",
|
| 66 |
+
"mightve": "might've",
|
| 67 |
+
"mustnt": "mustn't",
|
| 68 |
+
"mustve": "must've",
|
| 69 |
+
"neednt": "needn't",
|
| 70 |
+
"notve": "not've",
|
| 71 |
+
"oclock": "o'clock",
|
| 72 |
+
"oughtnt": "oughtn't",
|
| 73 |
+
"ow's'at": "'ow's'at",
|
| 74 |
+
"'ows'at": "'ow's'at",
|
| 75 |
+
"'ow'sat": "'ow's'at",
|
| 76 |
+
"shant": "shan't",
|
| 77 |
+
"shed've": "she'd've",
|
| 78 |
+
"she'dve": "she'd've",
|
| 79 |
+
"she's": "she's",
|
| 80 |
+
"shouldve": "should've",
|
| 81 |
+
"shouldnt": "shouldn't",
|
| 82 |
+
"shouldnt've": "shouldn't've",
|
| 83 |
+
"shouldn'tve": "shouldn't've",
|
| 84 |
+
"somebody'd": "somebodyd",
|
| 85 |
+
"somebodyd've": "somebody'd've",
|
| 86 |
+
"somebody'dve": "somebody'd've",
|
| 87 |
+
"somebodyll": "somebody'll",
|
| 88 |
+
"somebodys": "somebody's",
|
| 89 |
+
"someoned": "someone'd",
|
| 90 |
+
"someoned've": "someone'd've",
|
| 91 |
+
"someone'dve": "someone'd've",
|
| 92 |
+
"someonell": "someone'll",
|
| 93 |
+
"someones": "someone's",
|
| 94 |
+
"somethingd": "something'd",
|
| 95 |
+
"somethingd've": "something'd've",
|
| 96 |
+
"something'dve": "something'd've",
|
| 97 |
+
"somethingll": "something'll",
|
| 98 |
+
"thats": "that's",
|
| 99 |
+
"thered": "there'd",
|
| 100 |
+
"thered've": "there'd've",
|
| 101 |
+
"there'dve": "there'd've",
|
| 102 |
+
"therere": "there're",
|
| 103 |
+
"theres": "there's",
|
| 104 |
+
"theyd": "they'd",
|
| 105 |
+
"theyd've": "they'd've",
|
| 106 |
+
"they'dve": "they'd've",
|
| 107 |
+
"theyll": "they'll",
|
| 108 |
+
"theyre": "they're",
|
| 109 |
+
"theyve": "they've",
|
| 110 |
+
"twas": "'twas",
|
| 111 |
+
"wasnt": "wasn't",
|
| 112 |
+
"wed've": "we'd've",
|
| 113 |
+
"we'dve": "we'd've",
|
| 114 |
+
"weve": "we've",
|
| 115 |
+
"werent": "weren't",
|
| 116 |
+
"whatll": "what'll",
|
| 117 |
+
"whatre": "what're",
|
| 118 |
+
"whats": "what's",
|
| 119 |
+
"whatve": "what've",
|
| 120 |
+
"whens": "when's",
|
| 121 |
+
"whered": "where'd",
|
| 122 |
+
"wheres": "where's",
|
| 123 |
+
"whereve": "where've",
|
| 124 |
+
"whod": "who'd",
|
| 125 |
+
"whod've": "who'd've",
|
| 126 |
+
"who'dve": "who'd've",
|
| 127 |
+
"wholl": "who'll",
|
| 128 |
+
"whos": "who's",
|
| 129 |
+
"whove": "who've",
|
| 130 |
+
"whyll": "why'll",
|
| 131 |
+
"whyre": "why're",
|
| 132 |
+
"whys": "why's",
|
| 133 |
+
"wont": "won't",
|
| 134 |
+
"wouldve": "would've",
|
| 135 |
+
"wouldnt": "wouldn't",
|
| 136 |
+
"wouldnt've": "wouldn't've",
|
| 137 |
+
"wouldn'tve": "wouldn't've",
|
| 138 |
+
"yall": "y'all",
|
| 139 |
+
"yall'll": "y'all'll",
|
| 140 |
+
"y'allll": "y'all'll",
|
| 141 |
+
"yall'd've": "y'all'd've",
|
| 142 |
+
"y'alld've": "y'all'd've",
|
| 143 |
+
"y'all'dve": "y'all'd've",
|
| 144 |
+
"youd": "you'd",
|
| 145 |
+
"youd've": "you'd've",
|
| 146 |
+
"you'dve": "you'd've",
|
| 147 |
+
"youll": "you'll",
|
| 148 |
+
"youre": "you're",
|
| 149 |
+
"youve": "you've",
|
| 150 |
+
}
|
| 151 |
+
self.manualMap = {
|
| 152 |
+
"none": "0",
|
| 153 |
+
"zero": "0",
|
| 154 |
+
"one": "1",
|
| 155 |
+
"two": "2",
|
| 156 |
+
"three": "3",
|
| 157 |
+
"four": "4",
|
| 158 |
+
"five": "5",
|
| 159 |
+
"six": "6",
|
| 160 |
+
"seven": "7",
|
| 161 |
+
"eight": "8",
|
| 162 |
+
"nine": "9",
|
| 163 |
+
"ten": "10",
|
| 164 |
+
}
|
| 165 |
+
self.articles = ["a", "an", "the"]
|
| 166 |
+
|
| 167 |
+
self.periodStrip = re.compile("(?!<=\d)(\.)(?!\d)")
|
| 168 |
+
self.commaStrip = re.compile("(\d)(,)(\d)")
|
| 169 |
+
self.punct = [
|
| 170 |
+
";",
|
| 171 |
+
r"/",
|
| 172 |
+
"[",
|
| 173 |
+
"]",
|
| 174 |
+
'"',
|
| 175 |
+
"{",
|
| 176 |
+
"}",
|
| 177 |
+
"(",
|
| 178 |
+
")",
|
| 179 |
+
"=",
|
| 180 |
+
"+",
|
| 181 |
+
"\\",
|
| 182 |
+
"_",
|
| 183 |
+
"-",
|
| 184 |
+
">",
|
| 185 |
+
"<",
|
| 186 |
+
"@",
|
| 187 |
+
"`",
|
| 188 |
+
",",
|
| 189 |
+
"?",
|
| 190 |
+
"!",
|
| 191 |
+
]
|
| 192 |
+
|
| 193 |
+
def evaluate(self, quesIds=None):
|
| 194 |
+
if quesIds == None:
|
| 195 |
+
quesIds = [quesId for quesId in self.params["question_id"]]
|
| 196 |
+
gts = {}
|
| 197 |
+
res = {}
|
| 198 |
+
for quesId in quesIds:
|
| 199 |
+
gts[quesId] = self.vqa.qa[quesId]
|
| 200 |
+
res[quesId] = self.vqaRes.qa[quesId]
|
| 201 |
+
|
| 202 |
+
# =================================================
|
| 203 |
+
# Compute accuracy
|
| 204 |
+
# =================================================
|
| 205 |
+
accQA = []
|
| 206 |
+
accQuesType = {}
|
| 207 |
+
accAnsType = {}
|
| 208 |
+
print("computing accuracy")
|
| 209 |
+
step = 0
|
| 210 |
+
for quesId in quesIds:
|
| 211 |
+
resAns = res[quesId]["answer"]
|
| 212 |
+
resAns = resAns.replace("\n", " ")
|
| 213 |
+
resAns = resAns.replace("\t", " ")
|
| 214 |
+
resAns = resAns.strip()
|
| 215 |
+
resAns = self.processPunctuation(resAns)
|
| 216 |
+
resAns = self.processDigitArticle(resAns)
|
| 217 |
+
gtAcc = []
|
| 218 |
+
gtAnswers = [ans["answer"] for ans in gts[quesId]["answers"]]
|
| 219 |
+
if len(set(gtAnswers)) > 1:
|
| 220 |
+
for ansDic in gts[quesId]["answers"]:
|
| 221 |
+
ansDic["answer"] = self.processPunctuation(ansDic["answer"])
|
| 222 |
+
for gtAnsDatum in gts[quesId]["answers"]:
|
| 223 |
+
otherGTAns = [
|
| 224 |
+
item for item in gts[quesId]["answers"] if item != gtAnsDatum
|
| 225 |
+
]
|
| 226 |
+
matchingAns = [item for item in otherGTAns if item["answer"] == resAns]
|
| 227 |
+
acc = min(1, float(len(matchingAns)) / 3)
|
| 228 |
+
gtAcc.append(acc)
|
| 229 |
+
quesType = gts[quesId]["question_type"]
|
| 230 |
+
ansType = gts[quesId]["answer_type"]
|
| 231 |
+
avgGTAcc = float(sum(gtAcc)) / len(gtAcc)
|
| 232 |
+
accQA.append(avgGTAcc)
|
| 233 |
+
if quesType not in accQuesType:
|
| 234 |
+
accQuesType[quesType] = []
|
| 235 |
+
accQuesType[quesType].append(avgGTAcc)
|
| 236 |
+
if ansType not in accAnsType:
|
| 237 |
+
accAnsType[ansType] = []
|
| 238 |
+
accAnsType[ansType].append(avgGTAcc)
|
| 239 |
+
self.setEvalQA(quesId, avgGTAcc)
|
| 240 |
+
self.setEvalQuesType(quesId, quesType, avgGTAcc)
|
| 241 |
+
self.setEvalAnsType(quesId, ansType, avgGTAcc)
|
| 242 |
+
if step % 100 == 0:
|
| 243 |
+
self.updateProgress(step / float(len(quesIds)))
|
| 244 |
+
step = step + 1
|
| 245 |
+
|
| 246 |
+
self.setAccuracy(accQA, accQuesType, accAnsType)
|
| 247 |
+
print("Done computing accuracy")
|
| 248 |
+
|
| 249 |
+
def processPunctuation(self, inText):
|
| 250 |
+
outText = inText
|
| 251 |
+
for p in self.punct:
|
| 252 |
+
if (p + " " in inText or " " + p in inText) or (
|
| 253 |
+
re.search(self.commaStrip, inText) != None
|
| 254 |
+
):
|
| 255 |
+
outText = outText.replace(p, "")
|
| 256 |
+
else:
|
| 257 |
+
outText = outText.replace(p, " ")
|
| 258 |
+
outText = self.periodStrip.sub("", outText, re.UNICODE)
|
| 259 |
+
return outText
|
| 260 |
+
|
| 261 |
+
def processDigitArticle(self, inText):
|
| 262 |
+
outText = []
|
| 263 |
+
tempText = inText.lower().split()
|
| 264 |
+
for word in tempText:
|
| 265 |
+
word = self.manualMap.setdefault(word, word)
|
| 266 |
+
if word not in self.articles:
|
| 267 |
+
outText.append(word)
|
| 268 |
+
else:
|
| 269 |
+
pass
|
| 270 |
+
for wordId, word in enumerate(outText):
|
| 271 |
+
if word in self.contractions:
|
| 272 |
+
outText[wordId] = self.contractions[word]
|
| 273 |
+
outText = " ".join(outText)
|
| 274 |
+
return outText
|
| 275 |
+
|
| 276 |
+
def setAccuracy(self, accQA, accQuesType, accAnsType):
|
| 277 |
+
self.accuracy["overall"] = round(100 * float(sum(accQA)) / len(accQA), self.n)
|
| 278 |
+
self.accuracy["perQuestionType"] = {
|
| 279 |
+
quesType: round(
|
| 280 |
+
100 * float(sum(accQuesType[quesType])) / len(accQuesType[quesType]),
|
| 281 |
+
self.n,
|
| 282 |
+
)
|
| 283 |
+
for quesType in accQuesType
|
| 284 |
+
}
|
| 285 |
+
self.accuracy["perAnswerType"] = {
|
| 286 |
+
ansType: round(
|
| 287 |
+
100 * float(sum(accAnsType[ansType])) / len(accAnsType[ansType]), self.n
|
| 288 |
+
)
|
| 289 |
+
for ansType in accAnsType
|
| 290 |
+
}
|
| 291 |
+
|
| 292 |
+
def setEvalQA(self, quesId, acc):
|
| 293 |
+
self.evalQA[quesId] = round(100 * acc, self.n)
|
| 294 |
+
|
| 295 |
+
def setEvalQuesType(self, quesId, quesType, acc):
|
| 296 |
+
if quesType not in self.evalQuesType:
|
| 297 |
+
self.evalQuesType[quesType] = {}
|
| 298 |
+
self.evalQuesType[quesType][quesId] = round(100 * acc, self.n)
|
| 299 |
+
|
| 300 |
+
def setEvalAnsType(self, quesId, ansType, acc):
|
| 301 |
+
if ansType not in self.evalAnsType:
|
| 302 |
+
self.evalAnsType[ansType] = {}
|
| 303 |
+
self.evalAnsType[ansType][quesId] = round(100 * acc, self.n)
|
| 304 |
+
|
| 305 |
+
def updateProgress(self, progress):
|
| 306 |
+
barLength = 20
|
| 307 |
+
status = ""
|
| 308 |
+
if isinstance(progress, int):
|
| 309 |
+
progress = float(progress)
|
| 310 |
+
if not isinstance(progress, float):
|
| 311 |
+
progress = 0
|
| 312 |
+
status = "error: progress var must be float\r\n"
|
| 313 |
+
if progress < 0:
|
| 314 |
+
progress = 0
|
| 315 |
+
status = "Halt...\r\n"
|
| 316 |
+
if progress >= 1:
|
| 317 |
+
progress = 1
|
| 318 |
+
status = "Done...\r\n"
|
| 319 |
+
block = int(round(barLength * progress))
|
| 320 |
+
text = "\rFinshed Percent: [{0}] {1}% {2}".format(
|
| 321 |
+
"#" * block + "-" * (barLength - block), int(progress * 100), status
|
| 322 |
+
)
|
| 323 |
+
sys.stdout.write(text)
|
| 324 |
+
sys.stdout.flush()
|
LAVIS-main/lavis/configs/datasets/aokvqa/defaults.yaml
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
aok_vqa:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url:
|
| 16 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_train.json
|
| 17 |
+
storage:
|
| 18 |
+
- aokvqa/annotations/aokvqa_v1p0_train.json
|
| 19 |
+
val:
|
| 20 |
+
url:
|
| 21 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_val.json
|
| 22 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
|
| 23 |
+
storage:
|
| 24 |
+
- aokvqa/annotations/aokvqa_v1p0_val.json
|
| 25 |
+
- aokvqa/annotations/specialized_vocab_train_lavis.json
|
| 26 |
+
# - aokvqa/annotations/large_vocab_train_lavis.json
|
| 27 |
+
test:
|
| 28 |
+
url:
|
| 29 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_test.json
|
| 30 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
|
| 31 |
+
storage:
|
| 32 |
+
- aokvqa/annotations/aokvqa_v1p0_test.json
|
| 33 |
+
- aokvqa/annotations/specialized_vocab_train_lavis.json
|
| 34 |
+
images:
|
| 35 |
+
storage: coco/images/
|
LAVIS-main/lavis/configs/datasets/aokvqa/defaults_instruct.yaml
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
aok_vqa_instruct:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: "clip_image_train"
|
| 14 |
+
image_size: 224
|
| 15 |
+
eval:
|
| 16 |
+
name: "clip_image_eval"
|
| 17 |
+
image_size: 224
|
| 18 |
+
|
| 19 |
+
text_processor:
|
| 20 |
+
train:
|
| 21 |
+
name: blip_instruction
|
| 22 |
+
modality: image
|
| 23 |
+
task: qa
|
| 24 |
+
eval:
|
| 25 |
+
name: blip_question
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 29 |
+
annotations:
|
| 30 |
+
train:
|
| 31 |
+
url:
|
| 32 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_train.json
|
| 33 |
+
storage:
|
| 34 |
+
- aokvqa/annotations/aokvqa_v1p0_train.json
|
| 35 |
+
# val:
|
| 36 |
+
# url:
|
| 37 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_val.json
|
| 38 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
|
| 39 |
+
# storage:
|
| 40 |
+
# - aokvqa/annotations/aokvqa_v1p0_val.json
|
| 41 |
+
# - aokvqa/annotations/specialized_vocab_train_lavis.json
|
| 42 |
+
# # - aokvqa/annotations/large_vocab_train_lavis.json
|
| 43 |
+
# test:
|
| 44 |
+
# url:
|
| 45 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/aokvqa_v1p0_test.json
|
| 46 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/aokvqa/specialized_vocab_train.json
|
| 47 |
+
# storage:
|
| 48 |
+
# - aokvqa/annotations/aokvqa_v1p0_test.json
|
| 49 |
+
# - aokvqa/annotations/specialized_vocab_train_lavis.json
|
| 50 |
+
images:
|
| 51 |
+
# storage: /coco/images
|
| 52 |
+
storage: /export/share/datasets/vision/coco/images
|
LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap.yaml
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
audiocaps_mm_caption: # name of the dataset builder
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
sampling_rate: 16000
|
| 12 |
+
eval:
|
| 13 |
+
name: beats_audio
|
| 14 |
+
sampling_rate: 16000
|
| 15 |
+
|
| 16 |
+
text_processor:
|
| 17 |
+
train:
|
| 18 |
+
name: "blip_instruction"
|
| 19 |
+
modality: audio
|
| 20 |
+
task: caption
|
| 21 |
+
eval:
|
| 22 |
+
name: "blip_caption"
|
| 23 |
+
|
| 24 |
+
data_type: [audio]
|
| 25 |
+
|
| 26 |
+
build_info:
|
| 27 |
+
kwargs:
|
| 28 |
+
missing_ids: [2sh7ZkazyO8, 966jA2-z0mQ, 52RlolYyjAE, HVAc9hm4jjk, 8lPjqvYWNyM, eXgPnnE3TuQ]
|
| 29 |
+
annotations:
|
| 30 |
+
train:
|
| 31 |
+
url:
|
| 32 |
+
- https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/train.csv
|
| 33 |
+
storage:
|
| 34 |
+
- audiocaps/annotations/train.csv
|
| 35 |
+
|
| 36 |
+
val:
|
| 37 |
+
url:
|
| 38 |
+
- https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/val.csv
|
| 39 |
+
storage:
|
| 40 |
+
- audiocaps/annotations/val.csv
|
| 41 |
+
|
| 42 |
+
test:
|
| 43 |
+
url:
|
| 44 |
+
- https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/test.csv
|
| 45 |
+
storage:
|
| 46 |
+
- audiocaps/annotations/test.csv
|
| 47 |
+
|
| 48 |
+
audio:
|
| 49 |
+
storage: /export/einstein-vision/audio_datasets/audiocaps/AUDIOCAPS_32000Hz/audio
|
LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
audiocaps_mm_caption_instruct: # name of the dataset builder
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
sampling_rate: 16000
|
| 12 |
+
eval:
|
| 13 |
+
name: beats_audio
|
| 14 |
+
sampling_rate: 16000
|
| 15 |
+
|
| 16 |
+
text_processor:
|
| 17 |
+
train:
|
| 18 |
+
name: "blip_instruction"
|
| 19 |
+
modality: audio
|
| 20 |
+
task: caption
|
| 21 |
+
eval:
|
| 22 |
+
name: "blip_caption"
|
| 23 |
+
|
| 24 |
+
data_type: [audio]
|
| 25 |
+
|
| 26 |
+
missing_ids: [2sh7ZkazyO8, 966jA2-z0mQ, 52RlolYyjAE, HVAc9hm4jjk, 8lPjqvYWNyM, eXgPnnE3TuQ]
|
| 27 |
+
|
| 28 |
+
build_info:
|
| 29 |
+
kwargs:
|
| 30 |
+
cached: False
|
| 31 |
+
cached_dir: /export/einstein-vision/audio_datasets/audiocaps/beats_features
|
| 32 |
+
annotations:
|
| 33 |
+
train:
|
| 34 |
+
url:
|
| 35 |
+
- https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/train.csv
|
| 36 |
+
storage:
|
| 37 |
+
- audiocaps/annotations/train.csv
|
| 38 |
+
|
| 39 |
+
# val:
|
| 40 |
+
# url:
|
| 41 |
+
# - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/val.csv
|
| 42 |
+
# storage:
|
| 43 |
+
# - audiocaps/annotation/val.csv
|
| 44 |
+
|
| 45 |
+
# test:
|
| 46 |
+
# url:
|
| 47 |
+
# - https://raw.githubusercontent.com/cdjkim/audiocaps/master/dataset/test.csv
|
| 48 |
+
# storage:
|
| 49 |
+
# - /export/einstein-vision/audio_datasets/audiocaps/dataset/test.csv
|
| 50 |
+
|
| 51 |
+
audio:
|
| 52 |
+
storage: /export/einstein-vision/audio_datasets/audiocaps/AUDIOCAPS_32000Hz/audio
|
LAVIS-main/lavis/configs/datasets/audiocaps/defaults_mm_qa.yaml
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
audiocaps_mm_qa: # name of the dataset builder
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
sampling_rate: 16000
|
| 12 |
+
eval:
|
| 13 |
+
name: beats_audio
|
| 14 |
+
sampling_rate: 16000
|
| 15 |
+
is_eval: True
|
| 16 |
+
|
| 17 |
+
text_processor:
|
| 18 |
+
train:
|
| 19 |
+
name: "blip_instruction"
|
| 20 |
+
modality: audio
|
| 21 |
+
task: qa
|
| 22 |
+
eval:
|
| 23 |
+
name: "blip_question"
|
| 24 |
+
|
| 25 |
+
data_type: [audio]
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
kwargs:
|
| 29 |
+
cached: False
|
| 30 |
+
# add_binary: True
|
| 31 |
+
cached_dir: /export/einstein-vision/audio_datasets/audiocaps/beats_features
|
| 32 |
+
missing_ids: [2sh7ZkazyO8, 966jA2-z0mQ, 52RlolYyjAE, HVAc9hm4jjk, 8lPjqvYWNyM, eXgPnnE3TuQ]
|
| 33 |
+
annotations:
|
| 34 |
+
train:
|
| 35 |
+
url:
|
| 36 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/audiocaps/audio_qa_final_train.csv
|
| 37 |
+
# - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_train.csv
|
| 38 |
+
storage:
|
| 39 |
+
- audiocaps_qa/annotations/train.csv
|
| 40 |
+
# - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_train.csv
|
| 41 |
+
|
| 42 |
+
# val:
|
| 43 |
+
# url:
|
| 44 |
+
# # - https://storage.googleapis.com/sfr-xinstructblip-data-research/data/audiocaps/audio_qa_final_val.csv
|
| 45 |
+
# - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_val.csv
|
| 46 |
+
# storage:
|
| 47 |
+
# # - audiocaps_qa/annotations/val.csv
|
| 48 |
+
# - /export/home/LAVIS-xgen_mm/projects/xinstructblip/data_aug/audio_qa_data/audio_qa_final_val.csv
|
| 49 |
+
|
| 50 |
+
audio:
|
| 51 |
+
storage: /export/einstein-vision/audio_datasets/audiocaps/AUDIOCAPS_32000Hz/audio
|
LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap.yaml
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
audioset_mm_caption: # 14141
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
sampling_rate: 16000
|
| 12 |
+
eval:
|
| 13 |
+
name: beats_audio
|
| 14 |
+
sampling_rate: 16000
|
| 15 |
+
is_eval: False
|
| 16 |
+
|
| 17 |
+
text_processor:
|
| 18 |
+
train:
|
| 19 |
+
name: blip_instruction
|
| 20 |
+
modality: audio
|
| 21 |
+
task: classification
|
| 22 |
+
eval:
|
| 23 |
+
name: blip_caption
|
| 24 |
+
|
| 25 |
+
data_type: [audio]
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
annotations:
|
| 29 |
+
train:
|
| 30 |
+
url:
|
| 31 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data//audioset/balanced_train_clean.csv
|
| 32 |
+
# - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
|
| 33 |
+
- http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
|
| 34 |
+
storage:
|
| 35 |
+
- audioset/balanced_train_clean.csv
|
| 36 |
+
# - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
|
| 37 |
+
- audioset/annotations/class_labels_indices.csv
|
| 38 |
+
|
| 39 |
+
# val:
|
| 40 |
+
# url:
|
| 41 |
+
# - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/eval_segments.csv
|
| 42 |
+
# - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
|
| 43 |
+
# storage:
|
| 44 |
+
# - audioset/annotations/eval_segments.csv
|
| 45 |
+
# - audioset/annotations/class_labels_indices.csv
|
| 46 |
+
audio:
|
| 47 |
+
storage: /export/einstein-vision/audio_datasets/AudioSet/all_audio
|
LAVIS-main/lavis/configs/datasets/audioset/defaults_mm_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
audioset_mm_caption_instruct: # 14141
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
sampling_rate: 16000
|
| 12 |
+
eval:
|
| 13 |
+
name: beats_audio
|
| 14 |
+
sampling_rate: 16000
|
| 15 |
+
is_eval: False
|
| 16 |
+
|
| 17 |
+
text_processor:
|
| 18 |
+
train:
|
| 19 |
+
name: blip_instruction
|
| 20 |
+
modality: audio
|
| 21 |
+
task: classification
|
| 22 |
+
eval:
|
| 23 |
+
name: blip_caption
|
| 24 |
+
|
| 25 |
+
data_type: [audio]
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
annotations:
|
| 29 |
+
train:
|
| 30 |
+
url:
|
| 31 |
+
# - https://storage.googleapis.com/sfr-xinstructblip-data-research/data//audioset/balanced_train_clean.csv
|
| 32 |
+
- /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
|
| 33 |
+
- http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
|
| 34 |
+
storage:
|
| 35 |
+
- audioset/annotations/balanced_train_clean.csv
|
| 36 |
+
# - /export/home/LAVIS-xgen_mm/lavis/configs/datasets/audioset/balanced_train_clean.csv
|
| 37 |
+
- audioset/annotations/class_labels_indices.csv
|
| 38 |
+
|
| 39 |
+
# val:
|
| 40 |
+
# url:
|
| 41 |
+
# - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/eval_segments.csv
|
| 42 |
+
# - http://storage.googleapis.com/us_audioset/youtube_corpus/v1/csv/class_labels_indices.csv
|
| 43 |
+
# storage:
|
| 44 |
+
# - audioset/annotations/eval_segments.csv
|
| 45 |
+
# - audioset/annotations/class_labels_indices.csv
|
| 46 |
+
|
| 47 |
+
audio:
|
| 48 |
+
storage: /export/einstein-vision/audio_datasets/AudioSet/all_audio
|
LAVIS-main/lavis/configs/datasets/avsd/defaults_dial.yaml
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
avsd_dialogue: # name of the dataset builder
|
| 8 |
+
dataset_card: dataset_card/avsd_dialogue.md
|
| 9 |
+
data_type: features #extracted features of videos (I3D, VGGish) # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_train.json
|
| 16 |
+
storage: avsd/annotations/train.json
|
| 17 |
+
val:
|
| 18 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_val.json
|
| 19 |
+
storage: avsd/annotations/val.json
|
| 20 |
+
test:
|
| 21 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_test.json
|
| 22 |
+
storage: avsd/annotations/test.json
|
| 23 |
+
features:
|
| 24 |
+
storage: avsd/features/
|
LAVIS-main/lavis/configs/datasets/avsd/defaults_mm_dial_instruct.yaml
ADDED
|
@@ -0,0 +1,65 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
avsd_mm_dialogue_instruct: # name of the dataset builder
|
| 8 |
+
data_type: [video, audio]
|
| 9 |
+
|
| 10 |
+
audio_processor:
|
| 11 |
+
train:
|
| 12 |
+
name: beats_audio
|
| 13 |
+
sampling_rate: 16000
|
| 14 |
+
eval:
|
| 15 |
+
name: beats_audio
|
| 16 |
+
sampling_rate: 16000
|
| 17 |
+
|
| 18 |
+
video_processor:
|
| 19 |
+
train:
|
| 20 |
+
name: alpro_video_train
|
| 21 |
+
n_frms: 4
|
| 22 |
+
image_size: 224
|
| 23 |
+
min_scale: 0.9
|
| 24 |
+
max_scale: 1.0
|
| 25 |
+
full_video: True
|
| 26 |
+
eval:
|
| 27 |
+
name: alpro_video_eval
|
| 28 |
+
n_frms: 4
|
| 29 |
+
image_size: 224
|
| 30 |
+
min_scale: 0.9
|
| 31 |
+
max_scale: 1.0
|
| 32 |
+
full_video: True
|
| 33 |
+
|
| 34 |
+
text_processor:
|
| 35 |
+
train:
|
| 36 |
+
name: "blip_caption"
|
| 37 |
+
eval:
|
| 38 |
+
name: "blip_caption"
|
| 39 |
+
|
| 40 |
+
build_info:
|
| 41 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 42 |
+
annotations:
|
| 43 |
+
train:
|
| 44 |
+
url:
|
| 45 |
+
- https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_train.json
|
| 46 |
+
storage:
|
| 47 |
+
- avsd/annotations/train.json
|
| 48 |
+
val:
|
| 49 |
+
url:
|
| 50 |
+
- https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_val.json
|
| 51 |
+
storage:
|
| 52 |
+
- avsd/annotations/val.json
|
| 53 |
+
test:
|
| 54 |
+
url:
|
| 55 |
+
- https://storage.googleapis.com/sfr-vision-language-research/datasets/avsd_dstc7_test.json
|
| 56 |
+
storage:
|
| 57 |
+
- avsd/annotations/test.json
|
| 58 |
+
templates: null
|
| 59 |
+
|
| 60 |
+
audio:
|
| 61 |
+
storage: /export/video-language-dataset/data/charade/videos
|
| 62 |
+
|
| 63 |
+
video:
|
| 64 |
+
storage: /export/video-language-dataset/data/charade/videos
|
| 65 |
+
|
LAVIS-main/lavis/configs/datasets/blip_diffusion_datasets/defaults.yaml
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
blip_diffusion_finetune: # name of the dataset builder
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
images:
|
| 14 |
+
storage: ""
|
LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap.yaml
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
capfilt14m: # 13873136
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: "clip_image_train"
|
| 14 |
+
image_size: 224
|
| 15 |
+
text_processor:
|
| 16 |
+
train:
|
| 17 |
+
name: blip_caption
|
| 18 |
+
|
| 19 |
+
build_info:
|
| 20 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 21 |
+
annotations:
|
| 22 |
+
train:
|
| 23 |
+
url:
|
| 24 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/capfilt14m/annotation.json
|
| 25 |
+
# - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
|
| 26 |
+
storage:
|
| 27 |
+
- capfilt14m/annotations/annotation.json
|
| 28 |
+
# - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
|
| 29 |
+
images:
|
| 30 |
+
storage: /export/share/datasets/vision/coco/images
|
LAVIS-main/lavis/configs/datasets/capfilt14m/defaults_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
capfilt14m_instruct: # 13873136
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: "clip_image_train"
|
| 14 |
+
image_size: 224
|
| 15 |
+
|
| 16 |
+
text_processor:
|
| 17 |
+
train:
|
| 18 |
+
name: blip_instruction
|
| 19 |
+
modality: image
|
| 20 |
+
task: caption
|
| 21 |
+
|
| 22 |
+
build_info:
|
| 23 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 24 |
+
annotations:
|
| 25 |
+
train:
|
| 26 |
+
url:
|
| 27 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/capfilt14m/annotation.json
|
| 28 |
+
# - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
|
| 29 |
+
storage:
|
| 30 |
+
- capfilt14m/annotations/annotation.json
|
| 31 |
+
# - /export/share/datasets/vision_language/capfilt_14m_new/annotation.json
|
| 32 |
+
|
| 33 |
+
images:
|
| 34 |
+
storage: /export/share/datasets/vision/coco/images
|
LAVIS-main/lavis/configs/datasets/charade/defaults_cap.yaml
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
charade_caption: # name of the dataset builder
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: videos # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: alpro_video_train
|
| 14 |
+
n_frms: 4
|
| 15 |
+
image_size: 224
|
| 16 |
+
min_scale: 0.9
|
| 17 |
+
max_scale: 1.0
|
| 18 |
+
full_video: False
|
| 19 |
+
eval:
|
| 20 |
+
name: alpro_video_eval
|
| 21 |
+
n_frms: 4
|
| 22 |
+
image_size: 224
|
| 23 |
+
min_scale: 0.9
|
| 24 |
+
max_scale: 1.0
|
| 25 |
+
full_video: False
|
| 26 |
+
|
| 27 |
+
text_processor:
|
| 28 |
+
train:
|
| 29 |
+
name: blip_caption
|
| 30 |
+
eval:
|
| 31 |
+
name: blip_caption
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
build_info:
|
| 35 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 36 |
+
annotations:
|
| 37 |
+
train:
|
| 38 |
+
url:
|
| 39 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/train_lavis.json
|
| 40 |
+
# - /export/video-language-dataset/data/charade/train_lavis.json
|
| 41 |
+
storage:
|
| 42 |
+
- charade/annotations/train.json
|
| 43 |
+
# - /export/video-language-dataset/data/charade/train_lavis.json
|
| 44 |
+
val:
|
| 45 |
+
url:
|
| 46 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/val_lavis.json
|
| 47 |
+
# - /export/video-language-dataset/data/charade/val_lavis.json
|
| 48 |
+
storage:
|
| 49 |
+
- charade/annotations/val.json
|
| 50 |
+
# - /export/video-language-dataset/data/charade/val_lavis.json
|
| 51 |
+
videos:
|
| 52 |
+
storage: /export/video-language-dataset/data/charade/videos
|
LAVIS-main/lavis/configs/datasets/charade/defaults_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
charade_caption_instruct: # name of the dataset builder
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: videos # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: alpro_video_train
|
| 14 |
+
n_frms: 4
|
| 15 |
+
image_size: 224
|
| 16 |
+
min_scale: 0.9
|
| 17 |
+
max_scale: 1.0
|
| 18 |
+
full_video: False
|
| 19 |
+
eval:
|
| 20 |
+
name: alpro_video_eval
|
| 21 |
+
n_frms: 4
|
| 22 |
+
image_size: 224
|
| 23 |
+
min_scale: 0.9
|
| 24 |
+
max_scale: 1.0
|
| 25 |
+
full_video: False
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
text_processor:
|
| 29 |
+
train:
|
| 30 |
+
name: blip_instruction
|
| 31 |
+
modality: video
|
| 32 |
+
task: caption
|
| 33 |
+
eval:
|
| 34 |
+
name: blip_caption
|
| 35 |
+
|
| 36 |
+
build_info:
|
| 37 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 38 |
+
annotations:
|
| 39 |
+
train:
|
| 40 |
+
url:
|
| 41 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/train_lavis.json
|
| 42 |
+
# - /export/video-language-dataset/data/charade/train_lavis.json
|
| 43 |
+
storage:
|
| 44 |
+
- charade/annotations/train.json
|
| 45 |
+
# - /export/video-language-dataset/data/charade/train_lavis.json
|
| 46 |
+
val:
|
| 47 |
+
url:
|
| 48 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/charade/val_lavis.json
|
| 49 |
+
# - /export/video-language-dataset/data/charade/val_lavis.json
|
| 50 |
+
storage:
|
| 51 |
+
- charade/annotations/val.json
|
| 52 |
+
# - /export/video-language-dataset/data/charade/val_lavis.json
|
| 53 |
+
videos:
|
| 54 |
+
storage: /export/video-language-dataset/data/charade/videos
|
LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap.yaml
ADDED
|
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
clothov2: # name of the dataset builder
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
eval:
|
| 12 |
+
name: beats_audio
|
| 13 |
+
|
| 14 |
+
text_processor:
|
| 15 |
+
train:
|
| 16 |
+
name: "blip_caption"
|
| 17 |
+
eval:
|
| 18 |
+
name: "blip_caption"
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
data_type: [audio]
|
| 22 |
+
|
| 23 |
+
build_info:
|
| 24 |
+
kwargs:
|
| 25 |
+
clotho_root: /export/einstein-vision/audio_datasets/clothov2/
|
| 26 |
+
split: eval
|
| 27 |
+
|
| 28 |
+
annotations:
|
| 29 |
+
train:
|
| 30 |
+
url:
|
| 31 |
+
- https://zenodo.org/record/4783391/files/clotho_captions_development.csv
|
| 32 |
+
storage:
|
| 33 |
+
- clothov2/annotations/clotho_captions_development.csv
|
| 34 |
+
val:
|
| 35 |
+
url:
|
| 36 |
+
- https://zenodo.org/record/4783391/files/clotho_captions_evaluation.csv
|
| 37 |
+
storage:
|
| 38 |
+
- clothov2/annotations/clotho_captions_evaluation.csv
|
| 39 |
+
audio:
|
| 40 |
+
storage: /export/einstein-vision/audio_datasets/clothov2/CLOTHO_v2.1/clotho_audio_files/
|
| 41 |
+
|
LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
clothov2_instruct: # name of the dataset builder
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
eval:
|
| 12 |
+
name: beats_audio
|
| 13 |
+
|
| 14 |
+
text_processor:
|
| 15 |
+
train:
|
| 16 |
+
name: "blip_instruction"
|
| 17 |
+
modality: audio
|
| 18 |
+
task: caption
|
| 19 |
+
eval:
|
| 20 |
+
name: "blip_caption"
|
| 21 |
+
|
| 22 |
+
data_type: [audio]
|
| 23 |
+
|
| 24 |
+
build_info:
|
| 25 |
+
kwargs:
|
| 26 |
+
clotho_root: /export/einstein-vision/audio_datasets/clothov2/
|
| 27 |
+
split: eval
|
| 28 |
+
|
| 29 |
+
annotations:
|
| 30 |
+
train:
|
| 31 |
+
url:
|
| 32 |
+
- https://zenodo.org/record/4783391/files/clotho_captions_development.csv
|
| 33 |
+
storage:
|
| 34 |
+
- clothov2/annotations/clotho_captions_development.csv
|
| 35 |
+
val:
|
| 36 |
+
url:
|
| 37 |
+
- https://zenodo.org/record/4783391/files/clotho_captions_evaluation.csv
|
| 38 |
+
storage:
|
| 39 |
+
- clothov2/annotations/clotho_captions_evaluation.csv
|
| 40 |
+
audio:
|
| 41 |
+
storage: /export/einstein-vision/audio_datasets/clothov2/CLOTHO_v2.1/clotho_audio_files/
|
| 42 |
+
|
LAVIS-main/lavis/configs/datasets/clotho/defaults_mm_qa.yaml
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
clotho_qa: # name of the dataset builder
|
| 8 |
+
audio_processor:
|
| 9 |
+
train:
|
| 10 |
+
name: beats_audio
|
| 11 |
+
eval:
|
| 12 |
+
name: beats_audio
|
| 13 |
+
|
| 14 |
+
text_processor:
|
| 15 |
+
train:
|
| 16 |
+
name: "blip_caption"
|
| 17 |
+
eval:
|
| 18 |
+
name: "blip_caption"
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
data_type: [audio]
|
| 22 |
+
|
| 23 |
+
build_info:
|
| 24 |
+
|
| 25 |
+
annotations:
|
| 26 |
+
train:
|
| 27 |
+
url:
|
| 28 |
+
- https://zenodo.org/records/6473207/files/clotho_aqa_train.csv
|
| 29 |
+
storage:
|
| 30 |
+
- clotho_Qa/annotations/clotho_aqa_train.csv
|
| 31 |
+
val:
|
| 32 |
+
url:
|
| 33 |
+
- https://zenodo.org/records/6473207/files/clotho_aqa_val.csv
|
| 34 |
+
storage:
|
| 35 |
+
- clotho_qa/annotations/clotho_aqa_val.csv
|
| 36 |
+
|
| 37 |
+
test:
|
| 38 |
+
url:
|
| 39 |
+
- https://zenodo.org/records/6473207/files/clotho_aqa_test.csv
|
| 40 |
+
storage:
|
| 41 |
+
- clotho_qa/annotations/clotho_aqa_test.csv
|
| 42 |
+
audio:
|
| 43 |
+
storage: /export/einstein-vision/audio_datasets/clotho-aqa/audio_files
|
| 44 |
+
|
LAVIS-main/lavis/configs/datasets/coco/defaults_cap.yaml
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coco_caption: # name of the dataset builder
|
| 8 |
+
dataset_card: dataset_card/coco_caption.md
|
| 9 |
+
# data_dir: ${env.data_dir}/datasets
|
| 10 |
+
data_type: images # [images|videos|features]
|
| 11 |
+
|
| 12 |
+
build_info:
|
| 13 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 14 |
+
annotations:
|
| 15 |
+
train:
|
| 16 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_train.json
|
| 17 |
+
md5: aa31ac474cf6250ebb81d18348a07ed8
|
| 18 |
+
storage: coco/annotations/coco_karpathy_train.json
|
| 19 |
+
val:
|
| 20 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_val.json
|
| 21 |
+
md5: b273847456ef5580e33713b1f7de52a0
|
| 22 |
+
storage: coco/annotations/coco_karpathy_val.json
|
| 23 |
+
test:
|
| 24 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_test.json
|
| 25 |
+
md5: 3ff34b0ef2db02d01c37399f6a2a6cd1
|
| 26 |
+
storage: coco/annotations/coco_karpathy_test.json
|
| 27 |
+
images:
|
| 28 |
+
storage: coco/images/
|
LAVIS-main/lavis/configs/datasets/coco/defaults_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coco_caption_instruct: # name of the dataset builder
|
| 8 |
+
dataset_card: dataset_card/coco_caption.md
|
| 9 |
+
# data_dir: ${env.data_dir}/datasets
|
| 10 |
+
data_type: images # [images|videos|features]
|
| 11 |
+
|
| 12 |
+
vis_processor:
|
| 13 |
+
train:
|
| 14 |
+
name: "clip_image_train"
|
| 15 |
+
image_size: 224
|
| 16 |
+
eval:
|
| 17 |
+
name: "clip_image_eval"
|
| 18 |
+
image_size: 224
|
| 19 |
+
|
| 20 |
+
text_processor:
|
| 21 |
+
train:
|
| 22 |
+
name: blip_instruction
|
| 23 |
+
modality: image
|
| 24 |
+
task: caption
|
| 25 |
+
eval:
|
| 26 |
+
name: blip_caption
|
| 27 |
+
|
| 28 |
+
build_info:
|
| 29 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 30 |
+
annotations:
|
| 31 |
+
train:
|
| 32 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_train.json
|
| 33 |
+
md5: aa31ac474cf6250ebb81d18348a07ed8
|
| 34 |
+
storage: coco/annotations/coco_karpathy_train.json
|
| 35 |
+
# val:
|
| 36 |
+
# url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_val.json
|
| 37 |
+
# md5: b273847456ef5580e33713b1f7de52a0
|
| 38 |
+
# storage: coco/annotations/coco_karpathy_val.json
|
| 39 |
+
# test:
|
| 40 |
+
# url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_test.json
|
| 41 |
+
# md5: 3ff34b0ef2db02d01c37399f6a2a6cd1
|
| 42 |
+
# storage: coco/annotations/coco_karpathy_test.json
|
| 43 |
+
images:
|
| 44 |
+
storage: /export/share/datasets/vision/coco/images
|
LAVIS-main/lavis/configs/datasets/coco/defaults_ret.yaml
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coco_retrieval:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_train.json
|
| 16 |
+
md5: aa31ac474cf6250ebb81d18348a07ed8
|
| 17 |
+
storage: coco/annotations/coco_karpathy_train.json
|
| 18 |
+
val:
|
| 19 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_val.json
|
| 20 |
+
md5: b273847456ef5580e33713b1f7de52a0
|
| 21 |
+
storage: coco/annotations/coco_karpathy_val.json
|
| 22 |
+
test:
|
| 23 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/datasets/coco_karpathy_test.json
|
| 24 |
+
md5: 3ff34b0ef2db02d01c37399f6a2a6cd1
|
| 25 |
+
storage: coco/annotations/coco_karpathy_test.json
|
| 26 |
+
images:
|
| 27 |
+
storage: coco/images/
|
LAVIS-main/lavis/configs/datasets/coco/defaults_vqa.yaml
ADDED
|
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coco_vqa:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url:
|
| 16 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_train.json
|
| 17 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val.json
|
| 18 |
+
storage:
|
| 19 |
+
- coco/annotations/vqa_train.json
|
| 20 |
+
- coco/annotations/vqa_val.json
|
| 21 |
+
val:
|
| 22 |
+
url:
|
| 23 |
+
# TODO make this order insensitive
|
| 24 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val_eval.json
|
| 25 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
|
| 26 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_OpenEnded_mscoco_val2014_questions.json
|
| 27 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_mscoco_val2014_annotations.json
|
| 28 |
+
storage:
|
| 29 |
+
- coco/annotations/vqa_val_eval.json
|
| 30 |
+
- coco/annotations/answer_list.json
|
| 31 |
+
- coco/annotations/v2_OpenEnded_mscoco_val2014_questions.json
|
| 32 |
+
- coco/annotations/v2_mscoco_val2014_annotations.json
|
| 33 |
+
test:
|
| 34 |
+
url:
|
| 35 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_test.json
|
| 36 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
|
| 37 |
+
storage:
|
| 38 |
+
- coco/annotations/vqa_test.json
|
| 39 |
+
- coco/annotations/answer_list.json
|
| 40 |
+
images:
|
| 41 |
+
storage: coco/images/
|
LAVIS-main/lavis/configs/datasets/coco/defaults_vqa_instruct.yaml
ADDED
|
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coco_vqa_instruct:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: "clip_image_train"
|
| 14 |
+
image_size: 224
|
| 15 |
+
eval:
|
| 16 |
+
name: "clip_image_eval"
|
| 17 |
+
image_size: 224
|
| 18 |
+
|
| 19 |
+
text_processor:
|
| 20 |
+
train:
|
| 21 |
+
name: blip_instruction
|
| 22 |
+
modality: image
|
| 23 |
+
task: qa
|
| 24 |
+
eval:
|
| 25 |
+
name: blip_caption
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 29 |
+
annotations:
|
| 30 |
+
train:
|
| 31 |
+
url:
|
| 32 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_train.json
|
| 33 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val.json
|
| 34 |
+
storage:
|
| 35 |
+
- coco/annotations/vqa_train.json
|
| 36 |
+
- coco/annotations/vqa_val.json
|
| 37 |
+
# val:
|
| 38 |
+
# url:
|
| 39 |
+
# # TODO make this order insensitive
|
| 40 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val_eval.json
|
| 41 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
|
| 42 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_OpenEnded_mscoco_val2014_questions.json
|
| 43 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_mscoco_val2014_annotations.json
|
| 44 |
+
# storage:
|
| 45 |
+
# - coco/annotations/vqa_val_eval.json
|
| 46 |
+
# - coco/annotations/answer_list.json
|
| 47 |
+
# - coco/annotations/v2_OpenEnded_mscoco_val2014_questions.json
|
| 48 |
+
# - coco/annotations/v2_mscoco_val2014_annotations.json
|
| 49 |
+
# test:
|
| 50 |
+
# url:
|
| 51 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_test.json
|
| 52 |
+
# - https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
|
| 53 |
+
# storage:
|
| 54 |
+
# - coco/annotations/vqa_test.json
|
| 55 |
+
# - coco/annotations/answer_list.json
|
| 56 |
+
images:
|
| 57 |
+
storage: /export/share/datasets/vision/coco/images
|
LAVIS-main/lavis/configs/datasets/coco/eval_vqa.yaml
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coco_vqa:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
val:
|
| 15 |
+
url:
|
| 16 |
+
# TODO make this order insensitive
|
| 17 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/vqa_val_eval.json
|
| 18 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/answer_list.json
|
| 19 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_OpenEnded_mscoco_val2014_questions.json
|
| 20 |
+
- https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/vqav2/v2_mscoco_val2014_annotations.json
|
| 21 |
+
storage:
|
| 22 |
+
- coco/annotations/vqa_val_eval.json
|
| 23 |
+
- coco/annotations/answer_list.json
|
| 24 |
+
- coco/annotations/v2_OpenEnded_mscoco_val2014_questions.json
|
| 25 |
+
- coco/annotations/v2_mscoco_val2014_annotations.json
|
| 26 |
+
images:
|
| 27 |
+
storage: coco/images/
|
LAVIS-main/lavis/configs/datasets/coin/defaults_cap.yaml
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coin_caption: # name of the dataset builder
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: videos # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: alpro_video_train
|
| 14 |
+
n_frms: 4
|
| 15 |
+
image_size: 224
|
| 16 |
+
min_scale: 0.9
|
| 17 |
+
max_scale: 1.0
|
| 18 |
+
full_video: False
|
| 19 |
+
eval:
|
| 20 |
+
name: alpro_video_eval
|
| 21 |
+
n_frms: 4
|
| 22 |
+
image_size: 224
|
| 23 |
+
min_scale: 0.9
|
| 24 |
+
max_scale: 1.0
|
| 25 |
+
full_video: False
|
| 26 |
+
|
| 27 |
+
text_processor:
|
| 28 |
+
train:
|
| 29 |
+
name: blip_caption
|
| 30 |
+
eval:
|
| 31 |
+
name: blip_caption
|
| 32 |
+
|
| 33 |
+
build_info:
|
| 34 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 35 |
+
annotations:
|
| 36 |
+
train:
|
| 37 |
+
url:
|
| 38 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/train.json
|
| 39 |
+
# - /export/video-language-dataset/data/coin/annotations/train_lavis.json
|
| 40 |
+
storage:
|
| 41 |
+
- coin/annotations/train.json
|
| 42 |
+
# - /export/video-language-dataset/data/coin/annotations/train_lavis.json
|
| 43 |
+
val:
|
| 44 |
+
url:
|
| 45 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/val.json
|
| 46 |
+
# - /export/video-language-dataset/data/coin/annotations/val_lavis.json
|
| 47 |
+
storage:
|
| 48 |
+
- coin/annotations/val.json
|
| 49 |
+
# - /export/video-language-dataset/data/coin/annotations/val_lavis.json
|
| 50 |
+
videos:
|
| 51 |
+
storage: /export/video-language-dataset/data/coin/annotations/videos/
|
LAVIS-main/lavis/configs/datasets/coin/defaults_cap_instruct.yaml
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
coin_caption_instruct: # name of the dataset builder
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: videos # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: alpro_video_train
|
| 14 |
+
n_frms: 4
|
| 15 |
+
image_size: 224
|
| 16 |
+
min_scale: 0.9
|
| 17 |
+
max_scale: 1.0
|
| 18 |
+
full_video: False
|
| 19 |
+
eval:
|
| 20 |
+
name: alpro_video_eval
|
| 21 |
+
n_frms: 4
|
| 22 |
+
image_size: 224
|
| 23 |
+
min_scale: 0.9
|
| 24 |
+
max_scale: 1.0
|
| 25 |
+
full_video: False
|
| 26 |
+
|
| 27 |
+
text_processor:
|
| 28 |
+
train:
|
| 29 |
+
name: blip_instruction
|
| 30 |
+
task: caption
|
| 31 |
+
modality: image
|
| 32 |
+
eval:
|
| 33 |
+
name: blip_caption
|
| 34 |
+
|
| 35 |
+
build_info:
|
| 36 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 37 |
+
annotations:
|
| 38 |
+
train:
|
| 39 |
+
url:
|
| 40 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/train.json
|
| 41 |
+
# - /export/video-language-dataset/data/coin/annotations/train_lavis.json
|
| 42 |
+
storage:
|
| 43 |
+
- coin/annotations/train.json
|
| 44 |
+
# - /export/video-language-dataset/data/coin/annotations/train_lavis.json
|
| 45 |
+
val:
|
| 46 |
+
url:
|
| 47 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/coin/val.json
|
| 48 |
+
# - /export/video-language-dataset/data/coin/annotations/val_lavis.json
|
| 49 |
+
storage:
|
| 50 |
+
- coin/annotations/val.json
|
| 51 |
+
# - /export/video-language-dataset/data/coin/annotations/val_lavis.json
|
| 52 |
+
videos:
|
| 53 |
+
storage: /export/video-language-dataset/data/coin/annotations/videos/
|
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m.yaml
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
conceptual_caption_12m:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url:
|
| 16 |
+
- /export/home/workspace/datasets/cc12m.json
|
| 17 |
+
storage:
|
| 18 |
+
- conceptual_caption/annotations/cc12m.json
|
| 19 |
+
images:
|
| 20 |
+
storage: conceptual_caption/images_12m
|
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_12m_instruct.yaml
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
conceptual_caption_12m_instruct:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: "clip_image_train"
|
| 14 |
+
image_size: 224
|
| 15 |
+
eval:
|
| 16 |
+
name: "clip_image_eval"
|
| 17 |
+
image_size: 224
|
| 18 |
+
|
| 19 |
+
text_processor:
|
| 20 |
+
train:
|
| 21 |
+
name: blip_instruction
|
| 22 |
+
task: caption
|
| 23 |
+
modality: image
|
| 24 |
+
eval:
|
| 25 |
+
name: blip_caption
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 29 |
+
annotations:
|
| 30 |
+
train:
|
| 31 |
+
url:
|
| 32 |
+
- https://storage.googleapis.com/sfr-xinstructblip-data-research/data/cc12m/x_instructblip_clean.json
|
| 33 |
+
# - /export/home/workspace/datasets/cc12m.json
|
| 34 |
+
storage:
|
| 35 |
+
- conceptual_caption/annotations/cc12m.json
|
| 36 |
+
images:
|
| 37 |
+
storage: conceptual_caption/images_12m
|
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m.yaml
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
conceptual_caption_3m:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url:
|
| 16 |
+
- /export/home/workspace/datasets/cc3m.json
|
| 17 |
+
storage:
|
| 18 |
+
- conceptual_caption/annotations/cc3m.json
|
| 19 |
+
images:
|
| 20 |
+
storage: conceptual_caption/images
|
LAVIS-main/lavis/configs/datasets/conceptual_caption/defaults_3m_instruct.yaml
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2023, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
conceptual_caption_3m_instruct:
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: images # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
vis_processor:
|
| 12 |
+
train:
|
| 13 |
+
name: "clip_image_train"
|
| 14 |
+
image_size: 224
|
| 15 |
+
eval:
|
| 16 |
+
name: "clip_image_eval"
|
| 17 |
+
image_size: 224
|
| 18 |
+
|
| 19 |
+
text_processor:
|
| 20 |
+
train:
|
| 21 |
+
name: blip_instruction
|
| 22 |
+
task: caption
|
| 23 |
+
modality: image
|
| 24 |
+
eval:
|
| 25 |
+
name: blip_caption
|
| 26 |
+
|
| 27 |
+
build_info:
|
| 28 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 29 |
+
annotations:
|
| 30 |
+
train:
|
| 31 |
+
url:
|
| 32 |
+
- /export/home/workspace/datasets/cc3m.json
|
| 33 |
+
storage:
|
| 34 |
+
- conceptual_caption/annotations/cc3m.json
|
| 35 |
+
images:
|
| 36 |
+
storage: conceptual_caption/images
|
LAVIS-main/lavis/configs/datasets/didemo/defaults_ret.yaml
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright (c) 2022, salesforce.com, inc.
|
| 2 |
+
# All rights reserved.
|
| 3 |
+
# SPDX-License-Identifier: BSD-3-Clause
|
| 4 |
+
# For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause
|
| 5 |
+
|
| 6 |
+
datasets:
|
| 7 |
+
didemo_retrieval: # name of the dataset builder
|
| 8 |
+
# data_dir: ${env.data_dir}/datasets
|
| 9 |
+
data_type: videos # [images|videos|features]
|
| 10 |
+
|
| 11 |
+
build_info:
|
| 12 |
+
# Be careful not to append minus sign (-) before split to avoid itemizing
|
| 13 |
+
annotations:
|
| 14 |
+
train:
|
| 15 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/didemo/retrieval_train.json
|
| 16 |
+
storage: didemo/annotations/retrieval_train.json
|
| 17 |
+
val:
|
| 18 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/didemo/retrieval_val.json
|
| 19 |
+
storage: didemo/annotations/retrieval_val.json
|
| 20 |
+
test:
|
| 21 |
+
url: https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/didemo/retrieval_test.json
|
| 22 |
+
storage: didemo/annotations/retrieval_test.json
|
| 23 |
+
videos:
|
| 24 |
+
storage: didemo/videos
|
| 25 |
+
# storage: /export/share/dongxuli/data/didemo_retrieval/videos
|