File size: 4,561 Bytes
1c25c2f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
import csv
import json
import random

# 文件路径
csv_file_path = '/mnt/afs/xueyingyi/meme/generate/E_text_1.csv'  # CSV文件路径
user_input_jsonl_path = '/mnt/afs/xueyingyi/meme/pause_data/all_item/C_generate_pause.jsonl'  # user_input.jsonl文件路径
output_jsonl_path = '/mnt/afs/xueyingyi/meme/pause_data/all_item/Cjson/C_generate_pause.jsonl'  # 输出JSONL文件路径
train_jsonl_path = '/mnt/afs/xueyingyi/meme/pause_data/all_item/Cjson/C_generate_train_pause.jsonl'  # 训练集路径
eval_jsonl_path = '/mnt/afs/xueyingyi/meme/pause_data/all_item/Cjson/C_generate_eval_pause.jsonl'  # 测试集路径
train_config_path = '/mnt/afs/xueyingyi/meme/pause_data/all_item/C_generate_train_pause.jsonl'  # 训练集配置文件路径
eval_config_path = '/mnt/afs/xueyingyi/meme/pause_data/all_item/C_generate_eval_pause.jsonl'  # 测试集配置文件路径

# 读取CSV文件
csv_data = {}
with open(csv_file_path, 'r', encoding='utf-8') as csv_file:
    csv_reader = csv.DictReader(csv_file)
    for row in csv_reader:
        file_name = row['file_name']
        text = row['text'].strip()
        csv_data[file_name] = text  # 存储file_name和text的映射关系

# 读取user_input.jsonl文件
user_input_data = []
with open(user_input_jsonl_path, 'r', encoding='utf-8') as f:
    for line in f:
        user_input_data.append(json.loads(line.strip()))

# 构建JSONL数据
jsonl_data = []
for idx, item in enumerate(user_input_data):
    file_name = item['file_name']
    user_input = item['user_input']
    
    # 检查file_name是否在CSV数据中
    if file_name not in csv_data:
        print(f"警告: {file_name} 在CSV文件中未找到,跳过此条数据")
        continue
    
    # 获取对应的text
    text = csv_data[file_name]
    
    # 构建提示词
    with open('/mnt/afs/xueyingyi/vl2.5/InternVL/inference/text_new.txt', 'r') as prompt_file:
        PROMPT = prompt_file.read()

    with open('/mnt/afs/xueyingyi/vl2.5/InternVL/inference/text_example.txt', 'r') as prompt_file:
        PROMPT_example = prompt_file.read()
    
    # 构建conversations
    conversations = [
        {
            "from": "human",
            "value": f"{PROMPT}<image>\n{PROMPT_example}\n<image>\n{user_input}"
        },
        {
            "from": "gpt",
            "value": text  # 使用CSV中的文字
        }
    ]
    
    # 构建JSON对象(单图像)
    # json_obj = {
    #     "id": idx,
    #     "image": f"/mnt/afs/xueyingyi/image_vague/inpainting_demo/{file_name}",
    #     "conversations": conversations
    # }
    # 构建JSON对象(多图像)
    json_obj = {
        "id": idx,
        "image": [
        f"/mnt/afs/xueyingyi/vl2.5/InternVL/inference/example.jpg",
        f"/mnt/afs/xueyingyi/image_vague/inpainting_demo/{file_name}"
        ], 
        "conversations": conversations
    }
    
    jsonl_data.append(json_obj)

# 保存为JSONL文件
with open(output_jsonl_path, 'w', encoding='utf-8') as f:
    for item in jsonl_data:
        f.write(json.dumps(item, ensure_ascii=False) + '\n')

# 划分训练集和测试集
random.seed(42)  # 设置随机种子以确保可重复性
random.shuffle(jsonl_data)  # 打乱数据
train_size = int(len(jsonl_data) * 0.9)  # 90%训练集
train_data = jsonl_data[:train_size]
eval_data = jsonl_data[train_size:]

# 保存训练集
with open(train_jsonl_path, 'w', encoding='utf-8') as f:
    for item in train_data:
        f.write(json.dumps(item, ensure_ascii=False) + '\n')

# 保存测试集
with open(eval_jsonl_path, 'w', encoding='utf-8') as f:
    for item in eval_data:
        f.write(json.dumps(item, ensure_ascii=False) + '\n')

# 生成训练集配置文件
train_config = {
    "classification_C": {
        "root": "/mnt/afs/xueyingyi/image_vague/inpainting_demo",
        "annotation": train_jsonl_path,
        "data_augment": False,
        "repeat_time": 1,
        "length": len(train_data)
    }
}
with open(train_config_path, 'w', encoding='utf-8') as f:
    json.dump(train_config, f, ensure_ascii=False, indent=4)

# 生成测试集配置文件
eval_config = {
    "classification_C": {
        "root": "/mnt/afs/xueyingyi/image_vague/inpainting_demo",
        "annotation": eval_jsonl_path,
        "data_augment": False,
        "repeat_time": 1,
        "length": len(eval_data)
    }
}
with open(eval_config_path, 'w', encoding='utf-8') as f:
    json.dump(eval_config, f, ensure_ascii=False, indent=4)

print("数据处理完成!")
print(f"训练集大小: {len(train_data)}")
print(f"测试集大小: {len(eval_data)}")